No more culture wars? Corporate America is finally realizing that getting ‘woke’ is bad for business

Major brands and Hollywood studios appear to be questioning the benefits of waging “culture wars” as investors reverse course