No more culture wars? Corporate America is finally realizing that getting ‘woke’ is bad for business
Major brands and Hollywood studios appear to be questioning the benefits of waging “culture wars” as investors reverse course