Something is wrong with American journalism. Long before "fake news" became the calling card of the Right, Americans had lost faith in their news media. But lately, the feeling that something is off has become impossible to ignore. That's because the majority of our mainstream news is no longer just liberal; it's woke. Today's newsrooms are propagating radical ideas that were fringe as recently as a decade ago, including "antiracism," intersectionality,...