Friday, February 17, 2023

When Motives Don't Matter


    There's quite a bit of talk lately about truth and accuracy in speech and print, and what we are to make of the various ways in which we are being misled. It has been recently reported that the AI wonder, ChatGPT, is capable of being wildly wrong, pretty much at the top of its artificial voice. It lies, it argues with people when called on its lies, and it cites to sources that do not exist. That's a lot like Tucker Carlson, whom (among others at Fox) has been just outed in a lawsuit as being well-aware of the deliberate nature of his duplicitous double-dealing -- for example, praising Trump on-air while calling him a "demonic force" behind his back, capable of "easily destroying (Fox News)." If only. Then of course there's the Liar-In-Chief himself and all his lying MAGA minions, of which little more needs to be said. 
    One of the more interesting questions about all this is the extent to which these walking, talking misinformation generators are dissembling because they
A. Believe their bullshit is true, or
B. Don't believe what they're spreading, but do it just to agitate the snowflakes, or
C. Are playing some kind of "Overton Window" game, or
D. Are self-dealing Fascists who will do and say whatever it takes.
    I recognize that those are not exclusive categories, and that human psychology is such that a person could hold all of those perspectives virtually simultaneously. (See "Tucker Carlson" and "Trump," above.)  That being the case, I therefore propose the following rule: If someone is lying, and they reject valid evidence of their mistake -- much like ChatGTP -- just pull the plug. Save your breath. Move on and don't look back. You may be the sort of person who suffers fools (gladly or otherwise), but liars are a whole next level.

No comments: