info@ceoworld.biz
Tuesday, March 19, 2024
CEOWORLD magazine - Latest - CEO Agenda - Defending Your Organization From Misinformation via Neuroscience

CEO Agenda

Defending Your Organization From Misinformation via Neuroscience

Whenever you hear something repeated, it feels more true when you hear it repeated. In other words, repetition makes any statement seem more true. So anything you hear will feel more true each time you hear it again.

Do you see what I did there? Each of the three sentences above conveyed the same message. Yet each time you read the next sentence, it felt more and more true. Cognitive neuroscientists like myself call this the “illusory truth effect.”

Go back and recall your experience reading the first sentence. It probably felt strange and disconcerting, perhaps with a tone of outrage, as in “I don’t believe things more if they’re repeated!”

Reading the second sentence did not inspire such a strong reaction. Your reaction to the third sentence was tame by comparison.

Why? Because of a phenomenon called “cognitive fluency,” meaning how easily we process information. Much of our vulnerability to deception in all areas of life – including misinformation – revolves around cognitive fluency in one way or another.

Now think about how rumors spread in your organization’s grapevine. It works on the same principle. Employees hear a rumor – say about a proposed headquarters move, just like Elon Musk’s move of Tesla’s HQ to Texas. It feeds into their fears, which is a very cognitively fluid part of our mind.

They repeat the rumor, and it goes around, and then they keep hearing it from others. It begins to seem more and more true, regardless of reality. Before you know it, those who want to stay where they are looking for another job, even though you might never have intended to move your headquarters!

Fortunately, we can learn about these mental errors, which helps us address misinformation and make our workplaces more truthful.

The Lazy Brain

Our brains are lazy. The more effort it takes to process information, the more uncomfortable we feel about it and the more we dislike and distrust it.

By contrast, the more we like certain data and are comfortable with it, the more we feel that it’s accurate. This intuitive feeling in our gut is what we use to judge what’s true and false.

Yet no matter how often you heard that you should trust your gut and follow your intuition, that advice is wrong. You should not trust your gut when evaluating information where you don’t have expert-level knowledge, at least when you don’t want to screw up. Structured information gathering and decision-making processes help us avoid the numerous errors we make when we follow our intuition. And even experts can make serious errors when they don’t rely on such decision aids.

These mistakes happen due to mental errors that scholars call “cognitive biases.” The illusory truth effect is one of these mental blindspots; there are over 100 altogether. These mental blindspots impact all areas of our life, from health and politics to relationships.

Other Important Cognitive Biases

Besides illusory truth, what are some other cognitive biases you need to beware of to protect your organization from misinformation? If you’ve heard of any cognitive biases, you’ve likely heard of the “confirmation bias.” That refers to our tendency to look for and interpret information in ways that conform to our prior beliefs, intuitions, feelings, desires, and preferences, as opposed to the facts.

Again, cognitive fluency deserves blame. It’s much easier to build neural pathways to information that we already possess, especially that around which we have strong emotions; it’s much more difficult to break well-established neural pathways if we need to change our mind based on new information. Consequently, we instead look for information that’s easy to accept, that which fits our prior beliefs. In turn, we ignore and even actively reject information that doesn’t fit our beliefs.

Moreover, the more educated we are, the more likely we are to engage in such active rejection. After all, our smarts give us more ways of arguing against new information that counters our beliefs. That’s why research demonstrates that the more educated you are, the more polarized your beliefs will be around scientific issues that have religious or political value overtones, such as stem cell research, human evolution, and climate change. Where might you and your team be letting your smarts get in the way of the facts?

Our minds like to interpret the world through stories, meaning explanatory narratives that link cause and effect in a clear and simple manner. Such stories are a balm to our cognitive fluency, as our mind constantly looks for patterns that explain the world around us in an easy-to-process manner. That leads to the “narrative fallacy,” where we fall for convincing-sounding narratives regardless of the facts, especially if the story fits our predispositions and our emotions.

You ever wonder why politicians tell so many stories? What about the advertisements you see on TV or video advertisements on websites, which tell very quick visual stories? How about salespeople or fundraisers? Sure, sometimes they cite statistics and scientific reports, but they spend much, much more time telling stories: simple, clear, compelling narratives that seem to make sense and tug at our heartstrings.

Now, here’s something that’s actually true: the world doesn’t make sense. The world is not simple, clear, and compelling. The world is complex, confusing, and contradictory. Beware of simple stories! Look for complex, confusing, and contradictory scientific reports and high-quality statistics: they’re much more likely to contain the truth than the easy-to-process stories.

Fixing Our Brains

Unfortunately, knowledge only weakly protects us from cognitive biases; it’s important, but far from sufficient.

What can we do? You can use decision aid strategies to address cognitive biases to defend your organization from misinformation.

One of the most effective strategies is to help your employees and yourself build up a habit of automatically considering alternative possibilities to any claim you hear, especially claims that feel comfortable to you. Since our lazy brain’s default setting is to avoid questioning claims, which requires hard thinking, it really helps to develop a mental practice of going against this default. Be especially suspicious of repeated claims that make you feel comfortable without any additional evidence, which play on the illusory truth effect and the confirmation bias combined.

Another effective strategy involves cultivating a mental habit of questioning stories in particular. Whenever you hear a story, the brain goes into a listening and accepting mode. Remember that it’s very easy to cherry-pick stories to support whatever position the narrator wants to advance. Instead, look for thorough hard numbers, statistical evidence, and peer-reviewed research to support claims.

More broadly, you can encourage employees to make a personal commitment to the twelve truth-oriented behaviors of the Pro-Truth Pledge by signing the pledge at ProTruthPledge.org. All of these behaviors stem from cognitive neuroscience and behavioral economics research in the field called debiasing, which refers to counterintuitive, uncomfortable, but effective strategies to protect yourself from cognitive biases. Peer-reviewed research has shown that taking the Pro-Truth Pledge is effective for changing people’s behavior to be more truthful, both in their own statements and in interactions with others.

These quick mental habits will address the most fundamentally flawed aspects of our mind’s tendency to accept misinformation.


Written by Dr. Gleb Tsipursky. Have you read?
World’s Richest Race Car Drivers, Richest Musicians, and Richest Rappers.


Add CEOWORLD magazine to your Google News feed.
Follow CEOWORLD magazine headlines on: Google News, LinkedIn, Twitter, and Facebook.

This report/news/ranking/statistics has been prepared only for general guidance on matters of interest and does not constitute professional advice. You should not act upon the information contained in this publication without obtaining specific professional advice.
No representation or warranty (express or implied) is given as to the accuracy or completeness of the information contained in this publication, and, to the extent permitted by law, CEOWORLD magazine does not accept or assume any liability, responsibility or duty of care for any consequences of you or anyone else acting, or refraining to act, in reliance on the information contained in this publication or for any decision based on it. This publication (and any extract from it) must not be copied, redistributed or placed on any website, without CEOWORLD magazine' prior written consent.


Copyright 2024 The CEOWORLD magazine. All rights reserved. This material may not be published, broadcast, rewritten, or redistributed. For media queries, please contact: info@ceoworld.biz
SUBSCRIBE NEWSLETTER
CEOWORLD magazine - Latest - CEO Agenda - Defending Your Organization From Misinformation via Neuroscience
Dr. Gleb Tsipursky
Dr. Gleb Tsipursky, P.h.D, is the CEO of the boutique future-of-work consultancy Disaster Avoidance Experts. He is the best-selling author of seven books, including Never Go With Your Gut: How Pioneering Leaders Make the Best Decisions and Avoid Business Disasters and Leading Hybrid and Remote Teams: A Manual on Benchmarking to Best Practices for Competitive Advantage. His expertise comes from over 20 years of consulting for Fortune 500 companies from Aflac to Xerox and over 15 years in academia as a behavioral scientist at UNC-Chapel Hill and Ohio State.


Dr. Gleb Tsipursky is an opinion columnist for the CEOWORLD magazine. Connect with him through LinkedIn. For more information, visit the author’s website.