The perils of assuming everything is fine: Normalcy bias and the rushed approval of the Boeing 737 Max 10 jet
Find out why normally bias, and the perils of assuming everything is fine, has worryingly led to the rushed approval of the Boeing 737 Max 10 jet – and what you can learn from this lesson.
Congress just cleared the Boeing 737 Max 10 jet for certification in the omnibus end-of-year spending bill, without further safety enhancements.
That’s despite significant opposition by those demanding a safety upgrade: from the union representing the 15,000 pilots at American Airlines, from the families of those killed in the two deadly crashes in 2019, and from Rep. Peter DeFazio, chair of the House Transportation Committee.
Rep. DeFazio led the key congressional investigation into the Max crashes, and said the language in the spending bill was included over his objection.
Neither Boeing nor Congress has learned the lesson of Boeing’s earlier 737 Max fiasco
This rushed clearance stemmed from the pressure of lobbying by Boeing and its allies. It suggests neither Boeing nor Congress learned the lesson of Boeing’s earlier 737 Max fiasco when 346 people lost their lives, Boeing lost $5 billion in direct revenue and over $25 billion when counting damage to the brand and losing customers, and Boeing fired its CEO Dennis Muilenburg.
What caused the disaster for Boeing? At a high level, it was the company’s desire to keep up with Airbus’s newer, more fuel-efficient aircraft, the Airbus 320. To do this, Boeing rushed the production of the 737 Max and provided misleading information to the Federal Aviation Administration (FAA) in order to receive fast approval for the plane.
Why Boeing’s disaster was caused by an error called ‘normalcy bias’
The root cause of the disaster at Boeing can be traced back to a cognitive error known as normalcy bias. This bias causes people to overestimate the likelihood that things will continue as they have been and underestimate the potential consequences of a disaster occurring.
Ironically, the transformation of the airline industry in recent decades to make airplanes much safer and accidents incredibly rare is key to understanding Boeing’s disaster. The Boeing leadership was overconfident in the safety record of their airplanes and saw the FAA certification process as an obstacle to doing business rather than a necessary safety measure.
This normalcy bias contributed to their decision to rush the production of the 737 Max and overlook known software issues.
Boeing’s 737 Max disaster is a classic case of the normalcy bias. The Boeing leadership felt utter confidence in the safety record of the airplanes it produced in the last couple of decades, deservedly so, according to statistics on crashes.
From their perspective, it would be impossible to imagine that the 737 Max would be less safe than these other recent-model airplanes. They saw the typical FAA certification process as simply another bureaucratic hassle that got in the way of doing business and competing with Airbus, as opposed to ensuring safety.
Think it’s only big companies? Think again.
The normalcy bias is a big reason for bubbles: in stocks, housing prices, loans, and other areas. It’s as though we’re incapable of remembering the previous bubble, even if occurred only a few years ago.
Similarly, the normalcy bias helps explain why leaders at companies of all sizes were so vastly underprepared for COVID-19 and its impact. While pandemics post a major threat, it’s a low-likelihood, high-impact, slow-moving disaster. The normalcy bias keeps tripping us up on such disasters, unless we take effective steps to deal with this problem.
How normalcy bias led to the failure of a tech start-up
Of course, the normalcy bias hits mid-size and small companies hard as well.
At one of my frequent trainings for small and mid-size company executives, Brodie, a tech entrepreneur shared the story of a startup he founded with a good friend. They complemented each other well: Brodie had strong technical skills, and his friend brought strong marketing and selling capacity.
Things went great for the first two and a half years, with a growing client list – until his friend got into a bad motorcycle accident that left him unable to talk. Brodie had to deal not only with the emotional trauma, but also with covering his co-founder’s work roles.
Unfortunately, his co-founder failed to keep good notes. He also did not introduce Brodie to his contacts at the client companies. In turn, Brodie – a strong introvert – struggled with selling. Eventually, the startup burned through its cash and had to close its doors.
The normalcy bias is one of many dangerous judgment errors, mental blindspots resulting from how our brains are wired. Researchers in cognitive neuroscience and behavioral economics call them cognitive biases. Fortunately, recent research in these fields shows how you can use pragmatic strategies to address these dangerous judgment errors in your professional life.
How to prevent normalcy bias disasters
It really helps to use the strategy of considering and addressing potential alternative futures that are much more negative than you intuitively feel are likely. That’s the strategy that Brodie and I explored in my coaching with him after the training session, as he felt ready to get back to the startup world.
While Brodie definitely knew he wouldn’t be up to starting a new business himself, he also wanted to avoid the previous problems. So we discussed how he would from the start push for creating systems and processes that would enable each co-founder to back up the other in cases of emergencies.
Moreover, the co-founders would commit to sharing important contacts from their side of the business with each other, so that relationships could be maintained if the other person was out of commission for a while.
So what are the broader principles here?
- Be much more pessimistic about the possibility and impact of disasters than you intuitively feel or can easily imagine, to get over the challenges caused by the normalcy bias.
- Use effective strategic planning techniques to scan for potential disasters and try to address them in advance, as Brodie did with his plans for the new business.
- Of course, you can’t predict everything, so retain some extra capacity in your system – of time, money, and other resources – that you can use to deal with unknown unknowns, also called black swans.
- Finally, if you see a hint of a disaster, react much more quickly than you intuitively feel you should to overcome the gut reaction’s dismissal of the likelihood and impact of disasters.
Unfortunately, Boeing – and Congress – did not appear to learn this lesson in the rushed approval of the new 737 Max model. The fact that they failed to make the safety upgrade demanded by so many diverse external stakeholders signals that more deadly lessons may be in store for us in the future.
Dr. Gleb Tsipursky helps tech and finance industry executives drive collaboration, innovation, and retention in hybrid work. He serves as the CEO of the boutique future-of-work consultancy Disaster Avoidance Experts.
He is the best-selling author of seven books, including Never Go With Your Gut and Leading Hybrid and Remote Teams. His cutting-edge thought leadership was featured in over 650 articles in prominent venues such asHarvard Business Review, Fortune, and Forbes.
His expertise comes from over 20 years of consulting for Fortune 500 companies from Aflac to Xerox and over 15 years in academia as a behavioral scientist at UNC-Chapel Hill and Ohio State. A proud Ukrainian American, Dr. Gleb lives in Columbus, Ohio.
Photo by Tim Trad