Hidden flaws in strategy

Can insights from behavioral economics explain why good executives back bad strategies?

May 2003| byCharles Roxburgh

After nearly 40 years, the theory of business strategy is well developed and widely disseminated. Pioneering work by academics such as Michael E. Porter and Henry Mintzberg has established a rich literature on good strategy. Most senior executives have been trained in its principles, and large corporations have their own skilled strategy departments.

Podcast

Hidden flaws in strategy

Yet the business world remains littered with examples of bad strategies. Why? What makes chief executives back them when so much know-how is available? Flawed analysis, excessive ambition, greed, and other corporate vices are possible causes, but this article doesn’t attempt to explore all of them. Rather, it looks at one contributing factor that affects every strategist: the human brain.

The brain is a wondrous organ. As scientists uncover more of its inner workings through brain-mapping techniques,11.See Rita Carter, Mapping the Mind, London: Phoenix, 2000.
our understanding of its astonishing abilities increases. But the brain isn’t the rational calculating machine we sometimes imagine. Over the millennia of its evolution, it has developed shortcuts, simplifications, biases, and basic bad habits. Some of them may have helped early humans survive on the savannas of Africa ("if it looks like a wildebeest and everyone else is chasing it, it must be lunch"), but they create problems for us today. Equally, some of the brain’s flaws may result from education and socialization rather than nature. But whatever the root cause, the brain can be a deceptive guide for rational decision making.

These implications of the brain’s inadequacies have been rigorously studied by social scientists and particularly by behavioral economists, who have found that the underlying assumption behind modern economics—human beings as purely rational economic decision makers—doesn’t stack up against the evidence. As most of the theory underpinning business strategy is derived from the rational world of microeconomics, all strategists should be interested in behavioral economics.

Insights from behavioral economics have been used to explain bad decision making in the business world,22.See, for example, J. Edward Russo and Paul J. H. Schoemaker, Decision Traps: The Ten Barriers to Brilliant Decision-Making and How to Overcome Them, New York: Fireside, 1990; and John S. Hammond III, Ralph L. Keeney, and Howard Raiffa, 'The hidden traps in decision making,' Harvard Business Review, September–October 1998, pp. 47–57.
and bad investment decision making in particular. Some private equity firms have successfully remodeled their investment processes to counteract the biases predicted by behavioral economics. Likewise, behavioral economics has been applied to personal finance,33.See Gary Belsky and Thomas Gilovich, Why Smart People Make Big Money Mistakes and How to Correct Them, New York: Simon and Schuster, 1999.
thereby providing an easier route to making money than any hot stock tip. However, the field hasn’t permeated the day-to-day world of strategy formulation.

This article aims to help rectify that omission by highlighting eight44.This is far from a complete list of all the flaws in the way we make decisions. For a full description of the irrational biases in decision making, see Jonathan Baron, Thinking and Deciding, New York: Cambridge University Press, 1994.
insights from behavioral economics that best explain some examples of bad strategy. Each insight illustrates a common flaw that can draw us to the wrong conclusions and increase the risk of betting on bad strategy. All the examples come from a field with which I am familiar—European financial services—but equally good ones could be culled from any industry.

Several examples come from the dot-com era, a particularly rich period for students of bad strategy. But don’t make the mistake of thinking that this was an era of unrepeatable strategic madness. Behavioral economics tells us that the mistakes made in the late 1990s were exactly the sorts of errors our brains are programmed to make—and will probably make again.

Flaw 1: Overconfidence

Our brains are programmed to make us feel overconfident. This can be a good thing; for instance, it requires great confidence to launch a new business. Only a few start-ups will become highly successful. The world would be duller and poorer if our brains didn’t inspire great confidence in our own abilities. But there is a downside when it comes to formulating and judging strategy.

The brain is particularly overconfident of its ability to make accurate estimates. Behavioral economists often illustrate this point with simple quizzes: guess the weight of a fully laden jumbo jet or the length of the River Nile, say. Participants are asked to offer not a precise figure but rather a range in which they feel 90 percent confidence—for example, the Nile is between 2,000 and 10,000 miles long. Time and again, participants walk into the same trap: rather than playing safe with a wide range, they give a narrow one and miss the right answer. (I scored 0 out of 15 on such a test, which was one of the triggers of my interest in this field!) Most of us are unwilling and, in fact, unable to reveal our ignorance by specifying a very wide range. Unlike John Maynard Keynes, most of us prefer being precisely wrong rather than vaguely right.

We also tend to be overconfident of our own abilities.55.In a 1981 survey, for example, 90 percent of Swedes described themselves as above-average drivers.
This is a particular problem for strategies based on assessments of core capabilities. Almost all financial institutions, for instance, believe their brands to be of "above-average" value.

Related to overconfidence is the problem of overoptimism. Other than professional pessimists such as financial regulators, we all tend to be optimistic, and our forecasts tend toward the rosier end of the spectrum. The twin problems of overconfidence and overoptimism can have dangerous consequences when it comes to developing strategies, as most of them are based on estimates of what may happen—too often on unrealistically precise and overoptimistic estimates of uncertainties.

One leading investment bank sensibly tested its strategy against a pessimistic scenario—the market conditions of 1994, when a downturn lasted about nine months—and built in some extra downturn. But this wasn’t enough. The 1994 scenario looks rosy compared with current conditions, and the bank, along with its peers, is struggling to make dramatic cuts to its cost base. Other sectors, such as banking services for the affluent and on-line brokerages, are grappling with the same problem.

There are ways to counter the brain’s overconfidence:

Test strategies under a much wider range of scenarios. But don’t give managers a choice of three, as they are likely to play safe and pick the central one. For this reason, the pioneers of scenario planning at Royal Dutch/Shell always insisted on a final choice of two or four options.66.See Pierre Wack’s two-part article, 'Scenarios: Uncharted waters ahead,' Harvard Business Review, September–October 1985, pp. 73–89; and 'Scenarios: Shooting the rapids,' Harvard Business Review, November–December 1985, pp. 139–50.

Add 20 to 25 percent more downside to the most pessimistic scenario.77.This rule of thumb was suggested by Belsky and Gilovich in Why Smart People Make Big Money Mistakes and How to Correct Them.
Given our optimism, the risk of getting pessimistic scenarios wrong is greater than that of getting the upside wrong. The Lloyd’s of London insurance market—which has learned these lessons the hard, expensive way—makes a point of testing the market’s solvency under a series of extreme disasters, such as two 747 aircraft colliding over central London. Testing the resilience of Lloyd’s to these conditions helped it build its reserves and reinsurance to cope with the September 11 disaster.

Build more flexibility and options into your strategy to allow the company to scale up or retrench as uncertainties are resolved. Be skeptical of strategies premised on certainty.

Flaw 2: Mental accounting

Richard Thaler, a pioneer of behavioral economics, coined the term "mental accounting," defined as "the inclination to categorize and treat money differently depending on where it comes from, where it is kept, and how it is spent."88.See Belsky and Gilovich.
Gamblers who lose their winnings, for example, typically feel that they haven’t really lost anything, though they would have been richer had they stopped while they were ahead.

Mental accounting pervades the boardrooms of even the most conservative and otherwise rational corporations. Some examples of this flaw include the following:

being less concerned with value for money on expenses booked against a restructuring charge than on those taken through the P&L

imposing cost caps on a core business while spending freely on a start-up

creating new categories of spending, such as "revenue-investment spend" or "strategic investment"

All are examples of spending that tends to be less scrutinized because of the way it is categorized, but all represent real costs.

These delusions can have serious strategic implications. Take cost caps. In some UK financial institutions during the dot-com era, core retail businesses faced stringent constraints on their ability to invest, however sound the proposal, while start-up Internet businesses spent with abandon. These banks have now written off much of their loss from dot-com investment and must reverse their underinvestment in core businesses.

Avoiding mental accounting traps should be easier if you adhere to a basic rule: that every pound (or dollar or euro) is worth exactly that, whatever the category. In this way, you will make sure that all investments are judged on consistent criteria and be wary of spending that has been reclassified. Be particularly skeptical of any investment labeled "strategic."

Flaw 3: The status quo bias

In one classic experiment,99.This is a simplified account of an experiment conducted by William Samuelson and Richard Zeck-hauser, described in 'Status-quo bias in decision making,' Journal of Risk and Uncertainty, 1, March 1988, pp. 7–59.
students were asked how they would invest a hypothetical inheritance. Some received several million dollars in low-risk, low-return bonds and typically chose to leave most of the money alone. The rest received higher-risk securities—and also left most of the money alone. What determined the students’ allocation in this experiment was the initial allocation, not their risk preference. People would rather leave things as they are. One explanation for the status quo bias is aversion to loss—people are more concerned about the risk of loss than they are excited by the prospect of gain. The students’ fear of switching into securities that might end up losing value prevented them from making the rational choice: rebalancing their portfolios.

A similar bias, the endowment effect, gives people a strong desire to hang on to what they own; the very fact of owning something makes it more valuable to the owner. Richard Thaler tested this effect with coffee mugs imprinted with the Cornell University logo. Students given one of them wouldn’t part with it for less than $5.25, on average, but students without a mug wouldn’t pay more than $2.75 to acquire it. The gap implies an incremental value of $2.50 from owning the mug.

The status quo bias, the aversion to loss, and the endowment effect contribute to poor strategy decisions in several ways. First, they make CEOs reluctant to sell businesses. McKinsey research shows that divestments are a major potential source of value creation but a largely neglected one.1010.See Lee Dranikoff, Tim Koller, and Antoon Schneider, 'Divestiture: Strategy’s missing link,' Harvard Business Review, May–June 2002, pp. 75–83; and Richard N. Foster and Sarah Kaplan, Creative Destruction: Why Companies That Are Built to Last Underperform the Market—and How to Successfully Transform Them, New York: Currency/Doubleday, 2001.
CEOs are prone to ask, "What if we sell for too little—how stupid will we look when this turns out to be a great buy for the acquirer?" Yet successful turnarounds, such as the one at Bankers Trust in the 1980s, often require a determined break with the status quo and an extensive reshaping of the portfolio—in that case, selling all of the bank’s New York retail branches.

These phenomena also make it hard for companies to shift their asset allocations. Before the recent market downturn, the UK insurer Prudential decided that equities were overvalued and made the bold decision to rebalance its fund toward bonds. Many other UK life insurers, unwilling to break with the status quo, stuck with their high equity weightings and have suffered more severe reductions in their solvency ratios.

This isn’t to say that the status quo is always wrong. Many investment advisers would argue that the best long-term strategy is to buy and hold equities (and, behavioral economists would add, not to check their value for many years, to avoid feeling bad when prices fall). In financial services, too, caution and conservatism can be strategic assets. The challenge for strategists is to distinguish between a status quo option that is genuinely the right course and one that feels deceptively safe because of an innate bias.

To make this distinction, strategists should take two approaches:

Adopt a radical view of all portfolio decisions. View all businesses as "up for sale." Is the company the natural parent, capable of extracting the most value from a subsidiary? View divestment not as a failure but as a healthy renewal of the corporate portfolio.

Subject status quo options to a risk analysis as rigorous as change options receive. Most strategists are good at identifying the risks of new strategies but less good at seeing the risks of failing to change.

Flaw 4: Anchoring

One of the more peculiar wiring flaws in the brain is called anchoring. Present the brain with a number and then ask it to make an estimate of something completely unrelated, and it will anchor its estimate on that first number. The classic illustration is the Genghis Khan date test. Ask a group of people to write down the last three digits of their phone numbers, and then ask them to estimate the date of Genghis Khan’s death. Time and again, the results show a correlation between the two numbers; people assume that he lived in the first millennium, when in fact he lived from 1162 to 1227.

Anchoring can be a powerful tool for strategists. In negotiations, naming a high sale price for a business can help secure an attractive outcome for the seller, as the buyer’s offer will be anchored around that figure. Anchoring works well in advertising too. Most retail-fund managers advertise their funds on the basis of past performance. Repeated studies have failed to show any statistical correlation between good past performance and future performance. By citing the past-performance record, though, the manager anchors the notion of future top-quartile performance to it in the consumer’s mind.

However, anchoring—particularly becoming anchored to the past—can be dangerous. Most of us have long believed that equities offer high real returns over the long term, an idea anchored in the experience of the past two decades. But in the 1960s and 1970s, UK equities achieved real annual returns of only 3.3 and 0.4 percent, respectively. Indeed, they achieved double-digit real annual returns during only 4 of the past 13 decades. Our expectations about equity returns have been seriously distorted by recent experience.

In the insurance industry, changes in interest rates have caused major problems due to anchoring. The United Kingdom’s Equitable Life Assurance Society assumed that high nominal interest rates would prevail for decades and sold guaranteed annuities accordingly. That assumption had severe financial consequences for the company and its policyholders. The banking industry may now be entering a period of much higher credit losses than it experienced during the past decade. Some banks may be caught out by the speed of change.

Besides remaining unswayed by the anchoring tactics of others, strategists should take a long historical perspective. Put trends in the context of the past 20 or 30 years, not the past 2 or 3; for certain economic indicators, such as equity returns or interest rates, use a very long time series of 50 or 75 years. Some commentators who spotted the dot-com bubble early did so by drawing comparisons with previous technology bubbles—for example, the uncannily close parallels between radio stocks in the 1920s and Internet stocks in the 1990s.

Flaw 5: The sunk-cost effect

A familiar problem with investments is called the sunk-cost effect, otherwise known as "throwing good money after bad." When large projects overrun their schedules and budgets, the original economic case no longer holds, but companies still keep investing to complete them.

Financial institutions often face this dilemma over large-scale IT projects. There are numerous examples, most of which remain private. One of the more public cases was the London Stock Exchange’s automated-settlement system, Taurus. It took the intervention of the Bank of England to force a cancellation, write off the expenses, and take control of building a replacement.

Executives making strategic-investment decisions can also fall into the sunk-cost trap. Certain European banks spent fortunes building up large equities businesses to compete with the global investment-banking firms. It then proved extraordinarily hard for some of these banks to face up to the strategic reality that they had no prospect of ever competing successfully against the likes of Goldman Sachs, Merrill Lynch, and Morgan Stanley in the equities business. Some banks in the United Kingdom took the agonizing decision to write off their investments; other European institutions are still caught in the trap.

Why is it so hard to avoid? One explanation is based on loss aversion: we would rather spend an additional $10 million completing an uneconomic $110 million project than write off $100 million. Another explanation relies on anchoring: once the brain has been anchored at $100 million, an additional $10 million doesn’t seem so bad.

What should strategists do to avoid the trap?

Apply the full rigor of investment analysis to incremental investments, looking only at incremental prospective costs and revenues. This is the textbook response to the sunk-cost fallacy, and it is right.

Be prepared to kill strategic experiments early. In an increasingly uncertain world, companies will often pursue several strategic options.1111.See Eric D. Beinhocker, 'Robust adaptive strategies,' Sloan Management Review, spring 1999, pp. 95–106; Hugh Courtney, Jane Kirkland, and Patrick Viguerie, 'Strategy under uncertainty,' Harvard Business Review, November–December 1997, pp. 67–79; and Lowell L. Bryan, 'Just-in-time strategy for a turbulent world,' The McKinsey Quarterly, 2002 Number 2 special edition: Risk and resilience, pp. 16–27.
Successfully managing a portfolio of them entails jettisoning the losers. The more quickly you get out, the lower the sunk costs and the easier the exit.

Use "gated funding" for strategic investments, much as pharmaceutical companies do for drug development: release follow-on funding only once strategic experiments have met previously agreed targets.

Flaw 6: The herding instinct

The banking industry, like many others, shows a strong herding instinct. It tends to lend too much money to the same kinds of borrowers at the same time—to UK property developers in the 1970s, less-developed countries in the 1980s, and technology, media, and telecommunications companies more recently. And banks tend to pursue the same strategies, be it creating Internet banks with strange-sounding names during the dot-com boom or building integrated investment banks at the time of the "big bang," when the London stock market was liberalized.

This desire to conform to the behavior and opinions of others is a fundamental human trait and an accepted principle of psychology.1212.See Belsky and Gilovich.
Warren Buffett put his finger on this flaw when he wrote, "Failing conventionally is the route to go; as a group, lemmings may have a rotten image, but no individual lemming has ever received bad press."1313.Warren Buffett, 'Letter from the chairman,' Berkshire Hathaway Annual Report, 1984.
For most CEOs, only one thing is worse than making a huge strategic mistake: being the only person in the industry to make it.

We all felt the tug of the herd during the dot-com era. It was lonely being a Luddite, arguing the case against setting up a stand-alone Internet bank or an on-line brokerage. At times of mass enthusiasm for a strategic trend, pressure to follow the herd rather than rely on one’s own information and analysis is almost irresistible. Yet the best strategies break away from the trend. Some actions may be necessary to match the competition—imagine a bank without ATMs or a good on-line banking offer. But these are not unique sources of strategic advantage, and finding such sources is what strategy is all about. "Me-too" strategies are often simply bad ones.1414.See Philipp M. Nattermann, 'Best practice does not equal best strategy,' The McKinsey Quarterly, 2000 Number 2, pp. 22–31.
Seeking out the new and the unusual should therefore be the strategist’s aim. Rather than copying what your most established competitors are doing, look to the periphery1515.See Foster and Kaplan.
for innovative ideas, and look outside your own industry.

Initially, an innovative strategy might draw skepticism from industry experts. They may be right, but as long as you kill a failing strategy early, your losses will be limited, and when they are wrong, the rewards will be great.

Flaw 7: Misestimating future hedonic states

What does it mean, in plain English, to misestimate future hedonic states? Simply that people are bad at estimating how much pleasure or pain they will feel if their circumstances change dramatically. Social scientists have shown that when people undergo major changes in circumstances, their lives typically are neither as bad nor as good as they had expected—another case of how bad we are at estimating. People adjust surprisingly quickly, and their level of pleasure (hedonic state) ends up, broadly, where it was before.

This research strikes a chord with anyone who has studied compensation trends in the investment-banking industry. Ever-higher compensation during the 1990s led only to ever-higher expectations—not to a marked change in the general level of happiness on the Street. According to Tom Wolfe’s Sherman McCoy, in Bonfire of the Vanities, it was hard to make ends meet in New York on $1 million a year in 1987. Back then, that was shocking hubris from a (fictional) top bond salesman. By 2000, even adjusted for inflation, it would have seemed a perfectly reasonable lament from a relatively junior managing director.

Another illustration of our poor ability to judge future hedonic states in the business world is the way we deal with a loss of independence. More often than not, takeovers are seen as the corporate equivalent of death, to be avoided at all costs. Yet sometimes they are the right move. Two once great British banks—Midland and National Westminster—both struggled to maintain their independence. Midland gave in to HSBC’s advances in 1992; NatWest was taken over by the Royal Bank of Scotland in 2000. At both institutions, the consequences were positive for customers, shareholders, and most employees on any test of the "greatest good of the greatest number." The employees ended up being part of better-managed, stronger, more respected institutions. Morale at NatWest has gone up. Midland has achieved what was, for an independent bank, an unrealistic goal: to become part of a great global bank.

Often, top management is blamed for resisting any loss of independence. Certainly part of the problem is the desire of managements and boards to hang on to the status quo. That said, frontline staff members often resist a takeover or merger however much they are frustrated with the existing top management. Some deeper psychological factor appears to be at work. We do seem very bad at estimating how we would feel if our circumstances changed dramatically—changes in corporate control, like changes in our personal health or wealth.

How can the strategist avoid this pitfall?

In takeovers, adopt a dispassionate and unemotional view. Easier said than done—especially for a management team with years of committed service to an institution and a personal stake in the status quo. Nonexecutives, however, should find it easier to maintain a detached view.

Keep things in perspective. Don’t overreact to apparently deadly strategic threats or get too excited by good news. During the high and low points of the crisis at Lloyd’s of London in the mid-1990s, the chairman used to quote Field Marshall Slim—"In battle nothing is ever as good or as bad as the first reports of excited men would have it." This is a good guide for every strategist trying to navigate a crisis, with the inevitable swings in emotion and morale.

Flaw 8: False consensus

People tend to overestimate the extent to which others share their views, beliefs, and experiences—the false-consensus effect. Research shows many causes, including these:

confirmation bias, the tendency to seek out opinions and facts that support our own beliefs and hypotheses

selective recall, the habit of remembering only facts and experiences that reinforce our assumptions

biased evaluation, the quick acceptance of evidence that supports our hypotheses, while contradictory evidence is subjected to rigorous evaluation and almost certain rejection; we often, for example, impute hostile motives to critics or question their competence

groupthink,1616.Famously described in Irving Lester Janis’s study of the Bay of Pigs and Cuban missile crises, among others. See his book Groupthink: Psychological Studies of Policy Decisions and Fiascoes, Boston: Houghton Mifflin, June 1982.
the pressure to agree with others in team-based cultures

Consider how many times you may have heard a CEO say something like, "the executive team is 100 percent behind the new strategy" (groupthink); "the chairman and the board are fully supportive and they all agree with our strategy" (false consensus); "I’ve heard only good things from dealers and customers about our new product range" (selective recall); "OK, so some analysts are still negative, but those ’teenage scribblers’ don’t understand our business—their latest reports were superficial and full of errors" (biased evaluation). This hypothetical CEO might be right but more likely is heading for trouble. The role of any strategic adviser should be to provide a counterbalance to this tendency toward false consensus. CEOs should welcome the challenge.

False consensus, which ranks among the brain’s most pernicious flaws, can lead strategists to miss important threats to their companies and to persist with doomed strategies. But it can be extremely difficult to uncover—especially if those proposing a strategy are strong role models. We are easily influenced by dominant individuals and seek to emulate them. This can be a force for good if the role models are positive. But negative ones can prove an irresistible source of strategic error.

Many of the worst financial-services strategies can be attributed to over-dominant individuals. The failure of several Lloyd’s syndicates in the 1980s and 1990s was due to powerful underwriters who controlled their own agencies. And overdominant individuals are associated with several more recent insurance failures. In banking, one European institution struggled to impose effective risk disciplines because its seemingly most successful employees were, in the eyes of junior staff, cavalier in their approach to compliance. Their behavior set the tone and created a culture of noncompliance.

The dangers of false consensus can be minimized in several ways:

Create a culture of challenge. As part of the strategic debate, management teams should value open and constructive criticism. Criticizing a fellow director’s strategy should be seen as a helpful, not a hostile, act. CEOs and strategic advisers should understand criticisms of their strategies, seek contrary views on industry trends, and, if in doubt, take steps to assure themselves that opposing views have been well researched. They shouldn’t automatically ascribe to critics bad intentions or a lack of understanding.

Ensure that strong checks and balances control the dominant role models. A CEO should be particularly wary of dominant individuals who dismiss challenges to their own strategic proposals; the CEO should insist that these proposals undergo an independent review by respected experts. The board should be equally wary of a domineering CEO.

Don’t "lead the witness." Instead of asking for a validation of your strategy, ask for a detailed refutation. When setting up hypotheses at the start of a strategic analysis, impose contrarian hypotheses or require the team to set up equal and opposite hypotheses for each key analysis. Establish a "challenger team" to identify the flaws in the strategy being proposed by the strategy team.

An awareness of the brain’s flaws can help strategists steer around them. All strategists should understand the insights of behavioral economics just as much as they understand those of other fields of the "dismal science." Such an understanding won’t put an end to bad strategy; greed, arrogance, and sloppy analysis will continue to provide plenty of textbook cases of it. Understanding some of the flaws built into our thinking processes, however, may help reduce the chances of good executives backing bad strategies.

Stay connected

About Insights & Publications

The creation of knowledge supports McKinsey’s core mission: helping our clients achieve distinctive, lasting, and substantial performance improvements. We publish our insights and those of external experts to help advance the practice of management and provide leaders with facts on which to base business and policy decisions. Views expressed by third-party authors are theirs alone.