All Chapters

The Missing Chapter

The Beautiful Mind's Ugliest Insight

Nash equilibrium: where everyone is doing the best they can, and everything is terrible.

An extension of Jordan Ellenberg's "How Not to Be Wrong"

Chapter 25

A One-Sentence Letter

In the fall of 1948, a twenty-year-old from Bluefield, West Virginia, arrived at Princeton with a letter of recommendation that read, in its entirety: "This man is a genius." The letter writer was Richard Duffin, a mathematician at Carnegie Tech. He wasn't prone to hyperbole.

John Forbes Nash Jr. had applied to Princeton's math department with essentially no credentials beyond that sentence and a few undergraduate research results. He'd been rejected by Harvard — they offered money, but not enough, and Nash took Princeton's fuller fellowship instead. It was one of the great accidents of intellectual history. At Princeton, Nash would encounter the one-year-old field of game theory, and within two years he'd transform it with a twenty-seven-page doctoral thesis that would eventually win the Nobel Prize in Economics.1

But here's the thing about Nash's great insight: it's not a story about how to win. It's a story about getting stuck.

· · ·

What the Movie Gets Wrong

If you've seen A Beautiful Mind — Ron Howard's 2001 biopic of Nash, starring Russell Crowe — you probably remember the bar scene. Nash and his friends spot a group of women, including one stunning blonde. Nash has his eureka moment: if all the men pursue the blonde, they'll block each other and then get rejected by the other women too (who don't want to be second choices). The optimal strategy, movie-Nash declares, is for everyone to ignore the blonde and each approach a different brunette.2

It's a charming scene. It's also completely wrong about Nash equilibrium.

Here's why: if all your friends are ignoring the blonde, then you should go talk to the blonde. No competition! The "everyone ignores the blonde" strategy is unstable precisely because any individual player can do better by deviating from it. And that instability is exactly what Nash equilibrium rules out.

A Nash equilibrium is a set of strategies — one for each player — where no player can improve their outcome by unilaterally changing their strategy, given what everyone else is doing.

It's not about cooperation. It's not about what's optimal for the group. It's about what's stable: a situation where nobody has an incentive to budge.

Think about that for a moment. Nash didn't discover how to win games. He discovered which outcomes are self-enforcing — which situations persist because nobody wants to be the one to change. And sometimes those stable situations are terrible for everyone involved.

The Definition (It's Short)

Nash's thesis was twenty-seven pages. The core proof used Brouwer's fixed-point theorem — a deep result from topology that says, roughly, that if you continuously stir a cup of coffee, at least one molecule ends up back where it started.3 Nash used this to prove something remarkable:

Nash's Existence Theorem (1950)

Every finite game — any game with a finite number of players and a finite number of strategies — has at least one Nash equilibrium, possibly in mixed strategies.

"Mixed strategies" means players can randomize. You might play Rock with probability 1/3, Paper with probability 1/3, and Scissors with probability 1/3. In Rock-Paper-Scissors, that uniform mix is the Nash equilibrium — no matter what your opponent does, your expected payoff stays the same, so you can't improve by shifting your mix.

Let's make this concrete. The most famous game in game theory is the Prisoner's Dilemma:

Player 2 Cooperate Defect Player 1 Cooperate Defect 3 , 3 0 , 5 5 , 0 1 , 1 ← Nash Equilibrium

The Prisoner's Dilemma: both players defecting is the only Nash equilibrium, even though both cooperating would be better for everyone.

If Player 2 cooperates, Player 1 gets 3 from cooperating but 5 from defecting. Defect is better. If Player 2 defects, Player 1 gets 0 from cooperating but 1 from defecting. Defect is still better. No matter what the other player does, defecting is the best response. Both players defect. Both get 1. Both could have gotten 3.

This is Nash equilibrium's dark poetry. The stable outcome — the one nobody has a reason to deviate from — leaves everyone worse off than the cooperative outcome that nobody can sustain.

· · ·

Try It Yourself: The Equilibrium Finder

Don't take my word for it. Here's a 2×2 game where you set the payoffs. The finder will compute the Nash equilibria — pure strategy, mixed strategy, or both — and highlight the best responses for each player.

Nash Equilibrium Finder

Enter payoffs as (Row, Column) for each cell. Click "Find Equilibria" to analyze.

Col: LeftCol: Right
Row: Up
,
,
Row: Down
,
,
· · ·

The Zoo of Games

The Prisoner's Dilemma gets all the press, but it's only one species in a menagerie of strategic situations. Each has its own flavor of trouble.

Chicken (a.k.a. Hawk-Dove)

Two drivers race toward each other. Each can swerve or go straight. If both go straight, catastrophe. If one swerves and the other doesn't, the swerver is a "chicken" and the other wins. If both swerve, it's a draw. This game has two pure Nash equilibria — (Swerve, Straight) and (Straight, Swerve) — and a mixed equilibrium where each player randomizes. The problem isn't that you're stuck in a bad outcome; it's that you can't agree on which good outcome to land on.4

Battle of the Sexes

A couple wants to spend the evening together. One prefers opera, the other prefers a football game. Going together to either event beats going alone to your preferred one. Again, two pure equilibria — both at opera, or both at football — and a mixed one. The challenge is coordination: how do you end up at the same place?

Stag Hunt

Two hunters can hunt a stag together (big payoff, requires cooperation) or each hunt a hare alone (small but guaranteed payoff). Both hunting stag is a Nash equilibrium. Both hunting hare is also a Nash equilibrium. The stag equilibrium is better for everyone, but the hare equilibrium is safer — you can't get burned by your partner defecting.5

Pure Conflict Pure Coordination Zero-Sum Prisoner's Dilemma Chicken Stag Hunt Battle of the Sexes Multiple equilibria → coordination problem

The spectrum of games: from pure conflict (zero-sum) to pure coordination, with mixed-motive games in between.

What all these games share is the coordination problem. When a game has multiple equilibria, how do players end up at one rather than another? This is where culture, convention, and history come in — and where Nash equilibrium, as a concept, starts to feel incomplete.

The Coordination Game

Here's a game that demonstrates the problem viscerally. You'll play against a simple AI opponent across multiple rounds. You both choose between two options. You score when you match. But neither option is inherently better — you just need to find each other.

The Coordination Game

Choose A or B each round. You score 1 point when you match the other player. The AI adapts to your patterns — can a convention emerge?

Round 1 — Make your choice!
0
Matches
0
Round
Match %

If you played enough rounds, you probably noticed something: a convention emerged. You and the AI settled into a pattern — maybe both choosing A, maybe alternating. This is exactly how real-world conventions work. We drive on the right side of the road in the US, on the left in the UK. Neither is inherently better. The Nash equilibrium is just "do whatever everyone else is doing."6

· · ·

When the Real World Gets Stuck

Nash equilibrium would be merely an elegant mathematical concept if it didn't describe so many real situations with such uncomfortable accuracy.

Traffic and Braess's Paradox

Every morning, thousands of commuters independently choose their routes. Each driver picks the route that's fastest given what everyone else is doing. The result is a Nash equilibrium — no single driver can improve their commute by switching routes. But the equilibrium can be dreadful. In 1968, Dietrich Braess proved something astonishing: adding a new road to a network can make everyone's commute worse. The new road creates a Nash equilibrium where everyone uses it, causing congestion that makes all routes slower.7

This isn't hypothetical. In 1990, New York City closed 42nd Street for Earth Day and traffic actually improved. Seoul demolished a six-lane highway in 2003 and commute times went down.

The Arms Race

The Cold War was a Prisoner's Dilemma at civilizational scale. The US and USSR each faced the same logic: whatever the other side does, building more nuclear weapons is the "best response." If they disarm, you should arm (dominance). If they arm, you should arm (deterrence). Both countries armed to the teeth. Both were worse off than mutual disarmament. But mutual disarmament wasn't a Nash equilibrium — each side had an incentive to cheat.

OPEC and Oil Production

The Organization of the Petroleum Exporting Countries agrees to limit oil production to keep prices high. Every member benefits from the cartel. But every individual member benefits even more from secretly overproducing while everyone else restricts. The Nash equilibrium is for everyone to overproduce, which is exactly what OPEC countries routinely do. The cartel survives not because of Nash's theory but despite it — through monitoring, side payments, and Saudi Arabia's willingness to be the swing producer.

Technology Standards

VHS versus Betamax. Blu-ray versus HD DVD. Lightning versus USB-C. These are coordination games. Everyone benefits from having a single standard, but which one? Once enough people adopt one standard, the Nash equilibrium is for everyone to use it — even if it's technically inferior. VHS won not because it was better than Betamax (it wasn't) but because it reached the tipping point first.

The Price of Anarchy

In 1999, Elias Koutsoupias and Christos Papadimitriou formalized this gap between stable and optimal. They defined the Price of Anarchy: the ratio between the worst Nash equilibrium and the socially optimal outcome.8

The Price of Anarchy
PoA = Cost(worst Nash equilibrium) ÷ Cost(social optimum)
A Price of Anarchy of 1 means selfish behavior is costless. Higher values mean society pays dearly for individual rationality.

A Price of Anarchy of 1 means the Nash equilibrium is as good as the optimum — selfishness costs nothing. A Price of Anarchy of 2 means the stable outcome is twice as costly as what a benevolent dictator could arrange. For general routing networks, Tim Roughgarden proved the Price of Anarchy is at most 4/3 for linear latency functions — not great, not catastrophic. But for other games, it can be arbitrarily large.

Optimal Nash Equilibrium Price of Anarchy Total social cost (lower is better)

The Price of Anarchy: the gap between what rational selfishness produces and what coordinated action could achieve.

This is, I think, the deepest lesson of Nash equilibrium. The mathematics of individual rationality does not promise collective flourishing. Every commuter choosing the fastest route doesn't produce the fastest commute. Every country choosing the safest military posture doesn't produce the safest world. Every company choosing the most profitable strategy doesn't produce the most productive economy.

Nash equilibrium tells us where we'll end up. It doesn't promise we'll like it there.

The Beautiful and the Damned

Nash himself lived a life that mirrored his theorem's tensions. His brilliant early career was followed by decades of schizophrenia. He wandered the halls of Princeton as a ghost of his former self, scrawling strange equations on blackboards. Then, remarkably, he emerged. By the 1990s, his illness had receded. In 1994, he shared the Nobel Memorial Prize in Economics with John Harsanyi and Reinhard Selten.

In his Nobel autobiography, Nash wrote something striking: he wouldn't wish mental illness on anyone, but he was "grateful" for the experience in some ways, because without it he might have spent his life in conventional academic competition — another rat in the Nash equilibrium of academic prestige games.

I find that poignant. The man who discovered that rational individuals get stuck in suboptimal equilibria found that his own escape from such an equilibrium required a departure from rationality itself.

Mathematics can tell us where the stable outcomes are. What it can't tell us — what no formal system can — is how to muster the collective will to reach the outcomes we'd actually choose. That's not a math problem. That's a human one.

Notes & References

  1. Nash, John F. "Equilibrium Points in N-Person Games." Proceedings of the National Academy of Sciences 36, no. 1 (1950): 48–49. His full dissertation, "Non-Cooperative Games," was completed at Princeton in 1950.
  2. The scene appears about 25 minutes into A Beautiful Mind (2001). Sylvia Nasar's biography of the same name (Simon & Schuster, 1998) gives the more accurate history.
  3. Formally: every continuous function from a closed ball to itself has a fixed point. Nash's use of the Kakutani fixed-point theorem (a generalization) was the key technical contribution.
  4. The game of Chicken appears in Bertrand Russell's Common Sense and Nuclear Warfare (1959), where he used it as a metaphor for nuclear brinksmanship.
  5. The Stag Hunt originates from Jean-Jacques Rousseau's A Discourse on Inequality (1755). Brian Skyrms's The Stag Hunt and the Evolution of Social Structure (Cambridge, 2004) gives the full game-theoretic treatment.
  6. Thomas Schelling's The Strategy of Conflict (Harvard, 1960) introduced the concept of "focal points" — salient solutions that players gravitate toward without communication.
  7. Braess, Dietrich. "Über ein Paradoxon aus der Verkehrsplanung." Unternehmensforschung 12 (1968): 258–268. The New York City and Seoul examples are documented in Roughgarden, Tim. Selfish Routing and the Price of Anarchy (MIT Press, 2005).
  8. Koutsoupias, Elias, and Christos Papadimitriou. "Worst-case equilibria." Proceedings of the 16th Annual Symposium on Theoretical Aspects of Computer Science (1999): 404–413.