All Chapters

The Missing Chapter

The Art of Fermi Estimation

How to guess anything within an order of magnitude — using nothing but reasoning

An extension of Jordan Ellenberg's "How Not to Be Wrong"

Chapter 91

How Many Piano Tuners Are in Chicago?

You don't need data. You need reasoning.

How many piano tuners are there in Chicago? This was Enrico Fermi's favorite question to pose to his physics students at the University of Chicago, and it's a question that, at first glance, seems absolutely unanswerable without access to some very specific database that probably doesn't exist. But Fermi didn't want his students to look it up. He wanted them to think.

Here's how the thinking goes. Chicago has about 2.7 million people.1 That's roughly a million households. How many of those households have a piano? If you've spent any time in American living rooms, you might guess something like one in ten — some have grand pianos gathering dust in the corner, most don't. So that's about 100,000 pianos in Chicago.

Now, each piano should be tuned about once a year to keep it in decent shape. A piano tuner can tune a piano in about two hours, including travel time. Working an eight-hour day, that's four pianos per day. Working 250 days a year (a standard work year), each tuner handles about 1,000 pianos annually.

So: 100,000 pianos divided by 1,000 pianos per tuner gives us... about 100 piano tuners in Chicago.

The actual number? Somewhere between 100 and 200, depending on how you count part-timers.2 We got it right — not to the exact digit, but to the right order of magnitude. And we did it without looking up a single fact about the piano tuning industry.

The goal isn't precision. The goal is to be roughly right rather than precisely wrong.

This is Fermi estimation: the art of producing reasonable approximations to questions that seem impossible by breaking them into smaller, more tractable pieces. And it works far better than it has any right to.

· · ·
Chapter 91.1

Decomposition: The Heart of the Method

The key insight behind Fermi estimation is almost embarrassingly simple: you can decompose a hard question into easy questions. How many piano tuners are in Chicago? You can't answer that directly. But you can answer: How many people live in Chicago? What fraction own pianos? How often are pianos tuned? How many can one tuner handle?

Each of these sub-questions is something a reasonably informed adult can estimate to within a factor of two or three. And here's where the magic happens: when you multiply several rough estimates together, the errors in your individual guesses tend to cancel out.

How many piano tuners in Chicago? decompose Population 2.7M % with pianos ~4% Tunings/year 1 Tunings/tuner/yr 1,000 multiply & divide ≈ 100 tuners

Fermi decomposition: one impossible question becomes four tractable ones.

Think about what happens if your estimate of piano-owning households is too high by 50% (you guessed 15% instead of 10%), but your estimate of tunings per day is too low by a similar amount (you guessed 3 instead of 4). These errors push your final answer in opposite directions, and they partially cancel. When you're multiplying five or six rough factors, there's a good chance that the overestimates and underestimates roughly wash out.

This isn't just hand-waving. There's a precise mathematical reason for it, and it's connected to one of the most powerful theorems in all of mathematics: the Central Limit Theorem.3

· · ·
Chapter 91.2

The Central Limit Theorem on a Log Scale

Here's the mathematical punchline. When you multiply several estimated quantities together, you're adding their logarithms. If each estimate is off by some random factor — maybe 2× too high, maybe 0.5× too low — then on a logarithmic scale, each error is a random variable centered roughly around zero. The logarithm of your overestimate is positive; the logarithm of your underestimate is negative.

Fermi estimate as product of factors
Estimate = f₁ × f₂ × f₃ × ⋯ × fₙ
Taking logs: log(Estimate) = log(f₁) + log(f₂) + ⋯ + log(fₙ)

And now we're in the land of the Central Limit Theorem. If you're summing a bunch of independent random variables (our log-errors), their sum converges toward a normal distribution. The spread of that distribution grows as √n, while the number of terms grows as n. This means the average log-error gets smaller as you add more factors. More sub-estimates means more chances for errors to cancel, and the cancellation gets more reliable, not less.

This is why Fermi estimation works best when you can decompose your question into many sub-questions. Three factors is good. Seven factors is better. The more you slice the problem, the more the CLT works in your favor — as long as your individual estimates aren't systematically biased in the same direction.4

Why more factors help

If each of your n sub-estimates has independent log-error with standard deviation σ, the total log-error of the product has standard deviation σ√n. But to be off by 10× requires log-error of about 1 (base 10). With 5 factors each with σ ≈ 0.3 (roughly a 2× error), total σ is 0.3 × √5 ≈ 0.67. You'll usually land within one order of magnitude.

· · ·
Chapter 91.3

Fermi at Trinity

The method's namesake didn't just pose cute classroom puzzles. Enrico Fermi used estimation to tackle one of the most consequential questions of the twentieth century.

On July 16, 1945, the first nuclear weapon was detonated at the Trinity test site in New Mexico. Fermi stood ten miles away, watching the fireball bloom on the horizon. As the blast wave arrived about 40 seconds later, he tore small pieces of paper and dropped them — first from about six feet high while the air was still, then again as the blast wave hit.5

The paper scraps were displaced about 2.5 meters by the blast wave. From this — just the distance some confetti traveled — Fermi calculated the energy yield of the bomb at roughly 10 kilotons of TNT. The actual yield, painstakingly measured by sophisticated instruments, was about 21 kilotons. Fermi was right to within a factor of two, using paper and his brain, before any instrument readings were available.

This is the Fermi method at its most dramatic. You don't always need precise instruments or massive datasets. Sometimes a few well-chosen observations and some clear thinking will get you remarkably close to the truth.

Fermi's Trinity Estimate on a Log Scale 1 kt 10 kt 100 kt 1,000 kt Fermi's estimate ~10 kt Actual yield ~21 kt ~2× off 10× would be here

On a log scale, Fermi's paper-scrap estimate was remarkably close to the measured yield.

· · ·
Chapter 91.4

Your Turn: The Fermi Workshop

Enough theory. Let's see if you can think like Fermi. Below, pick a classic Fermi question and build your answer from sub-estimates. Each slider represents a factor in the decomposition. Adjust them according to your best judgment, and see how close you get.

🔬 Fermi Workshop

Select a question, adjust the sub-estimates, and see how close your product lands to the real answer.

Your Fermi estimate

Notice how, even when individual sliders are somewhat off, the final product often lands in the right ballpark. That's error cancellation at work.

· · ·
Chapter 91.5

Why the Errors Cancel

Let's make the error cancellation phenomenon concrete. Suppose the true answer to some question is a product of five factors, and you're estimating each one with up to 50% error in either direction. Intuitively, that sounds catastrophic — five errors of 50% each, compounding? Shouldn't you end up wildly off?

But you don't. And the reason is that "50% too high" and "50% too low" don't compound — they cancel. When errors are random and independent, the ones that push your answer up are roughly balanced by the ones that push it down. The result is that your final product is typically much closer to the truth than any individual factor.

📊 Error Cancellation Demo

Watch how multiplying noisy estimates still converges to the right answer. Each factor has up to ±50% random error. Click "Run Trial" to see a new set of noisy factors, or "Run 50 Trials" to see convergence.

Number of factors
5

Latest Trial

True product:
Noisy product:

Cumulative Results

Run some trials to see convergence...

Run it a few dozen times. You'll see that even with ±50% error on each factor, the geometric mean of many trials clusters remarkably close to the true product. This is the Central Limit Theorem doing its quiet, reliable work on the log scale. The individual logarithmic errors sum to something approximately normal with mean near zero.6

· · ·
Chapter 91.6

Applications: From Google to the Galaxy

Fermi estimation isn't just a party trick for physicists. It has become a core tool across an astonishing range of fields.

Tech interviews. Google famously popularized Fermi questions in job interviews — "How many golf balls fit in a school bus?" The point was never the answer. It was watching how a candidate structures their thinking. Do they freeze at the impossibility, or do they start decomposing? Can they identify the key variables? Are they comfortable with uncertainty? These are exactly the skills that matter when building products in unknown territory.

Startup market sizing. Every venture capitalist has seen a pitch that begins with "The global market for X is $50 billion." The good pitches build up to a number through Fermi reasoning: there are N potential customers, each would pay roughly $Y, with a Z% conversion rate. This bottom-up approach is inherently more trustworthy than top-down market reports, because each assumption can be interrogated independently.7

Scientific plausibility checks. Before running a million-dollar experiment, a quick Fermi estimate can tell you whether your expected effect is even in the detectable range. Physicists call these "back-of-the-envelope calculations," and they've saved countless hours and dollars by killing doomed experiments early.

The Drake Equation. Perhaps the grandest Fermi estimation of all: how many communicating extraterrestrial civilizations exist in our galaxy? Frank Drake decomposed this seemingly impossible question into seven factors: the rate of star formation, the fraction with planets, the fraction with habitable planets, and so on, down to the average lifetime of a communicating civilization.8

The Drake Equation
N = R★ × fp × ne × fl × fi × fc × L
R★
Rate of star formation in our galaxy
fp
Fraction of stars with planetary systems
ne
Habitable planets per system
fl
Fraction that develop life
fi
Fraction that develop intelligence
fc
Fraction that develop detectable technology
L
Years a civilization remains detectable

The Drake Equation is pure Fermi method. Nobody knows the answer to "how many alien civilizations?" But we can take stabs at each factor, and the product — however uncertain — frames the conversation productively. Depending on your assumptions, N ranges from essentially zero to millions. The equation doesn't give us a definitive answer, but it tells us where the uncertainty lives, which is arguably more valuable.

The Drake Funnel: Narrowing to N ~400 billion stars in the Milky Way ~50% have planets → 200B ~20% habitable → 40B life × intelligence × tech × lifespan → N = ???

The Drake Equation as a funnel: each factor narrows the field from hundreds of billions of stars to… maybe a handful of civilizations.

· · ·
Chapter 91.7

The Skill Behind the Trick

If there's one thing to take away from Fermi estimation, it's this: the ability to decompose an impossible question into tractable sub-questions is itself one of the most powerful intellectual skills you can develop.

It's not a math trick. It's a thinking habit. When someone asks "Is this startup idea viable?", the Fermi thinker doesn't throw up their hands or go looking for a market research report. They start decomposing: How many potential users? What would they pay? What's the cost of acquisition? What's the retention rate? Each question is answerable. The product tells a story.

When someone asks "Should we worry about asteroid impacts?", the Fermi thinker estimates: How many near-Earth asteroids are there? How often does one hit? What's the damage from various sizes? Suddenly the vague worry becomes a quantified risk.

Fermi estimation is, at its core, a refusal to be paralyzed by not knowing. It's the conviction that a rough answer obtained through clear reasoning is infinitely more valuable than no answer at all — or worse, a precise answer built on unexamined assumptions.

The question is never "do you know the answer?" It's "can you figure out how to figure it out?"

Fermi knew this. His students at Chicago learned it. And every time you catch yourself saying "I have no idea," you have an opportunity to practice it: stop, decompose, estimate, multiply. You'll be surprised how often you land remarkably close to the truth.

Because the world, it turns out, is more knowable than it looks — if you're willing to think about it one piece at a time.

Notes & References

  1. U.S. Census Bureau, 2020 Census. Chicago city population: 2,746,388. The metro area is much larger (~9.5 million), but for this estimate we stick with the city proper.
  2. Various sources converge on this range. The Bureau of Labor Statistics doesn't track piano tuners specifically, but Yellow Pages listings and industry surveys suggest 100–200 for the Chicago metro area. See Weinstein, L. & Adam, J., Guesstimation (Princeton University Press, 2008).
  3. For the Central Limit Theorem and its applications, see Chapter 63 of this series, or any good probability textbook. The key insight: the sum of many independent random variables converges to a normal distribution regardless of their individual distributions.
  4. Systematic bias is the enemy of Fermi estimation. If every factor is biased upward (e.g., optimism bias in market sizing), the errors compound rather than cancel. The CLT cancellation only works for independent errors centered around the truth.
  5. Fermi, E. "My Observations During the Explosion at Trinity on July 16, 1945." Reprinted in The Collected Papers of Enrico Fermi, Vol. II (University of Chicago Press, 1965), pp. 943–944.
  6. More precisely: if each factor's multiplicative error has mean 1 and log-variance σ², then the product of n such factors has log-variance nσ², giving standard deviation σ√n on the log scale. For the geometric mean of many trials (each being a product of n noisy factors), the CLT applies to the log-products.
  7. Blank, S. "How to Build a Startup: The Market Size Myth." Harvard Business Review, 2013. Bottom-up market sizing through Fermi decomposition consistently outperforms top-down analyst estimates for early-stage companies.
  8. Drake, F. "The Drake Equation Revisited." Presented at the Astrobiology Science Conference, 2010. Drake first wrote the equation in 1961 as an agenda for the first SETI meeting at Green Bank, West Virginia.