All Chapters

The Missing Chapter

The Central Limit Theorem

Why everything is bell-shaped — and when it spectacularly isn't

An extension of Jordan Ellenberg's "How Not to Be Wrong"

Chapter 63

The Supreme Law of Unreason

In 1889, the Victorian polymath Francis Galton built a machine to watch randomness organize itself. He called it the quincunx — a cabinet filled with pegs, through which small lead balls would tumble, bouncing left or right at each peg with equal probability. At the bottom, the balls piled up. And every time, no matter how many balls you dropped, the pile formed the same gentle, symmetrical curve. Galton was thunderstruck. "I know of scarcely anything so apt to impress the imagination," he wrote, as this demonstration of "the supreme law of Unreason."1

What Galton was watching — and what he couldn't quite explain with the mathematical tools at his disposal — was the Central Limit Theorem in action. It is, depending on who you ask, the most important theorem in all of statistics, or the most important theorem in all of applied mathematics, or simply the reason the bell curve shows up everywhere you look.

Here's the thing about the bell curve that should strike you as deeply weird: it shows up in places it has no business being. The heights of American women? Bell curve. The errors in astronomical measurements? Bell curve. The total weight of a truckload of oranges? Bell curve. The number of heads in a thousand coin flips? Bell curve. These are completely different phenomena. Heights are determined by genetics and nutrition. Measurement errors come from shaky hands and imperfect instruments. Orange weights are a function of sunlight and soil and pest management. And yet they all produce the same shape.

That's not a coincidence. It's a theorem.

• • •

A Machine for Making Bell Curves

Let's start where Galton started: with the quincunx. Imagine a vertical board studded with pegs arranged in rows. You drop a ball at the top. It hits the first peg and bounces either left or right — call it a fifty-fifty chance. Then it hits a peg in the next row, and bounces left or right again. And so on, row after row, until it lands in one of the bins at the bottom.

Galton's quincunx: each ball makes a series of random left-right decisions, yet the collective result is always a bell curve.

Each individual ball's path is random and unpredictable. Ball #47 might bounce left-left-right-left-right and end up slightly to the left of center. Ball #48 might go right-right-right-right-right and land way off on the right edge. But the aggregate — the pile — follows a beautiful, predictable pattern. The bins near the center fill up the most, the bins on the edges the least, and the shape of the pile is that famous bell.

Why? Because each ball's final position is the sum of many small random nudges. Left counts as −1, right counts as +1, and the ball's final position is the total of all those nudges. Most balls get a roughly even mix of lefts and rights, so they end up near the center. To land on the far edge, you'd need an extraordinary run of luck — all lefts or all rights — and that's rare. The bell curve is, in this sense, just a counting argument about what's likely and what isn't.

Try it yourself:

🔴 Galton Board (Quincunx)

Watch balls tumble through pegs and pile into a bell curve. Adjust rows to see how more random steps make a smoother curve.

• • •
Chapter 63

Adding Things Up

The quincunx is charming, but it only shows one source distribution: the fair coin flip (left or right with equal probability). The real power of the Central Limit Theorem is that it works for almost any source distribution. Uniform, lopsided, bimodal, weird — it doesn't matter. Add enough of them together, and you get a bell curve.

Let me show you with dice. Roll a single die: you get 1, 2, 3, 4, 5, or 6, each equally likely. That's the uniform distribution. A bar chart of outcomes from one die is perfectly flat — six bars, all the same height. Not a bell curve at all.

Now roll two dice and add them up. Suddenly you have the familiar distribution every Settlers of Catan player knows: 7 is the most common sum, and 2 and 12 are rare. The shape is triangular — a pointy tent, not a bell. But it's already more bell-like than one die was.

Roll ten dice and add them up. Now the distribution is nearly indistinguishable from a perfect Gaussian bell curve. And you started with the most boring, flat distribution imaginable.2

1 Die Uniform 2 Dice Triangular 10 Dice Nearly Gaussian
From flat to bell: one die is uniform, two dice make a triangle, ten dice make a near-perfect Gaussian. The dashed red line is the theoretical normal curve.

This is the central insight: sums of independent random things tend to be normally distributed, regardless of what the individual things look like. The original distribution could be flat, could be exponential, could be shaped like a camel with two humps — it doesn't matter. Add enough independent copies together, take the average, and the bell curve emerges like a universal attractor.

The Central Limit Theorem (Lindeberg-Lévy)

Let X₁, X₂, …, Xₙ be independent, identically distributed random variables with mean μ and finite variance σ². Then as n → ∞:

√n · ( μ) / σ N(0, 1)

The standardized sample mean converges in distribution to a standard normal.

The sample mean: (X₁ + X₂ + … + Xₙ) / n
μ
The population mean (expected value of each Xᵢ)
σ
The population standard deviation
n
The sample size
N(0,1)
The standard normal distribution

Read that statement carefully. It says finite variance. It says independent. Those conditions matter enormously, and we'll come back to them. But when they hold — and they hold a lot of the time in the real world — the theorem is stunningly powerful.

Try it for yourself. Pick any distribution you like — including some deeply un-bell-shaped ones — and watch what happens to the average as you increase the sample size:

📊 CLT Demonstrator

Choose a source distribution. We'll repeatedly take samples of size N, compute their mean, and histogram the results. As N grows, the histogram converges to a Gaussian (shown in red).

• • •
Chapter 63

A Brief History of the Bell

The Central Limit Theorem didn't arrive fully formed. It was built up over two centuries by mathematicians who each saw a piece of the picture.

The story starts with Abraham de Moivre, a French Huguenot living in exile in London, making his living by tutoring aristocrats in probability and frequenting the coffeehouses where gamblers came to him for advice. In 1733, de Moivre proved that the binomial distribution — the number of heads in n coin flips — approaches a bell-shaped curve as n grows large.3 This was the first version of the CLT, though de Moivre wouldn't have called it that. He was just trying to compute binomial probabilities without having to multiply enormous numbers together.

Pierre-Simon Laplace, the great French mathematician who wanted to calculate the probability of everything in the universe, generalized de Moivre's result. By 1810, Laplace had shown that the theorem worked not just for coin flips but for sums of other random variables too.4 He used it to analyze astronomical measurement errors — when you take the average of many observations, the error in that average is approximately normally distributed, which is exactly what astronomers needed to hear.

Then came Gauss, who put his name on the curve (somewhat unfairly — de Moivre and Laplace got there first). And over the next century, mathematicians gradually loosened the requirements, proving the CLT under weaker and weaker assumptions, until Lindeberg and Lévy produced the clean modern version in the 1920s: all you need is independence and finite variance.5

"The normal distribution is not normal because it's common. It's common because of a theorem."

Why It Works: The Convolution Argument

There's a beautifully intuitive way to see why the CLT works, if you think about it in terms of shapes.

When you add two independent random variables, their probability distributions get convolved. Convolution is a specific mathematical operation — you slide one distribution across the other and integrate the overlap — but visually, it's a smoothing operation. Take a rectangle (uniform distribution) and convolve it with itself: you get a triangle. Convolve again: you get a smoother bump. Each convolution sands off the corners and pushes the shape closer to a bell.

n = 1 n = 2 n = 4 n = 10 Repeated convolution smooths any shape into a Gaussian
Convolution in action: a uniform distribution convolved with itself repeatedly approaches a Gaussian. By n = 10, the fit is nearly perfect (dashed red = exact Gaussian).

There's an even more elegant way to see it if you know a little Fourier analysis. In Fourier space, convolution becomes multiplication. Each distribution has a characteristic function — its Fourier transform — and the characteristic function of the sum is the product of the individual characteristic functions. When you take the logarithm, products become sums. And by a Taylor expansion argument, the logarithm of any characteristic function, near the origin, looks like a quadratic — which is the log of a Gaussian.6 So the product of many characteristic functions converges to a Gaussian characteristic function. That's the CLT, proven in four sentences.

(Well, four sentences plus an appendix of rigorous analysis. But the idea is four sentences.)

• • •
Chapter 63

When the Bell Curve Breaks

I've been telling you that the CLT works for "almost any" distribution, and I should be more honest about that "almost." The theorem requires two things: independence and finite variance. When either condition fails, the bell curve doesn't just arrive a little late — it never shows up at all.

The Cauchy Distribution: A Cautionary Tale

The Cauchy distribution looks vaguely bell-shaped — it's symmetric and peaks at its center. But its tails are so fat that it has no mean and no variance. (The integrals diverge.) If you take the average of a million Cauchy random variables, the distribution of that average looks... exactly like a single Cauchy. No convergence. No smoothing. The CLT simply doesn't apply.7 (For more on heavy-tailed distributions and why they cause so much trouble, see Chapter 39: Power Laws.)

The independence condition is just as important, and more commonly violated. If your random variables are correlated — if one being large makes the next one more likely to be large — then the CLT may fail. This is what makes financial markets so treacherous. Stock returns on normal days look Gaussian, and if you assume they're independent, you conclude that catastrophic crashes are essentially impossible. But returns are not independent. Fear begets fear; selling triggers more selling. The resulting distributions have much fatter tails than a Gaussian predicts, which is why "once-in-a-century" market events seem to happen every decade or so.8

This is the dark side of the CLT's success. Because the bell curve shows up so reliably in so many contexts, people start assuming it shows up everywhere. They see a distribution that's roughly bell-shaped and conclude it must be Gaussian, without checking whether the conditions actually hold. And then they're shocked when the tails are fatter than expected — when the "impossible" event happens.

The theorem doesn't say everything is Gaussian. It says sums of independent, finite-variance random variables are approximately Gaussian. Those qualifiers are doing a lot of work.

• • •
Chapter 63

Why This Matters Outside of Math Class

The Central Limit Theorem is the invisible engine behind an enormous amount of modern life. Here are three examples:

Why polls work. When Gallup surveys 1,000 randomly selected Americans about their political views, each response is a random variable. The sample proportion is an average of those random variables. By the CLT, that average is approximately normally distributed around the true population proportion, with a standard deviation of roughly σ/√n. For a yes-no question (where σ ≤ 0.5), that gives a margin of error of about 1/√1000 ≈ 3%. That's why every poll you see reports a "±3% margin of error" — it's the CLT at work, quantifying exactly how much randomness remains after averaging over a thousand people.

Why manufacturing works. If a factory produces bolts whose lengths vary slightly due to dozens of small, independent sources of variation — temperature fluctuations, tool wear, material inconsistencies — then the total variation in bolt length will be approximately Gaussian. Engineers can use this to set tolerances, predict defect rates, and design quality control systems. The entire framework of Six Sigma is built on the assumption that manufacturing variation is normally distributed, which is really just the assumption that the CLT applies.

Why measurement errors are normally distributed. When you measure something — anything — the error is the sum of many small, independent sources of imprecision. The angle of your eye, the calibration of the instrument, the vibration of the table, the temperature of the room. Each contributes a tiny random error, and their sum, by the CLT, is approximately normal. This is why the normal distribution was historically called the "law of errors" — it's the natural distribution for the accumulation of many small mistakes.

"Galton's 'supreme law of Unreason' is really the supreme law of aggregation. Alone, we are unpredictable. Together, we ring like a bell."

The Central Limit Theorem tells us something profound about the nature of randomness: most of the randomness in the world is the additive kind, built from many small independent contributions. And additive randomness, aggregated, always converges to the same shape. The bell curve isn't imposed from outside — it emerges from within, an inevitable consequence of addition and independence.

Unless, of course, the variance is infinite. Or the variables aren't independent. In which case, all bets are off. And that, perhaps, is the most important lesson: the CLT is powerful precisely because it tells you both when the bell curve applies and when it doesn't. To know the conditions of a theorem is to know the boundaries of the world it describes.

Galton's quincunx still sits in museums, its lead balls still tumbling through pegs, still piling into that inevitable curve. Drop a single ball and you have no idea where it will land. Drop a thousand and you know exactly what shape they'll make. That's the Central Limit Theorem: the certainty that emerges from uncertainty, the order that hides inside chaos, the bell that rings every time you add enough things up.

Notes & References

  1. Francis Galton, Natural Inheritance (London: Macmillan, 1889), p. 63. Galton's original quincunx is preserved at the Galton Laboratory, University College London.
  2. The convergence rate depends on the shape of the original distribution. Symmetric distributions converge faster; highly skewed ones (like the exponential) need larger n. The Berry-Esseen theorem quantifies this: the error is O(1/√n).
  3. Abraham de Moivre, The Doctrine of Chances, 2nd edition (London, 1738). The normal approximation to the binomial first appeared in a 1733 supplement to the first edition.
  4. Pierre-Simon Laplace, Théorie analytique des probabilités (Paris: Courcier, 1812). Laplace's proof was the first to extend beyond the binomial case.
  5. J. W. Lindeberg, "Eine neue Herleitung des Exponentialgesetzes in der Wahrscheinlichkeitsrechnung," Mathematische Zeitschrift 15 (1922), pp. 211–225. Lévy independently proved similar results in his 1925 Calcul des probabilités.
  6. The characteristic function argument is sometimes called the "moment generating function proof." See Rick Durrett, Probability: Theory and Examples, 5th edition (Cambridge University Press, 2019), Chapter 3.
  7. The Cauchy distribution is the ratio of two independent standard normals, or equivalently, a Student's t-distribution with 1 degree of freedom. Its expected value does not exist because ∫|x|/(π(1+x²))dx diverges.
  8. Benoît Mandelbrot first documented the fat-tailed nature of financial returns in "The Variation of Certain Speculative Prices," The Journal of Business 36, no. 4 (1963), pp. 394–419. Nassim Taleb popularized the implications in The Black Swan (Random House, 2007).