ConceptsFundamentals → Normal approximation

When you can use a normal approximation — and when it'll lie to you.

Add enough independent dice together and the distribution starts to look like a bell curve. That's the central limit theorem, visible in 1d6 → 2d6 → 4d6 → 8d6. It is also the most over-applied piece of probability folklore in dice math — because most actually-interesting dice mechanics break the assumptions that make the bell curve work.

Where it works: sums of fair dice

The central limit theorem says: if you sum a large number of independent random variables with finite variance, the distribution of the sum, suitably standardised, approaches a normal (Gaussian) distribution. "Suitably standardised" means subtracting the mean and dividing by the standard deviation, so you end up looking at how many SDs from the mean a given outcome sits.

Visually, this is what's happening at every step from 1d6 to 2d6 to 4d6:

1d6
min 1 max 6 mean 3.50 7/2
P(finish at 4 HP) = 50.00% (1/2)
  • 1 16.67%
  • 2 16.67%
  • 3 16.67%
  • 4 16.67%
  • 5 16.67%
  • 6 16.67%
2d6
min 2 max 12 mean 7.00 7
P(finish at 7 HP) = 58.33% (7/12)
  • 2 2.78%
  • 3 5.56%
  • 4 8.33%
  • 5 11.11%
  • 6 13.89%
  • 7 16.67%
  • 8 13.89%
  • 9 11.11%
  • 10 8.33%
  • 11 5.56%
  • 12 2.78%
4d6
min 4 max 24 mean 14.00 14
P(finish at 14 HP) = 55.63% (721/1296)
  • 4 0.08%
  • 5 0.31%
  • 6 0.77%
  • 7 1.54%
  • 8 2.70%
  • 9 4.32%
  • 10 6.17%
  • 11 8.02%
  • 12 9.65%
  • 13 10.80%
  • 14 11.27%
  • 15 10.80%
  • 16 9.65%
  • 17 8.02%
  • 18 6.17%
  • 19 4.32%
  • 20 2.70%
  • 21 1.54%
  • 22 0.77%
  • 23 0.31%
  • 24 0.08%
8d6
min 8 max 48 mean 28.00 28
P(finish at 28 HP) = 54.05%
  • 8 0.00%
  • 9 0.00%
  • 10 0.00%
  • 11 0.01%
  • 12 0.02%
  • 13 0.05%
  • 14 0.10%
  • 15 0.20%
  • 16 0.37%
  • 17 0.62%
  • 18 1.00%
  • 19 1.52%
  • 20 2.18%
  • 21 2.99%
  • 22 3.92%
  • 23 4.90%
  • 24 5.88%
  • 25 6.77%
  • 26 7.48%
  • 27 7.94%
  • 28 8.09%
  • 29 7.94%
  • 30 7.48%
  • 31 6.77%
  • 32 5.88%
  • 33 4.90%
  • 34 3.92%
  • 35 2.99%
  • 36 2.18%
  • 37 1.52%
  • 38 1.00%
  • 39 0.62%
  • 40 0.37%
  • 41 0.20%
  • 42 0.10%
  • 43 0.05%
  • 44 0.02%
  • 45 0.01%
  • 46 0.00%
  • 47 0.00%
  • 48 0.00%

1d6 is flat. 2d6 is a tent. 4d6 is a clearly mound-shaped histogram. 8d6 is, to the eye, a bell curve. By the time you're summing a fistful of dice, the normal approximation is good enough for most practical purposes — within a percentage point or two on most queries.

With mean μ and standard deviation σ in hand, the normal approximation gives you a quick rule of thumb:

  • About 68% of outcomes within ±1σ of μ.
  • About 95% within ±2σ.
  • About 99.7% within ±3σ.

For 8d6: μ = 28, σ = √(70/3) ≈ 4.83. So roughly 68% of rolls land in [23, 33], 95% in [18, 38]. Compare to the panel: that's close to what the bars actually say. Not identical — but close.

Where it fails, and how loudly

Four common dice situations where the normal approximation gives you the wrong answer in ways that actually matter:

1. Few dice

1d20+5 is a single die plus a constant: a uniform distribution shifted right. Approximating a uniform distribution as a normal is visibly wrong — the tails of a normal stretch to infinity, while a uniform truncates abruptly. The central limit theorem needs enough dice for the bell shape to emerge; one die is not enough, and even 2d6 is still tent-shaped, not bell-shaped.

2. Heavy-tailed dice (exploding, fat-tail mechanics)

Exploding dice (1d6! — on a 6, roll again and add) have a long right tail with non-zero probability extending far past the original face count. The normal approximation systematically under-estimates the probability of those far-right outcomes, because the bell curve falls off too fast compared to a fat tail.

See exploding dice and fat tails for the worked-out version. Short version: don't normal-approximate exploding dice.

3. Asymmetric mechanics (advantage, disadvantage, kept-best-of)

2d20kh1 (advantage) is the maximum of two 1d20 rolls — an order statistic, not a sum. Order statistics are skewed, especially at the extremes. The distribution of kh1 bunches toward the high end; kl1 bunches toward the low end. Neither looks symmetric. A normal approximation, which is symmetric by construction, smooths over the skew that's the whole point of the mechanic.

4. Small mean, hard floor

A roll like 2d4 has mean 5 and SD ≈ 1.58. The normal approximation says ~16% of outcomes are below 3.4 (one SD down), but the actual minimum is 2 — and the distribution sits in a very narrow band. When μ − 3σ < min, you've got "the bell curve says probability extends below where the dice physically can land", which is always wrong by exactly whatever probability the bell curve assigns to the impossible region.

The rule of thumb

Use the normal approximation as a sanity check, not as the answer. For the four-or-more-dice flat sum case it's accurate to within a percentage point, which is fine for napkin work. For everything else the engine on this site computes the exact probability in microseconds — there is no reason to approximate.

A pragmatic test: compare the closed-form normal answer to the panel for the same expression on this site. If they agree within ~1% across the range you care about, the approximation is fine. If they don't, trust the engine — the rationals can't be wrong because they're not rounded.

Try it yourself

That closes the fundamentals track. From here, the counterintuitive results pillars build on every concept you now have: variance and kill probability uses both means and variances against thresholds; reliable-vs- nuke compares two distributions of equal mean and opposite shapes; expected strikes to kill is what happens when you sum many swings until a target falls.