Skip to content

S2 Chapter 3: Approximations and the Central Limit Theorem

S2 Statistics: Chapter 3 — Approximations and the Central Limit Theorem

Section titled “S2 Statistics: Chapter 3 — Approximations and the Central Limit Theorem”

Preface: The Quest for Computational Simplicity

Section titled “Preface: The Quest for Computational Simplicity”

Welcome, mathematical problem-solvers! Today we embark on a fascinating journey that bridges the gap between mathematical precision and practical computation. We’ll discover how some of history’s greatest mathematicians overcame seemingly insurmountable computational challenges through the elegant art of approximation.

Our story begins in an era before computers, when calculating even simple probabilities could take hours or days of tedious arithmetic. The question that drove mathematical innovation was simple yet profound: How can we make the impossible, possible?

1. The Computational Crisis: When Precision Becomes Impractical

Section titled “1. The Computational Crisis: When Precision Becomes Impractical”

Setting the Stage: The 19th Century Dilemma

Section titled “Setting the Stage: The 19th Century Dilemma”

Imagine you’re a 19th-century insurance actuary, tasked with calculating risk probabilities to set fair premiums. You need to compute probabilities from distributions like B(1000,0.01)B(1000, 0.01) or Po(25)\text{Po}(25).

2. Poisson Approximation: The Lightweight Solution

Section titled “2. Poisson Approximation: The Lightweight Solution”

From our study of the Poisson distribution, we know that:

Theorem (Poisson Limit of Binomial): As nn \to \infty and p0p \to 0 while maintaining np=λnp = \lambda (constant), we have:

B(n,p)Po(λ)B(n,p) \to \text{Po}(\lambda)

Example 1 (Quality Control Application):

A factory produces 1000 components per day with a defect rate of 0.005. What’s the probability of exactly 5 defective components?

Exact Calculation: P(X=5)P(X = 5) where XB(1000,0.005)X \sim B(1000, 0.005)

P(X=5)=(10005)×(0.005)5×(0.995)995P(X = 5) = \binom{1000}{5} \times (0.005)^5 \times (0.995)^{995}

This is computationally intensive!

Poisson Approximation: Since n=1000n = 1000 is large, p=0.005p = 0.005 is small, and np=510np = 5 \leq 10, we can use:

XPo(5) approximatelyX \sim \text{Po}(5) \text{ approximately} P(X=5)=e5×555!=e5×31251200.1755P(X = 5) = \frac{e^{-5} \times 5^5}{5!} = \frac{e^{-5} \times 3125}{120} \approx 0.1755

Much simpler to calculate!

Example 2 (Exercise):

In a certain region, 95%95\% of the population has lactose intolerance. A medical study randomly selects 8080 individuals from this population. Let XX represent the number of individuals WITHOUT lactose intolerance.

  1. Write down an expression for the exact value of P(X2)P(X \leq 2)
  2. Explain why the Poisson approximation is suitable for this problem.
  3. Using a Poisson approximation, estimate P(X2)P(X \leq 2)

Solution:

3. Normal Approximation: Breaking the Distribution Barriers

Section titled “3. Normal Approximation: Breaking the Distribution Barriers”

Normal Approximation to Binomial Distribution

Section titled “Normal Approximation to Binomial Distribution”

Theorem (De Moivre-Laplace Theorem): If XB(n,p)X \sim B(n,p) where nn is large and pp is not too close to 0 or 1, then:

XapproxN(μ,σ2)X \stackrel{\text{approx}}{\sim} N(\mu, \sigma^2)

where μ=np\mu = np and σ2=np(1p)\sigma^2 = np(1-p).

Rule of Thumb: Use when np>5np > 5 and n(1p)>5n(1-p) > 5.

Visual Demonstration: Normal Approximation Quality

Section titled “Visual Demonstration: Normal Approximation Quality”

Binomial Distribution B(50, 0.3) with Normal Overlay

Imagine a bar chart of B(50, 0.3) with a red normal curve N(15, 10.5) overlaid — the match is remarkably close.

Mean = 15, Variance = 10.5

Normal Approximation to Poisson Distribution

Section titled “Normal Approximation to Poisson Distribution”

Just as the binomial distribution approaches normality, so does the Poisson distribution for large parameters.

Theorem (Normal Approximation to Poisson): If XPo(λ)X \sim \text{Po}(\lambda) where λ\lambda is large (typically λ>10\lambda > 10), then:

XapproxN(λ,λ)X \stackrel{\text{approx}}{\sim} N(\lambda, \lambda)

Note the beautiful property: for Poisson distributions, the mean equals the variance!

Poisson Distribution Po(12) with Normal Overlay

Imagine a bar chart of Po(12) with a red normal curve N(12, 12) overlaid — again, a very close fit.

Mean = Variance = 12

4. Continuity Correction: The Bridge Between Discrete and Continuous

Section titled “4. Continuity Correction: The Bridge Between Discrete and Continuous”

When we approximate a discrete distribution with a continuous one, we face a conceptual problem:

Theorem (Continuity Correction): When approximating a discrete distribution with a continuous distribution, use these transformations:

DiscreteContinuous Approximation
P(X=a)P(X = a)P(a0.5<Y<a+0.5)P(a - 0.5 < Y < a + 0.5)
P(Xa)P(X \leq a)P(Y<a+0.5)P(Y < a + 0.5)
P(X<a)P(X < a)P(Y<a0.5)P(Y < a - 0.5)
P(Xa)P(X \geq a)P(Y>a0.5)P(Y > a - 0.5)
P(X>a)P(X > a)P(Y>a+0.5)P(Y > a + 0.5)

Example 3 (Continuity Correction in Practice):

A binomial random variable XB(100,0.3)X \sim B(100, 0.3) is approximated by YN(30,21)Y \sim N(30, 21). Find P(X=25)P(X = 25).

Without Continuity Correction: P(Y=25)=0P(Y = 25) = 0 (meaningless!)

With Continuity Correction:

P(X=25)P(24.5<Y<25.5)P(X = 25) \approx P(24.5 < Y < 25.5)

This gives a meaningful approximation that accounts for the discrete nature of the original distribution.

Example 4:

For each scenario, determine the most appropriate approximation method:

  1. XB(50,0.02)X \sim B(50, 0.02), find P(X=2)P(X = 2)
  2. XB(200,0.4)X \sim B(200, 0.4), find P(180X190)P(180 \leq X \leq 190)
  3. XPo(15)X \sim \text{Po}(15), find P(X>20)P(X > 20)

Solution:

Example 5:

A confectionery company produces chocolate bars, and during a special promotion, they place a golden ticket in 20%20\% of the bars. A convenience store receives a shipment of 6060 chocolate bars.

    1. Write down a suitable distribution to model the number of bars containing golden tickets.
    2. State one assumption required for this model to be valid.
  1. Find the probability that exactly 1515 bars contain golden tickets.
  2. Using a normal approximation with continuity correction, estimate the probability that fewer than 1010 bars contain golden tickets.
  3. The store manager wants to be 90%90\% confident that at least 88 customers will find golden tickets. Is this shipment size sufficient? Show your working.

Solution:

Exercise 1 (WST02/01/Jan15/7):

A multiple choice examination paper has nn questions where n>30n > 30.

Each question has 55 answers of which only 11 is correct. A pass on the paper is obtained by answering 3030 or more questions correctly.

The probability of obtaining a pass by randomly guessing the answer to each question should not exceed 0.02280.0228.

Use a normal approximation to work out the greatest number of questions that could be used.


Exercise 2 (WST02/01/Jan16/3):

Left-handed people make up 10%10\% of a population. A random sample of 6060 people is taken from this population. The discrete random variable YY represents the number of left-handed people in the sample.

    1. Write down an expression for the exact value of P(Y1)P(Y \leq 1)
    2. Evaluate your expression, giving your answer to 3 significant figures.
  1. Using a Poisson approximation, estimate P(Y1)P(Y \leq 1)
  2. Using a normal approximation, estimate P(Y1)P(Y \leq 1)
  3. Give a reason why the Poisson approximation is a more suitable estimate of P(Y1)P(Y \leq 1)

Exercise 3 (WST02/01/Jan17/3):

  1. State the condition under which the normal distribution may be used as an approximation to the Poisson distribution.

The number of reported first aid incidents per week at an airport terminal has a Poisson distribution with mean 3.5.

  1. Find the modal number of reported first aid incidents in a randomly selected week. Justify your answer.

The random variable XX represents the number of reported first aid incidents at this airport terminal in the next 2 weeks.

  1. Find P(X>5)P(X > 5)
  2. Given that there were exactly 6 reported first aid incidents in a 2 week period, find the probability that exactly 4 were reported in the first week.
  3. Using a suitable approximation, find the probability that in the next 40 weeks there will be at least 120 reported first aid incidents.

Exercise 4 (WST02/01/June17/2):

Crispy-crisps produces packets of crisps. During a promotion, a prize is placed in 25%25\% of the packets. No more than 11 prize is placed in any packet. A box contains 66 packets of crisps.

    1. Write down a suitable distribution to model the number of prizes found in a box.
    2. Write down one assumption required for the model.
  1. Find the probability that in 22 randomly selected boxes, only 11 box contains exactly 11 prize.
  2. Find the probability that a randomly selected box contains at least 22 prizes.

Neha buys 8080 boxes of crisps.

  1. Using a normal approximation, find the probability that no more than 3030 of the boxes contain at least 22 prizes.

6. The Central Limit Theorem: The Ultimate Foundation

Section titled “6. The Central Limit Theorem: The Ultimate Foundation”

All our approximation methods point to a deeper truth — one of the most important theorems in all of mathematics:

Theorem (Central Limit Theorem): Let X1,X2,,XnX_1, X_2, \ldots, X_n be independent and identically distributed random variables with finite mean μ\mu and variance σ2\sigma^2.

As nn \to \infty, the sum Sn=X1+X2++XnS_n = X_1 + X_2 + \cdots + X_n approaches a normal distribution:

SnapproxN(nμ,nσ2)S_n \stackrel{\text{approx}}{\sim} N(n\mu, n\sigma^2)

Equivalently, the standardized sum:

Zn=SnnμσnapproxN(0,1)Z_n = \frac{S_n - n\mu}{\sigma\sqrt{n}} \stackrel{\text{approx}}{\sim} N(0,1)

Discovering CLT Through Dice: A Visual Journey

Section titled “Discovering CLT Through Dice: A Visual Journey”

Let’s see the magic of CLT in action using the simple example of rolling dice.

Example 6 (Rolling Dice and Finding Normality):

Consider rolling a fair six-sided die. The outcome XX has:

  • Uniform distribution: P(X=k)=16P(X = k) = \frac{1}{6} for k=1,2,3,4,5,6k = 1, 2, 3, 4, 5, 6
  • Mean: E(X)=3.5E(X) = 3.5, Variance: Var(X)=35122.92\text{Var}(X) = \frac{35}{12} \approx 2.92

Exercise: Calculate the Distribution for n = 2

Now let’s roll two independent dice and find the sum S2=X1+X2S_2 = X_1 + X_2:

  1. Complete the table below to find all possible outcomes and their sums:
X1\X2X_1 \backslash X_2123456
1
2
3
4
5
6
  1. Count the frequency of each sum and complete the probability distribution:
Sum S2S_223456789101112
Frequency
P(S2=k)P(S_2 = k)
  1. Calculate E(S2)E(S_2) and Var(S2)\text{Var}(S_2) using the properties of sums:

E(S2)=E(X1)+E(X2)=E(S_2) = E(X_1) + E(X_2) = ______

Var(S2)=Var(X1)+Var(X2)=\text{Var}(S_2) = \text{Var}(X_1) + \text{Var}(X_2) = ______

Key Observation: Even with just two dice, we can see the distribution shifting from uniform (flat) to triangular (peaked). This is the beginning of the journey toward the normal distribution!

The Visual Journey of CLT:

Imagine three histograms side by side:

  • n = 1: Perfectly uniform — flat and rectangular, Mean = 3.5
  • n = 2: Triangular shape — the peak emerges, Mean = 7.0
  • n large: Beautiful bell curve — exactly the normal distribution, Mean = 3.5n

Now we can understand why our approximation methods work so well:

Example 7 (Why Binomial Becomes Normal):

Recall that if XBinomial(n,p)X \sim \text{Binomial}(n,p), then XX can be written as:

X=Y1+Y2++YnX = Y_1 + Y_2 + \cdots + Y_n

where each YiBernoulli(p)Y_i \sim \text{Bernoulli}(p) is independent.

By CLT, as nn gets large:

X=i=1nYiapproxN(np,np(1p))X = \sum_{i=1}^{n} Y_i \stackrel{\text{approx}}{\sim} N(n \cdot p, n \cdot p(1-p))

This is exactly the De Moivre-Laplace theorem we used earlier!

Example 8 (Why Poisson Becomes Normal):

Similarly, if XPoisson(λ)X \sim \text{Poisson}(\lambda), we can express XX as the sum of many small independent Poisson variables.

For large λ\lambda, we can write XX as the sum of nn independent Poisson(λ/n)\text{Poisson}(\lambda/n) variables. By CLT:

XapproxN(λ,λ)X \stackrel{\text{approx}}{\sim} N(\lambda, \lambda)

This explains our normal approximation to Poisson distribution!

Example 9 (CLT in Action: Quality Control):

A factory produces items where the final weight is affected by:

  • Raw material variations
  • Machine calibration drift
  • Temperature fluctuations
  • Operator differences
  • Measurement errors
  • … and many other small factors

Even if each individual factor has a completely different distribution, the total effect (sum of all factors) will be approximately normal by CLT.

This is why quality control charts always assume normal distributions!

Looking Forward: The Mathematical Foundation

Section titled “Looking Forward: The Mathematical Foundation”

The Mathematical Beauty: The Central Limit Theorem reveals a fundamental harmony in randomness — no matter how chaotic individual components might be, their collective behavior gravitates toward the same universal pattern: the normal distribution serves as nature’s “attractor” for randomness.