Graduate

GraduateProbability and StatisticsProbability Theory


Central Limit Theorem


The Central Limit Theorem (CLT) is one of the cornerstone concepts in probability theory and statistics. It explains why many distributions approximate the normal distribution, and it serves as the basis for many statistical procedures. In this comprehensive explanation, we will explore what the Central Limit Theorem is, why it is important, and how it works with various examples and visualizations.

Understanding the central limit theorem

At its core, the central limit theorem states that the distribution of the sum (or average) of a large number of independent and identically distributed (iid) random variables is approximately normal, regardless of the original distribution from which these variables are drawn, provided that the original distribution has a finite variance. This is a profound result because it implies that even if you start with a skewed or uniform distribution, if you take enough samples and find their average, the distribution of these averages will tend to a normal distribution.

P(X_1, X_2, ..., X_n) to N(mu, sigma^2/n)

Here, X_1, X_2, ..., X_n are samples from the population, mu is the mean of the population, and sigma^2 is the variance. N denotes the normal distribution.

Why is the central limit theorem important?

The central limit theorem is important because it justifies the use of the normal distribution in many scenarios where other models are more complex or less understood. Some of the key areas where the CLT is important include:

  • Statistical inference: This allows the use of confidence intervals and hypothesis tests that assume normally distributed data, even if the underlying data distribution is not normal.
  • Data science and machine learning: Many algorithms assume normality due to CLT, making models more robust.
  • Quality Control: CLT is helpful in quality control processes where sampling means are monitored.

Illustrating the central limit theorem through examples

Example 1: Throwing a dice

Consider throwing a fair six-sided dice. The outcome of throwing a dice is a discrete uniform distribution from 1 to 6. Each integer value from 1 to 6 has an equal probability of 1/6. This distribution is certainly not normal.

0.5 1.0 1 2 3 4 5 6

Now, imagine rolling two dice 1000 times and taking the average of both dice each time. Each roll of the two dice will be independent, and we will have 1000 averages. When you plot these averages, you will see that the shape of their distribution begins to resemble a bell curve. Now, if you increase the number of dice to 3, 4 and beyond, the distribution of these averages will continue to resemble a normal distribution better.

Example 2: Simulating a coin toss

Take 100 coins and toss them. Each coin toss can be viewed as a Bernoulli trial with a probability of 0.5 for heads and 0.5 for tails. Assume '1' for heads and '0' for tails.

If we perform this experiment and measure the number of heads (successes), we can treat each toss as an independent variable. The central limit theorem shows that if we repeat this 100-coin toss many times and plot the number of heads each time, the distribution of these counts will approach a normal distribution.

50 Counting the ends

Mathematical proof of the central limit theorem

Let's take a look at the mathematics that rigorously proves the Central Limit Theorem. The theorem was developed independently by several mathematicians, including Abraham de Moivre, Pierre-Simon Laplace, and Carl Friedrich Gauss. Here, we present a simplified version of the proof:

Let X_1, X_2, ..., X_n be iid random variables with mean mu and variance sigma^2. The expectation is given by

E[X_i] = mu

and the variance is

Var(X_i) = sigma^2

Define the sample mean as

bar{X} = frac{X_1 + X_2 + ... + X_n}{n}

The expectation of the sample mean is

E[bar{X}] = Eleft[frac{X_1 + X_2 + ... + X_n}{n}right] = mu

and its variance is

Var(bar{X}) = frac{1}{n^2}(Var(X_1) + Var(X_2) + ... + Var(X_n)) = frac{sigma^2}{n}

According to the standard central limit theorem, if n is large enough, the standardized sample mean is approximately normally distributed, with mean 0 and variance 1:

Z = frac{bar{X} - mu}{sigma/sqrt{n}} to N(0,1)

Conclusion

The central limit theorem is a powerful statistical principle that bridges the gap between different types of distributions and the normal distribution. Its versatility and reliability make it an important tool for statistical inference, justifying many methods and theories in both theoretical and applied statistics.

Whether throwing dice, tossing coins, or taking measurements in the real world, this theorem enables us to make informed statistical analysis and predictions. By understanding the CLT, you are better equipped to tackle a variety of statistical challenges across a variety of disciplines.


Graduate → 5.1.4


U
username
0%
completed in Graduate


Comments