The Law of Large Numbers (LLN) and the Central Limit Theorem (CLT) are fundamental concepts in probability and statistics.

1 Law of Large Numbers (LLN):

Definition: The Law of Large Numbers states that as the sample size increases (or as the number of trials in an experiment increases), the sample mean (or what) converges in probability to the population mean. In simpler terms, as you take more observations from a population, the average of those observations tends to get closer to the true average of the entire population.

Key Point: LLN is concerned with the behavior of sample averages as the sample size grows, and it emphasizes the stability and reliability of sample statistics as estimates of population parameters.

It doesn’t describe the distribution of individual samples; rather, it focuses on the convergence properties of the sample mean to the population mean.

2 Central Limit Theorem (CLT):

Definition: The Central Limit Theorem states that, regardless of the distribution of the population, the sampling distribution of the sample mean1 (or any sufficiently large sample statistic) will be approximately normally distributed. This holds true as long as the sample size is large enough.

Key Point: CLT is concerned with the shape of the distribution of the sample mean, suggesting that it becomes normal as the sample size increases, irrespective of the original distribution of the population.

2.0.1 Alternative Definition:

If you have a sufficiently large sample size drawn from a population with any shape of distribution (not necessarily normal), the distribution of the sample mean will be approximately normally distributed.

Specifically, the Central Limit Theorem states the following:

Let \(X_1​,X_2​,...,X_n\)​ be a sequence of independent and identically distributed (i.i.d.) random variables with a mean \(\mu\) and a finite standard deviation \(\sigma\). If \(n\) is sufficiently large, then the distribution of the sample mean \(\bar X\) (the average of the sample) will be approximately normally distributed with a mean \(\mu\) and a standard deviation \(\dfrac{\sigma}{\sqrt n}\).

Mathematically, this can be expressed as:

\(\bar X \sim N(\mu,\dfrac{\sigma}{\sqrt n}​)\)

This is a powerful result because it allows statisticians to make inferences about population parameters using the normal distribution, even when the underlying population distribution is not normal. The larger the sample size (n), the closer the distribution of the sample mean will be to a normal distribution, regardless of the original distribution of the population.

3 Similarity & Differences

3.1 Similarity:

  • Both LLN and CLT are concerned with the behavior of sample statistics, particularly the sample mean, as sample size increases.

  • Both are fundamental concepts in statistics and play crucial roles in inferential statistics.

3.2 Difference:

  1. Focus:

    • LLN: Focuses on the behavior of the sample mean as the sample size increases and how it converges to the population mean.

    • CLT: Focuses on the distribution of the sample mean, stating that it becomes approximately normal with a large sample size, regardless of the population distribution.

  2. Assumption:

    • LLN: Assumes the existence of the population mean and investigates the behavior of the sample mean as the sample size increases.

    • CLT: Does not necessarily assume the existence of a population mean; it deals with the distribution of the sample mean.

In summary, LLN deals with the convergence of sample statistics to population parameters, while CLT deals with the distributional properties of sample statistics, particularly the sample mean.

4 Central Limit Theorem (CLT) subsumes the Law of Large Numbers (LLN).

In other words, the CLT includes the LLN as a special case.

The Law of Large Numbers (LLN) is a fundamental concept that describes the behavior of sample averages as the sample size increases, stating that the sample mean converges in probability to the population mean. The focus of LLN is on the convergence of sample statistics to population parameters.

The Central Limit Theorem (CLT) is a more powerful and general result. It states that, under certain conditions, the distribution of the sum (or average) of a large number of independent, identically distributed random variables approaches a normal distribution, regardless of the shape of the original population distribution. This implies that the sample mean, for a sufficiently large sample size, will be approximately normally distributed.

Here’s the relationship:

  • LLN is a specific case or application of CLT when dealing with the sample mean. As the sample size becomes large, the CLT implies that the distribution of the sample mean becomes approximately normal, which aligns with the principles of the LLN.

In essence, the LLN can be viewed as a special case of the more general CLT when applied to the sample mean. Therefore, while LLN provides insights into the convergence of sample means, CLT extends these insights to describe the distributional properties of sample means as the sample size becomes large.

5


  1. The sampling distribution of the sample mean is a theoretical distribution that describes the possible values that the sample mean can take if you were to draw multiple random samples from a population and calculate the mean for each sample. This distribution is a key concept in statistics and is often used in hypothesis testing and constructing confidence intervals. In practice, if the population standard deviation (\(\sigma\)) is unknown, it can be estimated by the sample standard deviation (s),↩︎