Central Limit Theorem Calculator

Online calculator for computations related to the Central Limit Theorem

CLT Calculator

The Central Limit Theorem

The Central Limit Theorem (CLT) states that sample means, regardless of the original distribution, are approximately normally distributed.

Enter parameters
Population standard deviation
Number of samples (recommended n ≥30)
CLT Results
Standard error (σ̄):
Standard deviation of the sample means
CLT Properties

Core idea: For large samples (n ≥30) the distribution of sample means approaches a normal distribution

n ≥30 μ̄ = μ σ̄ = σ/√n

CLT Concept

Regardless of the original distribution, sample means become normally distributed.
The standard deviation decreases by a factor of √n.

Original distribution n ≥ 30 Sample means Arbitrary shape Normal distribution

Original distribution (any shape)
Sample means (approximately normal)


What is the Central Limit Theorem?

The Central Limit Theorem (CLT) is one of the most important theorems in statistics:

  • Statement: Sample means approach a normal distribution for large samples
  • Condition: Sample size n ≥30 (rule of thumb)
  • Generality: Holds independently of the original distribution
  • Application: Inferential statistics, confidence intervals, hypothesis testing
  • Importance: Enables statistical inference even for unknown distributions
  • Related: Law of Large Numbers, Normal distribution

The three statements of the CLT

The Central Limit Theorem makes three fundamental statements about sample means:

1. Expectation
\[\mathbb{E}[\overline{X}] = \mu\]

The expectation of sample means equals the population mean

2. Standard error
\[\text{Var}(\overline{X}) = \frac{\sigma^2}{n}\]

The variance of sample means is reduced by factor n

3. Normality
\[\overline{X} \sim \mathcal{N}\left(\mu, \frac{\sigma}{\sqrt{n}}\right)\]

The distribution of sample means is asymptotically normal

Applications of the Central Limit Theorem

The CLT is the theoretical basis for many statistical procedures:

Inferential statistics
  • Confidence intervals for means
  • Hypothesis tests (t-tests, z-tests)
  • Parameter estimation for large samples
  • Significance testing and p-values
Quality control
  • Statistical process control (SPC)
  • Control charts and tolerance limits
  • Sampling inspection in production
  • Six Sigma and process improvement
Market research & surveys
  • Opinion polls and election forecasting
  • Market share and preference analysis
  • A/B testing and experimental design
  • Sample size determination
Medicine & science
  • Clinical trials and drug testing
  • Epidemiological studies
  • Laboratory reference ranges
  • Biometric analyses

Formulas for the Central Limit Theorem

Expectation of sample means
\[\mu_{\overline{X}} = \mu\]

The mean of sample means equals the population mean

Standard error
\[\sigma_{\overline{X}} = \frac{\sigma}{\sqrt{n}}\]

Standard deviation of the sample means (standard error)

Normal distribution of sample means
\[\overline{X} \sim \mathcal{N}\left(\mu, \frac{\sigma^2}{n}\right) \quad \text{or} \quad \overline{X} \sim \mathcal{N}\left(\mu, \left(\frac{\sigma}{\sqrt{n}}\right)^2\right)\]

Asymptotic normality for large samples

Standardization (z-transformation)
\[Z = \frac{\overline{X} - \mu}{\sigma/\sqrt{n}}\]

Standard normal distribution N(0,1) for hypothesis testing

Confidence interval
\[\overline{x} \pm z_{\alpha/2} \cdot \frac{\sigma}{\sqrt{n}}\]

Confidence interval for the population mean

Finite population (finite population correction)
\[\sigma_{\overline{X}} = \frac{\sigma}{\sqrt{n}} \cdot \sqrt{\frac{N-n}{N-1}}\]

Correction for finite populations with N elements

Example calculations for the Central Limit Theorem

Example1: Production quality control
Machine part lengths, σ =0.5 mm, n =36
Given
  • Population standard deviation: σ =0.5 mm
  • Sample size: n =36 parts
  • Target mean: μ =100 mm
Sought: Standard error of the sample means
CLT calculation
\[\sigma_{\overline{X}} = \frac{\sigma}{\sqrt{n}} = \frac{0.5}{\sqrt{36}} = \frac{0.5}{6} =0.083 \text{ mm}\]
Interpretation: The standard deviation of sample means (0.083 mm) is much smaller than the individual measurement standard deviation (0.5 mm).95% of sample means lie between99.84 mm and100.16 mm.
Example2: Opinion poll
Election poll, p =0.4, n =100
Given
  • Population proportion: p =0.4 (40%)
  • Sample size: n =100 people
  • Binomial SD: σ = √(p(1-p)) = √(0.4×0.6) =0.49
Sought: Standard error of the sample proportion
Calculation
\[\sigma_{\hat{p}} = \sqrt{\frac{p(1-p)}{n}} = \sqrt{\frac{0.4 \times0.6}{100}}\] \[= \sqrt{\frac{0.24}{100}} = \sqrt{0.0024} =0.049\]
95% confidence interval:0.4 ±1.96 ×0.049 = [0.304,0.496] or30.4% to49.6%
Example3: Calculator defaults
σ =3, n =45
Direct calculation
\[\sigma_{\overline{X}} = \frac{\sigma}{\sqrt{n}} = \frac{3}{\sqrt{45}}\] \[= \frac{3}{6.708} =0.447\]
Improvement factor
Improvement factor:
\[\frac{\sigma}{\sigma_{\overline{X}}} = \sqrt{n} = \sqrt{45} =6.71\]
Accuracy improves by a factor of6.71
Comparison: Effect of sample size
Sample size n √n Standard error σ̄ (σ=3) Reduction vs σ
164.000.75075% less variability
255.000.60080% less variability
366.000.50083% less variability
456.710.44785% less variability
10010.000.30090% less variability
40020.000.15095% less variability
Conclusion: To double accuracy, sample size must be quadrupled.

Mathematical foundations of the Central Limit Theorem

The Central Limit Theorem is one of the cornerstones of probability theory and forms the theoretical basis of modern statistics. It explains why the normal distribution appears so often in nature and legitimizes many statistical methods.

Historical development

The development of the CLT spans several centuries:

  • Abraham de Moivre (1733): First version for binomial distributions
  • Pierre-Simon Laplace (1812): Generalization and early rigorous proofs
  • Aleksandr Lyapunov (1901): Modern form under general conditions
  • Paul Lévy (1925): Characteristic functions and weak convergence
  • William Feller (1950s): Systematic treatment in modern probability theory

Mathematical precision

The CLT makes a precise statement about asymptotic distribution:

Formal statement:
Let X₁, X₂, ..., Xₙ be independent, identically distributed random variables with E[Xᵢ] = μ and Var(Xᵢ) = σ² < ∞. Then: \[\frac{\sqrt{n}(\overline{X}_n - \mu)}{\sigma} \xrightarrow{d} \mathcal{N}(0,1)\]
Equivalently:
\[\overline{X}_n \xrightarrow{d} \mathcal{N}\left(\mu, \frac{\sigma^2}{n}\right)\]

Conditions and assumptions

The CLT holds under various conditions:

Classical conditions
  • Independence: The random variables must be independent
  • Identical distribution: Same mean and variance
  • Finite variance: σ² < ∞ is required
  • Large samples: n → ∞ for exact validity
Generalizations
  • Lyapunov CLT: For non-identical distributions
  • Lindeberg CLT: Weaker moment conditions
  • Martingale CLT: For dependent sequences
  • Multivariate CLT: For vectors of random variables

Rate of convergence

The Berry-Esseen inequality quantifies the convergence speed:

Berry-Esseen theorem

For the error of the normal approximation: \(|F_n(x) - \Phi(x)| \leq \frac{C \rho}{\sigma^3 \sqrt{n}}\) whereρ = E[|X₁ - μ|³] is the third central moment.

Practical implications

The approximation error is proportional to n⁻¹/². For skewed distributions (largeρ) larger samples are required for good approximation.

Related theorems

Law of Large Numbers

Describes convergence of the sample mean to the expectation. Weak LLN: convergence in probability. Strong LLN: almost sure convergence.

Delta method

Transfers the CLT to functions of sample means: If g is differentiable then √n(g(\overline{X}_n) - g(μ)) → N(0, [g'(μ)]²σ²).

Continuity correction

For discrete distributions the continuity correction improves the normal approximation: P(X ≤ k) ≈ Φ((k +0.5 - μ)/σ).

Local CLT

Describes convergence of probability densities, not only distribution functions.

Limits and exceptions

When the CLT does not apply
  • Infinite variance: Cauchy distribution, Pareto with α ≤2
  • Strong dependence: Slowly decaying autocorrelations
  • Extreme skewness: Very small samples with skewed distributions
  • Heavy tails: Stable distributions with α <2
Practical issues
  • Finite samples: n =30 is only a rule of thumb
  • Outliers: May slow convergence
  • Model misspecification: Wrong assumptions about the population distribution
  • Clustering: Violation of independence assumption

Modern developments

Bootstrap and resampling

Modern non-parametric methods partially avoid the need for the CLT via resampling techniques.

Robust statistics

Development of methods that work even when CLT assumptions are violated.

Summary

The Central Limit Theorem is the theoretical backbone of modern statistics. It explains not only why many natural phenomena are normally distributed but also justifies the use of the normal distribution in inferential statistics. Despite its universality, its limits and assumptions must be carefully considered in practical applications. In an era of big data and complex dependencies, understanding both the capabilities and limitations of the CLT is essential for responsible statistical analysis.

Is this page helpful?            
Thank you for your feedback!

Sorry about that

How can we improve it?

More Risk and Probability Functions

Birthday Paradox
Bayes Theorem
Central Limit Theorem