Central Limit Theorem (CLT)

The Central Limit Theorem states that the distribution of the sample mean can be approximated by a normal distribution although the original population may be non-normal.

The grand average, resulting from averaging sets of samples or the average of the averages, approaches the universe mean as the number of sample sets approaches infinity.

For a given population standard deviation, as the sample size (n) increases, the Standard Error of the Mean (SE) decreases. SE is the standard deviation of the sampling distribution.

See the formula for calculating the Standard Error of the Mean below:

This formula implies that as the denominator increases (which is the sample size), then the SE of the Mean decreases. 

Thinking about this practically, if you get more and more samples, the risk of error to the population should get smaller since you are getting closer and closer to actually using the entire population. Every one sample size more, gets you that much closer to the population. 

But this is not linear. Plug in a few values and you will see the effect of diminishing returns up to a point. Using simple number to illustrate the concept, if the population standard deviation is 1.0 and the sample size (n) = 1.0, then the SE Mean = 1.0.

If n = 9, SE Mean = 0.333

If n = 64, SE Mean = 0.125

If n = 10,000, SE Mean = 0.01

If n = 1,000,000, SE Mean = 0.001

In other words, as the sample size approaches infinity, the SE Mean approaches zero. However, in reality it is rare to practically obtain these large samples and therefore some error will exists. 

Large samples sizes:

- generate a better estimation of the population since the sampling error in minimized. 

- distribution curve gets narrower

- the average difference from the statistic to the parameter decreases. The values of x-bar will have less variation and get closer to the population mean, m.

CLT says that as n increases the SE Mean goes towards zero (as shown above) and the distribution of sample means will approach a normal distribution.

Sampling distributions tend to become normal distributions when sample sizes are large enough which is generally at n > 30 for unknown distributions. 


How is the SE related to the ME?

Click here to learn about the difference between the Standard Error of the Mean and the Margin of Error. Once you understand the Margin of Error, then the calculation for Confidence Intervals becomes simpler.  



Return to the MEASURE phase

Return to BASIC STATISTICS

Templates and Calculators

Find a career in Six Sigma - search active job postings

Return to the Six-Sigma-Material Home Page


Site Membership
CLICK HERE


Six Sigma
Online
Courses

Six Sigma Certification Courses

Six Sigma

Templates & Calculators


Six Sigma Modules

CLICK HERE

Green Belt Program (1,000+ Slides)

Basic Statistics

SPC

Process Mapping

Capability Studies

MSA

Cause & Effect Matrix

FMEA

Multivariate Analysis

Central Limit Theorem

Confidence Intervals

Hypothesis Testing

T Tests

1-Way Anova Test

Chi-Square Test

Correlation and Regression

Control Plan

Kaizen

Error Proofing


Six Sigma
Quick Reference Guide