The z-score calculation is used in the MEASURE and CONTROL phases of a DMAIC Six Sigma project. The score converts a point of interest (x) in terms of standard deviations from the population mean. It allows the comparison of observations from different normal distributions. If the value of x is less than the mean, then the z-score is negative and vice versa. Other names for the z-score are z-values, normal scores, standardized variables.
Determine a baseline z-score in the MEASURE phase after the MSA has passed. This preliminary value provided in the project contract may need refinement, this exercise is done in the MEASURE phase to get an accurate starting point.
Another z-score is calculated in the CONTROL phase. Sometimes it is done at the end of the IMPROVE phase but either way it is the final score indicating the change (better or worse) relative to the baseline measurement.
The z-score is most often used in the z-test in the Student's t-test for a population whose parameters are actually known (not estimates). But since it is rare to know the entire population, the t-test is more commonly used.
A benchmark sigma (z-bench) score offers an fair comparison for products or processes. A benchmark z-score is often unknown at the onset of the Project Contract and is a critical component for a GB/BB to determine as early as possible in the DMAIC project path. Once the baseline is determined then the contract is updated and all key stakeholders and revisit the potential savings and impact of the project and possibly redefine the scope.
The GB/BB's job is to collect the necessary data once the MSA passes to calculate and convert this data to a sigma score.
A couple simple useful formulas are:
If your data represents a shorter time frame or a short term sample then it is important to consider applying the 1.5 sigma shift to indicate what the longer term performance would actually be as a baseline. Further below is an explanation on converting short term to long term z-scores.
Click here to be redirected to a presentation that shows how to calculate z-scores on TI graphing calculators.
Motorola is credited with concluding that processes drift over long periods of time. The term is known as Long-Term Dynamic Mean Variation or in simpler terms, the "fudge factor". This variation between the short term samples was determined to be approximately 1.5 sigma.
Long term indicators contain both special cause and common cause variation. Short term data does not contain special cause variation and usually has a higher process capability than the long term data. That implies that the z-score for a short term sample should usually be higher than the z-score from long term data.
Six Sigma projects usually report process capability in terms of short-term performance. The true long term performance is estimated to have some special cause and unexpected performance (since Motorola confirmed this after years or study) and therefore estimated to be 1.5 sigma lower than the short term z-score.
It is possible to estimate the long or short term performance based on that value. A GB/BB can apply the 1.5 sigma shift if converting to or from long term sigma or short term sigma (if you are a believer in this 1.5 sigma shift).
If the z-score was calculated from a sample then the long-term performance can be estimated by subtracting 1.5 sigma from your calculated value for short term sigma.
Recall that if you arrive at DPMO that is considered a long term process indicator but it may not be if you have populations normally in the billions or trillions. It is important to use the shift appropriately. Using enough process data a GB/BB can even estimate the shift for their own process.
If the value you came up with was from a population it is a long term representation of the process performance. Therefore, if interested in the projected short-term performance then you would add 1.5 from your calculated z-score.
LONG TERM ---> SHORT TERM: ADD 1.5 SIGMA
SHORT TERM ---> LONG TERM: SUBTRACT 1.5 SIGMA
If the point of interest (x) is located on the mean then the z-score is zero. This seems logical since the point lies directly on the mean with zero standard deviation from the mean.
If the point of interest (original measurement, x) is below the mean then the z-score will be negative. The z-score is positive when the x is a value higher or to the right of the mean.
This distribution is represented by a normal distribution with a mean of 0 and a standard deviation of 1. Since integral calculus is challenging to most of us, there are tables to help solve z-distribution and normal distribution problems. Click here for normal distribution table.
The empirical rule for a normal distribution says that about 99.73% of all x values are within 3 standard deviations of the mean which, in terms of a z-distribution, is represented by z = -3 to z = +3.
When a population is normally distributed, the percentile rank may be determined from the z-score and statistical tables. It is important to understand the table being used as websites or textbooks display the values in different ways.
This is most often the most challenging part to a newer GB/BB and common mistake made on exams and projects. Get comfortable with converting z-scores to percentile ranks, PPM's, DPMO, and using tables.
Given a normally distributed set of data we know the following:
The mean is known to be 52.0
The standard deviation is 4.5
Our point of interest (x) is 61.0
Since the point of interest is greater than the mean you can expect a positive z-value. Substituting the values into the formula at the top of the page:
z = (61-52) / 4.5 = 2
The value of 61 is 2 standard deviations away from the mean.
Using the assumptions of a normal distributed bell curve only:
Given a normally distributed set of data we know the following:
The mean length is known to be 510.7 mm
The standard deviation of the lengths is 1.49 mm
The customer specifies that the LSL is 507 mm and the USL is 513 mm.
What is the total % of parts that can be expected to be defective?
This is a two-part problem.
FIRST: Determine the % of parts >513 mm.
z = (513 mm - 510.7 mm) / 1.49 mm = 1.54
Referring to a z-table that score equates to 6.8%
SECOND: Determine the % of parts <507 mm.
z = (507 mm - 510.7 mm) / 1.49 mm = -2.48
Referring to a z-table that score equates to 0.66%
The total expected defective parts related to the length are:
6.8% + 0.66% = 7.46%
What is the probability that z is <= to 1.96 and >= 1.40?
This is a two step problem since you are looking for an area within the standard random normal distribution curve. Using the table below
P(1.96) = 0.9750
Think of this as representing 97.5% of the total area under the entire bell curve. See the table below to see how the 0.9750 was found.
P(1.4) = 0.9192 consider 91.92% of the total area under the entire bell curve. See the table below to see how the 0.9192 was found.
The area of interest is found by taking the total area under P(1.96) and subtracting the area under P(1.40). In other words, take 97.5% of the area and subtract 91.92% of the area and the remaining is the area of interest.
P(1.40 <= Z <= 1.96) = 0.9750 - 0.9192 = 0.0558
5.58% of the total area is under the entire standard normal distribution curve is in this region. This is the probability that z is within the region.
Be careful on the table being used. The above table is a two tailed table. A one-tailed table would show P(1.96) = 0.4750 and P(1.40) = 0.4192. The difference in the two values also equals 0.0558.
0.4750 represents the area from the peak of the bell curve (x=0) to the point of z=1.96 which is 47.5% of half the bell curve.
Also, the answer would be the same if the question asked for the P(-1.96 <= Z <= -1.40). This is simply the same area (probability) but on the other side of the bell curve.
Six Sigma Certification
Six Sigma Modules
Green Belt Program (1,000+ Slides)
Cause & Effect Matrix
Central Limit Theorem
1-Way Anova Test
Correlation and Regression