The normal distribution is the most important probability distribution in statistics. Also called the bell curve or Gaussian distribution, it describes a pattern where most observations cluster around a central value and the frequency of observations decreases symmetrically as you move away from the center.
The normal distribution is fully determined by just two parameters: the mean (μ), which sets the center of the curve, and the standard deviation (σ), which controls how wide or narrow the curve is. A small standard deviation produces a tall, narrow bell; a large standard deviation produces a short, wide bell. Regardless of the specific values, the shape is always perfectly symmetric.
One of the most practical features of the normal distribution is the empirical rule, which tells us how data is distributed around the mean:
This rule provides a quick way to assess whether individual data points are typical or unusual. A value more than 2 standard deviations from the mean occurs less than 5% of the time — making it relatively rare.
LakeFront Retail Co. has analyzed years of shipping data and found that delivery times are approximately normally distributed with a mean of 3.2 days and a standard deviation of 0.6 days.
Using the empirical rule: 68% of deliveries arrive between 2.6 and 3.8 days, 95% arrive between 2.0 and 4.4 days, and 99.7% arrive between 1.4 and 5.0 days. A delivery taking more than 5 days would be extremely unusual and worth investigating.
Adjust the mean and standard deviation to see how the normal curve changes shape.
A raw data value by itself does not tell us much about how unusual it is. Is a daily sales figure of $4,800 impressive? It depends on the context — the average level and the typical amount of variation. A Z-score solves this problem by converting any raw value into a standardized measure that tells us exactly how many standard deviations it is from the mean.
Z-scores allow us to compare values across different scales. For example, you can compare a store's sales performance (measured in dollars) to its customer satisfaction score (measured on a 1-10 scale) by converting both to Z-scores. The Z-score strips away the original units and expresses everything in terms of standard deviations.
=STANDARDIZE(x, mean, std_dev)Z is the standardized score, x is the raw value, μ is the population mean, and σ is the population standard deviation.
LakeFront Store A has daily sales with a mean of $4,100 and standard deviation of $320. Yesterday, Store A had sales of $4,800.
The Z-score is: Z = (4,800 − 4,100) / 320 = 700 / 320 = 2.19.
This means yesterday's sales were 2.19 standard deviations above the mean — a very strong day. Using the empirical rule, values beyond Z = 2 occur less than 5% of the time, so this was an exceptionally good performance.
Enter a value, mean, and standard deviation to compute the Z-score and see its position on the normal curve.
The real power of Z-scores comes from connecting them to probabilities. Once you convert a value to a Z-score, you can use the standard normal table (Z-table) or Excel functions to find the probability of observing a value at or below that point. This probability corresponds to the area under the curve to the left of the Z-score.
=NORM.S.DIST(z, TRUE).=1-NORM.S.DIST(z, TRUE).Excel also offers =NORM.DIST(x, mean, std_dev, TRUE) which lets you skip the Z-score conversion entirely and calculate probabilities directly from raw values.
=NORM.S.DIST(z, TRUE)P(Z ≤ z), the area under the standard normal curve to the left of z.
=NORM.DIST(x, mean, std_dev, TRUE)P(X ≤ x) directly without needing to compute the Z-score first.
Z-scores are the bridge between raw data and probability. By standardizing any value — regardless of its original scale — you can determine exactly how unusual it is and calculate the probability of observing values above, below, or between any two points. In business, this translates directly to setting thresholds, identifying outliers, and making data-driven decisions about what counts as “normal” versus “exceptional.”
In this chapter, we explored the normal distribution and learned how Z-scores let us standardize any data value and connect it to probabilities. Here is what you should take away:
Normal Distribution: The bell curve is defined by its mean (μ) and standard deviation (σ). The empirical rule (68-95-99.7) gives a quick estimate of how data is spread around the center.
Z-Scores: Standardizing a value tells you how many standard deviations it is from the mean, enabling comparisons across different scales and identifying outliers.
Probabilities: By converting to Z-scores, you can find the probability (area under the curve) of any range of values using tables or Excel functions like NORM.S.DIST and NORM.DIST.
The Big Picture: The normal distribution underpins most of the inferential statistics we will study next — including confidence intervals, hypothesis tests, and sampling theory. Mastering Z-scores now pays dividends in every chapter that follows.