In the previous chapter, we used the Z-test when the population standard deviation σ was known. In practice, this is rare. Most of the time, we estimate σ using the sample standard deviation s. When we make this substitution, the test statistic no longer follows a standard normal distribution — it follows a Student’s t-distribution.
The t-distribution was discovered by William Sealy Gosset, who published under the pseudonym “Student” while working at the Guinness brewery. It is similar to the normal distribution but has heavier tails, reflecting the additional uncertainty that comes from estimating σ with s.
GreatLakes has a new production line that has only produced 15 shafts so far. The population standard deviation is unknown — they only have the sample standard deviation s = 0.18 mm. Because σ is unknown and n is small, a t-test is the appropriate choice.
=(xbar-mu0)/(s/SQRT(n)) is the sample mean, μ0 is the hypothesized population mean, s is the sample standard deviation, n is the sample size, and df = n − 1.
You can also use Excel’s =T.TEST() function for comparing arrays, or compute the t-statistic manually with =(xbar-mu0)/(s/SQRT(n)).
The one-sample t-test follows the same logical framework as the Z-test: formulate hypotheses, compute a test statistic, compare it to a critical value (or compute a p-value), and make a decision. The only difference is that we use the t-distribution instead of the Z-distribution, and critical values come from t-tables with the appropriate degrees of freedom.
The one-sample t-test relies on several assumptions:
The t-test is robust to moderate departures from normality, especially when n is large. With n ≥ 30, the Central Limit Theorem ensures that the sampling distribution of is approximately normal regardless of the population distribution. For small samples, check for severe skewness or outliers before relying on t-test results.
In GreatLakes’ case, with n = 15, the engineer should verify that the shaft diameter data does not show extreme skewness or outliers before trusting the t-test conclusion. A simple histogram or normal probability plot can help with this assessment.
The t-test is the workhorse of hypothesis testing in practice because population standard deviations are rarely known. As the sample size grows, the t-distribution approaches the normal distribution, and the choice between t-test and Z-test matters less. For small samples, always check normality assumptions and watch for outliers.
In this chapter, we extended hypothesis testing to the more common scenario where the population standard deviation is unknown. Here is what you should take away:
When to Use the T-Test: Use a t-test instead of a Z-test when σ is unknown and must be estimated by the sample standard deviation s. The t-distribution has heavier tails, reflecting additional estimation uncertainty.
Degrees of Freedom: For a one-sample t-test, df = n − 1. As df increases, the t-distribution converges to the standard normal.
Same Logic, Different Distribution: The hypothesis testing procedure is identical to the Z-test; only the reference distribution and critical values change.
Assumptions: Random sampling, independence, and approximate normality. The t-test is robust to non-normality for larger samples.