Sampling Distributions
Medication dosages, IV drip rates, vital monitoring
If you take two different random samples from the same population, you will get different sample means. Take a third sample, and you will get yet another mean. This natural variability is not a flaw in the sampling process — it is an unavoidable feature of working with samples instead of entire populations. Sampling distributions describe exactly how much this variability is, what shape it takes, and how it shrinks as your sample size grows. Understanding sampling distributions is the key to understanding confidence intervals, hypothesis tests, and virtually all of statistical inference.
What Is a Sampling Distribution?
A sampling distribution is not the distribution of data in a single sample. That is a common and important misconception to clear up right away. A sampling distribution is the distribution of a statistic — such as the sample mean or the sample proportion — computed across many possible samples of the same size drawn from the same population.
Here is the thought experiment that makes this concrete:
- You have a population with some mean and standard deviation .
- You draw a random sample of size and compute the sample mean .
- You put that sample back (or draw from a very large population) and draw another sample of size . You compute again.
- You repeat this process 1,000 times — or 10,000 times, or a million times.
- Now you have 1,000 (or 10,000, or a million) values of . Plot them in a histogram.
That histogram is the sampling distribution of . It shows you all the possible values the sample mean could take and how likely each value is. In practice, you never actually draw thousands of samples — but the sampling distribution tells you what would happen if you did, and that knowledge is the foundation of statistical inference.
Sampling Distribution of the Sample Mean
The sampling distribution of has three key properties:
Center: The mean of the sampling distribution equals the population mean:
This says that the sample mean is an unbiased estimator of the population mean. If you averaged all possible sample means, you would get exactly . Individual samples may overshoot or undershoot, but on average they hit the target.
Spread: The standard deviation of the sampling distribution is:
This quantity is called the standard error of the mean. Notice the critical role of : as the sample size increases, the standard error decreases. Larger samples produce more consistent (less variable) estimates.
Shape: When is large enough, the sampling distribution of is approximately normal — regardless of the shape of the population. This remarkable fact is the Central Limit Theorem, covered on the next page. If the population itself is normal, then the sampling distribution of is exactly normal for any sample size.
Example 1: Adult Heights
Suppose adult heights in a population have cm and cm. You plan to take random samples of adults and compute for each sample.
The sampling distribution of has:
- Mean: cm
- Standard error: cm
Individual heights vary with a standard deviation of 8 cm. But sample means (with ) vary with a standard deviation of only 1 cm. The averaging process dramatically reduces variability — your sample mean will almost certainly be within a few centimeters of the true population mean.
Individual Heights vs. Sampling Distribution of the Mean (n = 64)
The diagram above illustrates the key idea: while individual heights (dashed blue curve) spread widely around 170 cm, sample means with (solid green curve) cluster tightly around 170 cm. The sampling distribution is far narrower than the population distribution.
Standard Error
The standard error (SE) is the standard deviation of a sampling distribution. It measures how much a statistic typically varies from sample to sample.
Standard error of the mean:
When the population standard deviation is unknown (which is typical in practice), we estimate it from the sample:
where is the sample standard deviation.
The key insight is the in the denominator: to cut the standard error in half, you must quadruple the sample size. This is the law of diminishing returns for sampling — initial increases in sample size produce large gains in precision, but each further increase yields less additional benefit.
Example 2: Effect of Sample Size
A population has . How does the standard error change as increases?
| Sample size () | ||
|---|---|---|
| 25 | 5 | 4.0 |
| 100 | 10 | 2.0 |
| 400 | 20 | 1.0 |
| 1,600 | 40 | 0.5 |
Going from to (a fourfold increase) cuts the SE from 4.0 to 2.0 — halving it. Going from to (another fourfold increase) cuts the SE from 2.0 to 1.0 — halving it again. Each time you want to halve the SE, you need four times as many observations. This relationship governs how researchers choose sample sizes for studies.
Sampling Distribution of the Sample Proportion
When working with categorical data (yes/no, pass/fail, support/oppose), the relevant statistic is the sample proportion . Its sampling distribution has analogous properties:
Center: The mean of equals the population proportion:
The sample proportion is an unbiased estimator of .
Spread: The standard error of is:
Shape: The sampling distribution of is approximately normal when both of these conditions are met:
These conditions ensure that there are enough “successes” and “failures” in the expected sample for the normal approximation to work well.
Example 3: Voter Support
A candidate has true support of in a population. A pollster takes a random sample of voters.
Standard error:
Normality check:
- --- condition met
- --- condition met
So is approximately normal with mean and standard error .
Interpretation: In repeated polls of 200 voters, the sample proportion would typically be within about (about 7 percentage points) of the true proportion. This is why polls report a “margin of error” — it comes directly from the standard error of the sampling distribution.
Why Sampling Distributions Matter
Sampling distributions are not just an abstract concept — they are the engine that powers statistical inference. Here is why:
- Confidence intervals use the standard error to build a range of plausible values around a sample statistic. A 95% confidence interval for a mean is approximately , which comes directly from the sampling distribution.
- Hypothesis tests ask: “If the true parameter were some specific value, how likely is the sample result we observed?” The sampling distribution provides the answer by telling us what values are typical and what values are surprising.
- Margin of error in polls and surveys is a direct application of the standard error of .
Without understanding sampling distributions, confidence intervals and p-values are just mysterious formulas. With this understanding, they become logical consequences of how statistics vary from sample to sample.
Real-World Application: Nursing — Precision of Blood Pressure Estimates
A hospital clinic wants to estimate the average systolic blood pressure of its patient population. From prior studies, mmHg is a reasonable estimate of the population standard deviation.
How precise is the estimate for different sample sizes?
| Patients measured () | 95% margin of error () | |
|---|---|---|
| 9 | mmHg | 12.0 mmHg |
| 36 | mmHg | 6.0 mmHg |
| 100 | mmHg | 3.6 mmHg |
| 225 | mmHg | 2.4 mmHg |
With only 9 patients, the estimate could easily be off by 12 mmHg — that is clinically meaningless because it spans the difference between normal and hypertensive. With 100 patients, the margin shrinks to 3.6 mmHg, which is precise enough for most clinical decisions. With 225 patients, the margin is 2.4 mmHg — very precise, but the improvement from 100 to 225 patients is modest compared to the improvement from 9 to 36.
This illustrates the practical tradeoff that nurses, physicians, and public health researchers face: more measurements improve precision, but with diminishing returns. The sampling distribution quantifies exactly how much precision you gain for each additional measurement.
Practice Problems
Test your understanding with these problems. Click to reveal each answer.
Problem 1: A population has and . For samples of size , find the mean and standard error of the sampling distribution of .
Mean of sampling distribution:
Standard error:
Answer: The sampling distribution of has mean 50 and standard error 2.
Problem 2: A researcher wants to cut the standard error of her estimate in half. Her current sample size is . What sample size does she need?
To halve the standard error, you must quadruple the sample size:
Verification: . --- confirmed.
Answer: She needs a sample size of 200.
Problem 3: A population proportion is and the sample size is . Find the standard error of and check whether the normal approximation is valid.
Standard error:
Normality check:
- --- met
- --- met
Answer: The standard error is approximately 0.039, and the normal approximation is valid because both conditions are satisfied.
Problem 4: Two researchers study the same population (). Researcher A uses and Researcher B uses . How do their standard errors compare?
Researcher A:
Researcher B:
Researcher B’s standard error is one-third of Researcher A’s. Since , the sample size increased by a factor of 9, so the standard error decreased by a factor of .
Answer: Researcher A has and Researcher B has . The ninefold increase in sample size reduced the standard error to one-third its original value.
Problem 5: In a large city, 20% of residents speak a language other than English at home (). A survey samples residents. What is the standard error of , and within what range would you expect 95% of sample proportions to fall?
Standard error:
95% range (approximately ):
So 95% of sample proportions would fall between 0.16 and 0.24 (i.e., between 16% and 24%).
Normality check: and --- valid.
Answer: The standard error is 0.02, and approximately 95% of sample proportions would fall between 0.16 and 0.24.
Key Takeaways
- A sampling distribution is the distribution of a statistic (like or ) across all possible samples of a given size — it describes how the statistic varies from sample to sample
- The sampling distribution of is centered at (unbiased) with standard error
- The sampling distribution of is centered at (unbiased) with standard error
- Standard error measures the typical distance between a sample statistic and the population parameter — it is the standard deviation of the sampling distribution
- Larger samples produce smaller standard errors — but you must quadruple to halve the SE
- The normal approximation for requires and
- Sampling distributions are the foundation of confidence intervals and hypothesis tests — they tell us what to expect when sampling from a population
Return to Statistics for more topics in this section.
Next Up in Statistics
All Statistics topicsLast updated: March 29, 2026