Instatistics, asampling distribution orfinite-sample distribution is theprobability distribution of a givenrandom-sample-basedstatistic. For an arbitrarily large number of samples where each sample, involving multiple observations (data points), is separately used to compute one value of a statistic (for example, thesample mean or samplevariance) per sample, the sampling distribution is the probability distribution of the values that the statistic takes on. In many contexts, only one sample (i.e., a set of observations) is observed, but the sampling distribution can be found theoretically.
Sampling distributions are important in statistics because they provide a major simplification en route tostatistical inference. More specifically, they allow analytical considerations to be based on the probability distribution of a statistic, rather than on thejoint probability distribution of all the individual sample values.
Thesampling distribution of a statistic is thedistribution of that statistic, considered as arandom variable, when derived from arandom sample of size. It may be considered as the distribution of the statistic forall possible samples from the same population of a given sample size. The sampling distribution depends on the underlyingdistribution of the population, the statistic being considered, the sampling procedure employed, and the sample size used. There is often considerable interest in whether the sampling distribution can be approximated by anasymptotic distribution, which corresponds to the limiting case either as the number of random samples of finite size, taken from an infinite population and used to produce the distribution, tends to infinity, or when just one equally-infinite-size "sample" is taken of that same population.
For example, consider anormal population with mean and variance. Assume we repeatedly take samples of a given size from this population and calculate thearithmetic mean for each sample – this statistic is called thesample mean. The distribution of these means, or averages, is called the "sampling distribution of the sample mean". This distribution is normal (n is the sample size) since the underlying population is normal, although sampling distributions may be close to normal even when the population distribution is not (seecentral limit theorem). An alternative to the sample mean is the samplemedian. When calculated from the same population, it has a different sampling distribution to that of the mean and is generally not normal (but it may be close for large sample sizes).
The mean of a sample from a population having a normal distribution is an example of a simple statistic taken from one of the simpleststatistical populations. For other statistics and other populations the formulas are more complicated, and often they do not exist inclosed-form. In such cases the sampling distributions may be approximated throughMonte-Carlo simulations,[1]bootstrap methods, orasymptotic distribution theory.
Thestandard deviation of the sampling distribution of astatistic is referred to as thestandard error of the statistic. For the case where the statistic is the sample mean, and samples are uncorrelated, the standard error is:where is the standard deviation of the population distribution of that quantity and is the sample size (number of items in the sample).
An important implication of this formula is that the sample size must be quadrupled (multiplied by 4) to achieve half (1/2) the measurement error. When designing statistical studies where cost is a factor, this may have a role in understanding cost–benefit tradeoffs.
For the case where the statistic is the sample total, and samples are uncorrelated, the standard error is:where, again, is the standard deviation of the population distribution of that quantity and is the sample size (number of items in the sample).

| Population | Statistic | Sampling distribution |
|---|---|---|
| Normal: | Sample mean from samples of sizen | . If the standard deviation is not known, one can consider, which follows theStudent's t-distribution with degrees of freedom. Here is the sample variance, and is apivotal quantity, whose distribution does not depend on. |
| Bernoulli: | Sample proportion of "successful trials" | |
| Two independent normal populations: and | Difference between sample means, | |
| Any absolutely continuous distributionF with densityf | Median from a sample of sizen = 2k − 1, where sample is ordered to | |
| Any distribution with distribution functionF | Maximum from a random sample of sizen |