Sample size determination orestimation is the act of choosing the number of observations orreplicates to include in astatistical sample. The sample size is an important feature of any empirical study in which the goal is to makeinferences about apopulation from a sample. In practice, the sample size used in a study is usually determined based on the cost, time, or convenience of collecting the data, and the need for it to offer sufficientstatistical power. In complex studies, different sample sizes may be allocated, such as in stratified surveys or experimental designs with multiple treatment groups. In acensus, data is sought for an entire population, hence the intended sample size is equal to the population. Inexperimental design, where a study may be divided into differenttreatment groups, there may be different sample sizes for each group.
Sample sizes may be chosen in several ways:
Sample size determination is a crucial aspect of research methodology that plays a significant role in ensuring the reliability and validity of study findings. In order to influence the accuracy of estimates, the power of statistical tests, and the general robustness of the research findings, it entails carefully choosing the number of participants or data points to be included in a study.
Consider the case where we are conducting a survey to determine the average satisfaction level of customers regarding a new product. To determine an appropriate sample size, we need to consider factors such as the desired level of confidence, margin of error, and variability in the responses. We might decide that we want a 95% confidence level, meaning we are 95% confident that the true average satisfaction level falls within the calculated range. We also decide on a margin of error, of ±3%, which indicates the acceptable range of difference between our sample estimate and the true population parameter. Additionally, we may have some idea of the expected variability in satisfaction levels based on previous data or assumptions.
Larger sample sizes generally lead to increasedprecision whenestimating unknown parameters. For instance, to accurately determine the prevalence of pathogen infection in a specific species of fish, it is preferable to examine a sample of 200 fish rather than 100 fish. Several fundamental facts of mathematical statistics describe this phenomenon, including thelaw of large numbers and thecentral limit theorem.
In some situations, the increase in precision for larger sample sizes is minimal, or even non-existent. This can result from the presence ofsystematic errors or strongdependence in the data, or if the data follows a heavy-tailed distribution, or because the data is strongly dependent or biased.
Sample sizes may be evaluated by the quality of the resulting estimates, as follows. It is usually determined on the basis of the cost, time or convenience of data collection and the need for sufficient statistical power. For example, if a proportion is being estimated, one may wish to have the 95%confidence interval be less than 0.06 units wide. Alternatively, sample size may be assessed based on thepower of a hypothesis test. For example, if we are comparing the support for a certain political candidate among women with the support for that candidate among men, we may wish to have 80% power to detect a difference in the support levels of 0.04 units.
A relatively simple situation is estimation of aproportion. It is a fundamental aspect of statistical analysis, particularly when gauging the prevalence of a specific characteristic within a population. For example, we may wish to estimate the proportion of residents in a community who are at least 65 years old.
Theestimator of aproportion is, whereX is the number of 'positive' instances (e.g., the number of people out of then sampled people who are at least 65 years old). When the observations areindependent, this estimator has a (scaled)binomial distribution (and is also thesamplemean of data from aBernoulli distribution). The maximumvariance of this distribution is 0.25, which occurs when the trueparameter isp = 0.5. In practical applications, where the true parameterp is unknown, the maximum variance is often employed for sample size assessments. If a reasonable estimate for p is known the quantity may be used in place of 0.25.
As the sample sizen grows sufficiently large, the distribution of will be closely approximated by anormal distribution.[1] Using this and theWald method for the binomial distribution, yields a confidence interval, with Z representing the standard Z-score for the desired confidence level (e.g., 1.96 for a 95% confidence interval), in the form:
To determine an appropriate sample sizen for estimating proportions, the equation below can be solved, where W represents the desired width of the confidence interval. The resulting sample size formula, is often applied with a conservative estimate ofp (e.g., 0.5):
forn, yielding the sample size
, in the case of using 0.5 as the most conservative estimate of the proportion.(Note: W/2 =margin of error.)
In the figure below one can observe how sample sizes for binomial proportions change given different confidence levels and margins of error.
Otherwise, the formula would be , which yields.For example, in estimating the proportion of the U.S. population supporting a presidential candidate with a 95% confidence interval width of 2 percentage points (0.02), a sample size of (1.96)2/ (0.022) = 9604 is required with the margin of error in this case is 1 percentage point. It is reasonable to use the 0.5 estimate for p in this case because the presidential races are often close to 50/50, and it is also prudent to use a conservative estimate. Themargin of error in this case is 1 percentage point (half of 0.02).
In practice, the formula : is commonly used to form a 95% confidence interval for the true proportion. The equation can be solved forn, providing a minimum sample size needed to meet the desired margin of errorW. The foregoing is commonly simplified:[2][3]n = 4/W2 = 1/B2 whereB is the error bound on the estimate, i.e., the estimate is usually given aswithin ± B. ForB = 10% one requiresn = 100, forB = 5% one needsn = 400, forB = 3% the requirement approximates ton = 1000, while forB = 1% a sample size ofn = 10000 is required. These numbers are quoted often in news reports ofopinion polls and othersample surveys. However, the results reported may not be the exact value as numbers are preferably rounded up. Knowing that the value of then is the minimum number ofsample points needed to acquire the desired result, the number of respondents then must lie on or above the minimum.
Simply speaking, if we are trying to estimate the average time it takes for people to commute to work in a city. Instead of surveying the entire population, you can take a random sample of 100 individuals, record their commute times, and then calculate the mean (average) commute time for that sample. For example, person 1 takes 25 minutes, person 2 takes 30 minutes, ..., person 100 takes 20 minutes. Add up all the commute times and divide by the number of people in the sample (100 in this case). The result would be your estimate of the mean commute time for the entire population. This method is practical when it's not feasible to measure everyone in the population, and it provides a reasonable approximation based on a representative sample.
In a precisely mathematical way, when estimating the population mean using an independent and identically distributed (iid) sample of sizen, where each data value has varianceσ2, thestandard error of the sample mean is:
This expression describes quantitatively how the estimate becomes more precise as the sample size increases. Using thecentral limit theorem to justify approximating the sample mean with a normal distribution yields a confidence interval of the form
To determine the sample sizen required for a confidence interval of width W, with W/2 as the margin of error on each side of the sample mean, the equation
.
For instance, if estimating the effect of a drug on blood pressure with a 95% confidence interval that is six units wide, and the known standard deviation of blood pressure in the population is 15, the required sample size would be, which would be rounded up to 97, since sample sizes must be integers and must meet or exceed the calculatedminimum value. Understanding these calculations is essential for researchers designing studies to accurately estimate population means within a desired level of confidence.
One of the prevalent challenges faced by statisticians revolves around the task of calculating the sample size needed to attain a specified statistical power for a test, all while maintaining a pre-determinedType I error rate α, which signifies the level of significance in hypothesis testing. It yields a certainpower for a test, given a predetermined. As follows, this can be estimated by pre-determined tables for certain values, by formulas, by simulation, by Mead's resource equation, or by thecumulative distribution function:
[4] Power | Cohen's d | ||
---|---|---|---|
0.2 | 0.5 | 0.8 | |
0.25 | 84 | 14 | 6 |
0.50 | 193 | 32 | 13 |
0.60 | 246 | 40 | 16 |
0.70 | 310 | 50 | 20 |
0.80 | 393 | 64 | 26 |
0.90 | 526 | 85 | 34 |
0.95 | 651 | 105 | 42 |
0.99 | 920 | 148 | 58 |
The table shown on the right can be used in atwo-sample t-test to estimate the sample sizes of anexperimental group and acontrol group that are of equal size, that is, the total number of individuals in the trial is twice that of the number given, and the desiredsignificance level is 0.05.[4] The parameters used are:
Calculating a required sample size is often not easy since the distribution of the test statistic under the alternative hypothesis of interest is usually hard to work with. Approximate sample size formulas for specific problems are available - some general references are[5]and[6]
The QuickSize algorithm[7]is a very general approach that is simple to use yet versatile enough to give an exact solution for a broad range of problems. It uses simulation together with a search algorithm.
Mead's resource equation is often used for estimating sample sizes oflaboratory animals, as well as in many other laboratory experiments. It may not be as accurate as using other methods in estimating sample size, but gives a hint of what is the appropriate sample size where parameters such as expected standard deviations or expected differences in values between groups are unknown or very hard to estimate.[8]
All the parameters in the equation are in fact thedegrees of freedom of the number of their concepts, and hence, their numbers are subtracted by 1 before insertion into the equation.
The equation is:[8]
where:
For example, if a study using laboratory animals is planned with four treatment groups (T=3), with eight animals per group, making 32 animals total (N=31), without any furtherstratification (B=0), thenE would equal 28, which is above the cutoff of 20, indicating that sample size may be a bit too large, and six animals per group might be more appropriate.[9]
LetXi,i = 1, 2, ...,n be independent observations taken from anormal distribution with unknown mean μ and known variance σ2. Consider two hypotheses, anull hypothesis:
and an alternative hypothesis:
for some 'smallest significant difference'μ* > 0. This is the smallest value for which we care about observing a difference. Now, for (1) to rejectH0 with a probability of at least 1 − β whenHa is true (i.e. apower of 1 − β), and (2) rejectH0 with probability α whenH0 is true, the following is necessary: Ifzα is the upper α percentage point of the standard normal distribution, then
and so
is adecision rule which satisfies (2). (This is a 1-tailed test.) In such a scenario, achieving this with a probability of at least 1−β when the alternative hypothesisHa is true becomes imperative. Here, the sample average originates from a Normal distribution with a mean ofμ*. Thus, the requirement is expressed as:
Through careful manipulation, this can be shown (seeStatistical power Example) to happen when
where is the normalcumulative distribution function.
With more complicated sampling techniques, such asstratified sampling, the sample can often be split up into sub-samples. Typically, if there areH such sub-samples (fromH different strata) then each of them will have a sample sizenh,h = 1, 2, ...,H. Thesenh must conform to the rule thatn1 +n2 + ... +nH =n (i.e., that the total sample size is given by the sum of the sub-sample sizes). Selecting thesenh optimally can be done in various ways, using (for example) Neyman's optimal allocation.
There are many reasons to use stratified sampling:[10] to decrease variances of sample estimates, to use partly non-random methods, or to study strata individually. A useful, partly non-random method would be to sample individuals where easily accessible, but, where not, sample clusters to save travel costs.[11]
In general, forH strata, a weighted sample mean is
with
The weights,, frequently, but not always, represent the proportions of the population elements in the strata, and. For a fixed sample size, that is,
which can be made a minimum if thesampling rate within each stratum is madeproportional to the standard deviation within each stratum:, where and is a constant such that.
An "optimum allocation" is reached when the sampling rates within the strataare made directly proportional to the standard deviations within the strataand inversely proportional to the square root of the sampling cost per elementwithin the strata,:
where is a constant such that, or, more generally, when
Qualitative research approaches sample size determination with a distinctive methodology that diverges from quantitative methods. Rather than relying on predetermined formulas or statistical calculations, it involves a subjective and iterative judgment throughout the research process. In qualitative studies, researchers often adopt a subjective stance, making determinations as the study unfolds. Sample size determination in qualitative studies takes a different approach. It is generally a subjective judgment, taken as the research proceeds.[16] One common approach is to continually include additional participants or materials until a point of "saturation" is reached. Saturation occurs when new participants or data cease to provide fresh insights, indicating that the study has adequately captured the diversity of perspectives or experiences within the chosen samplesaturation is reached.[17] The number needed to reach saturation has been investigated empirically.[18][19][20][21]
Unlike quantitative research, qualitative studies face a scarcity of reliable guidance regarding sample size estimation prior to beginning the research.Imagine conducting in-depth interviews with cancer survivors, qualitative researchers may use data saturation to determine the appropriate sample size. If, over a number of interviews, no fresh themes or insights show up, saturation has been reached and more interviews might not add much to our knowledge of the survivor's experience. Thus, rather than following a preset statistical formula, the concept of attaining saturation serves as a dynamic guide for determining sample size in qualitative research. There is a paucity of reliable guidance on estimating sample sizes before starting the research, with a range of suggestions given.[19][22][23][24] In an effort to introduce some structure to the sample size determination process in qualitative research, a tool analogous to quantitative power calculations has been proposed. This tool, based on thenegative binomial distribution, is particularly tailored forthematic analysis.[25][24]