This articleneeds additional citations forverification. Please helpimprove this article byadding citations to reliable sources. Unsourced material may be challenged and removed. Find sources: "Method of moments" statistics – news ·newspapers ·books ·scholar ·JSTOR(June 2020) (Learn how and when to remove this message) |
Instatistics, themethod of moments is a method ofestimation of populationparameters. The same principle is used to derive higher moments like skewness andkurtosis.
It starts by expressing the populationmoments (i.e., theexpected values of powers of therandom variable under consideration) as functions of the parameters of interest. Those expressions are then set equal to the sample moments. The number of such equations is the same as the number of parameters to be estimated. Those equations are then solved for the parameters of interest. The solutions are estimates of those parameters.
The method of moments was introduced byPafnuty Chebyshev in 1887 in the proof of thecentral limit theorem. The idea of matching empirical moments of a distribution to the population moments dates back at least toKarl Pearson.[1]
Suppose that the parameter = () characterizes thedistribution of the random variable.[1] Suppose the first moments of the true distribution (the "population moments") can be expressed as functions of thes:
Suppose a sample of size is drawn, resulting in the values. For, let be thej-th sample moment, an estimate of. The method of moments estimator for denoted by is defined to be the solution (if one exists) to the equations:[2]
The method described here for single random variables generalizes in an obvious manner to multiple random variables leading to multiple choices for moments to be used. Different choices generally lead to different solutions.[2][3]
The method of moments is fairly simple and yieldsconsistent estimators (under very weak assumptions), though these estimators are oftenbiased.
It is an alternative to themethod of maximum likelihood.
However, in some cases the likelihood equations may be intractable without computers, whereas the method-of-moments estimators can be computed much more quickly and easily. Due to easy computability, method-of-moments estimates may be used as the first approximation to the solutions of the likelihood equations, and successive improved approximations may then be found by theNewton–Raphson method. In this way the method of moments can assist in finding maximum likelihood estimates.
In some cases, infrequent with large samples but less infrequent with small samples, the estimates given by the method of moments are outside of the parameter space (as shown in the example below); it does not make sense to rely on them then. That problem never arises in the method ofmaximum likelihood[3] Also, estimates by the method of moments are not necessarilysufficient statistics, i.e., they sometimes fail to take into account all relevant information in the sample.
When estimating other structural parameters (e.g., parameters of autility function, instead of parameters of a known probability distribution), appropriate probability distributions may not be known, and moment-based estimates may be preferred to maximum likelihood estimation.
The equations to be solved in the method of moments (MoM) are in general nonlinear and there are no generally applicable guarantees that tractable solutions exist[citation needed]. But there is an alternative approach to using sample moments to estimate data model parameters in terms of known dependence of model moments on these parameters, and this alternative requires the solution of only linear equations or, more generally, tensor equations. This alternative is referred to as the Bayesian-Like MoM (BL-MoM), and it differs from the classical MoM in that it uses optimally weighted sample moments. Considering that the MoM is typically motivated by a lack of sufficient knowledge about the data model to determine likelihood functions and associateda posteriori probabilities of unknown or random parameters, it is odd that there exists a type of MoM that isBayesian-Like. But the particular meaning ofBayesian-Like leads to a problem formulation in which required knowledge ofa posteriori probabilities is replaced with required knowledge of only the dependence of model moments on unknown model parameters, which is exactly the knowledge required by the traditional MoM [1],[2].[3][2][4][5][6] The BL-MoM also uses knowledge ofa priori probabilities of the parameters to be estimated, when available, but otherwise uses uniform priors.[citation needed]
The BL-MoM has been reported on in only the applied statistics literature in connection with parameter estimation andhypothesis testing using observations of stochastic processes for problems in Information and Communications Theory and, in particular, communications receiver design in the absence of knowledge of likelihood functions or associateda posteriori probabilities[7] and references therein. In addition, the restatement of this receiver design approach forstochastic process models as an alternative to the classical MoM for any type of multivariate data is available in tutorial form at the university website.[8] The applications in[7] and references demonstrate some important characteristics of this alternative to the classical MoM, and a detailed list of relative advantages and disadvantages is given in,[8] but the literature is missing direct comparisons in specific applications of the classical MoM and the BL-MoM.[citation needed]
An example application of the method of moments is to estimate polynomial probability density distributions. In this case, an approximating polynomial of order is defined on an interval. The method of moments then yields a system of equations, whose solution involves the inversion of aHankel matrix.[9]
This section'sfactual accuracy isdisputed. Relevant discussion may be found on thetalk page. Please help to ensure that disputed statements arereliably sourced.(October 2024) (Learn how and when to remove this message) |
Let beindependent random variables with mean 0 and variance 1, then let. We can compute the moments of asExplicit expansion shows thatwhere the numerator is the number of ways to select distinct pairs of balls by picking one each from buckets, each containing balls numbered from to. At the limit, all moments converge to that of a standard normal distribution. More analysis then show that this convergence in moments imply a convergence in distribution.
Essentially this argument was published by Chebyshev in 1887.[10]
Consider theuniform distribution on the interval,. If then we have
Solving these equations gives
Given a set of samples we can use the sample moments and in these formulae in order to estimate and.
Note, however, that this method can produce inconsistent results in some cases. For example, the set of samples results in the estimate,. Since it is impossible for the set to have been drawn from in this case.