- Method of moments (statistics)
In statistics, the method of moments is a method of estimation of population parameters such as mean, variance, median, etc. (which need not be moments), by equating sample moments with unobservable population moments and then solving those equations for the quantities to be estimated.
Suppose that the problem is to estimate p unknown parameters θ1,θ2,...θp characterizing a distribution fW(w;θ). Suppose p of the moments of the true distribution can be expressed as functions of the θs:
Let be the j-th sample moment corresponding to the population moment μj. The method of moments estimator for θ1,θ2,...θp denoted by is defined by the solution (if there is one) to the equations:
for x > 0, and 0 for x < 0.
The first moment, i.e., the expected value, of a random variable with this probability distribution is
and the second moment, i.e., the expected value of its square, is
These are the "population moments".
The first and second "sample moments" m1 and m2 are respectively
Equating the population moments with the sample moments, we get
Solving these two equations for α and β, we get
We then use these 2 quantities as estimates, based on the sample, of the two unobservable population parameters α and β.
Advantages and disadvantages of this method
In some respects, when estimating parameters of a known family of probability distributions, this method was superseded by Fisher's method of maximum likelihood, because maximum likelihood estimators have higher probability of being close to the quantities to be estimated.
However, in some cases, as in the above example of the gamma distribution, the likelihood equations may be intractable without computers, whereas the method-of-moments estimators can be quickly and easily calculated by hand as shown above.
Estimates by the method of moments may be used as the first approximation to the solutions of the likelihood equations, and successive improved approximations may then be found by the Newton–Raphson method. In this way the method of moments and the method of maximum likelihood are symbiotic.
In some cases, infrequent with large samples but not so infrequent with small samples, the estimates given by the method of moments are outside of the parameter space; it does not make sense to rely on them then. That problem never arises in the method of maximum likelihood. Also, estimates by the method of moments are not necessarily sufficient statistics, i.e., they sometimes fail to take into account all relevant information in the sample.
When estimating other structural parameters (e.g., parameters of a utility function, instead of parameters of a known probability distribution), appropriate probability distributions may not be known, and moment-based estimates may be preferred to MLE.
Statistics Descriptive statisticsSummary tables Data collectionDesigning studiesUncontrolled studies Statistical inferenceFrequentist inferenceSpecific tests Correlation and regression analysisNon-standard predictorsPartition of variance Categorical, multivariate, time-series, or survival analysis Applications Category · Portal · Outline · Index
Wikimedia Foundation. 2010.
Look at other dictionaries:
Method of moments — may refer to: Method of moments (statistics), a method of parameter estimation in statistics Method of moments (probability theory), a way of proving convergence in distribution in probability theory Second moment method, a technique used in… … Wikipedia
Generalized method of moments — GMM may also mean Gaussian mixture model. : For the Thai entertainment company, see GMM Grammy. The generalized method of moments is a very general statistical method for obtaining estimates of parameters of statistical models. It is a… … Wikipedia
List of statistics topics — Please add any Wikipedia articles related to statistics that are not already on this list.The Related changes link in the margin of this page (below search) leads to a list of the most recent changes to the articles listed below. To see the most… … Wikipedia
Monte Carlo method — Not to be confused with Monte Carlo algorithm. Computational physics … Wikipedia
Cross-validation (statistics) — Cross validation, sometimes called rotation estimation, is a technique for assessing how the results of a statistical analysis will generalize to an independent data set. It is mainly used in settings where the goal is prediction, and… … Wikipedia
Official statistics — on Germany in 2010, published in UNECE Countries in Figures 2011. Official statistics are statistics published by government agencies or other public bodies such as international organizations. They provide quantitative or qualitative information … Wikipedia
Descriptive statistics — quantitatively describe the main features of a collection of data. Descriptive statistics are distinguished from inferential statistics (or inductive statistics), in that descriptive statistics aim to summarize a data set, rather than use the… … Wikipedia
Outline of statistics — The following outline is provided as an overview and guide to the variety of topics included within the subject of statistics: Statistics pertains to the collection, analysis, interpretation, and presentation of data. It is applicable to a wide… … Wikipedia
Multivariate statistics — is a form of statistics encompassing the simultaneous observation and analysis of more than one statistical variable. The application of multivariate statistics is multivariate analysis. Methods of bivariate statistics, for example simple linear… … Wikipedia
Standard error (statistics) — For a value that is sampled with an unbiased normally distributed error, the above depicts the proportion of samples that would fall between 0, 1, 2, and 3 standard deviations above and below the actual value. The standard error is the standard… … Wikipedia