Movatterモバイル変換


[0]ホーム

URL:


Jump to content
WikipediaThe Free Encyclopedia
Search

Autoregressive moving-average model

From Wikipedia, the free encyclopedia
Statistical model used in time series analysis
"ARMA model" redirects here. For other uses, seeARMA (disambiguation).

In thestatistical analysis oftime series, anautoregressive–moving-average (ARMA) model is used to represent a(weakly) stationary stochastic process by combining two components:autoregression (AR) andmoving average (MA). These models are widely used for analyzing the structure of a series and for forecasting future values.

The AR component specifies that the current value of the series depends linearly on its own past values (lags), while the MA component specifies that the current value depends on alinear combination of pasterror terms. An ARMA model is typically denoted as ARMA(p,q), wherep is the order of the autoregressive part andq is the order of the moving-average part.

The general ARMA model was described in the 1951 thesis ofPeter Whittle,Hypothesis testing in time series analysis, and it was popularized in the 1970 book byGeorge E. P. Box andGwilym Jenkins.

ARMA models can be estimated by using theBox–Jenkins method.

Mathematical formulation

[edit]

Autoregressive model

[edit]
Main article:Autoregressive model

The notation AR(p) refers to the autoregressive model of orderp. The AR(p) model is written as

Xt=i=1pφiXti+εt{\displaystyle X_{t}=\sum _{i=1}^{p}\varphi _{i}X_{t-i}+\varepsilon _{t}}

whereφ1,,φp{\displaystyle \varphi _{1},\ldots ,\varphi _{p}} areparameters and the random variableεt{\displaystyle \varepsilon _{t}} iswhite noise, usuallyindependent and identically distributed (i.i.d.)normal random variables.[1][2]

In order for the model to remainstationary, the roots of itscharacteristic polynomial must lie outside the unit circle. For example, processes in the AR(1) model with|φ1|1{\displaystyle |\varphi _{1}|\geq 1} are not stationary because the root of1φ1B=0{\displaystyle 1-\varphi _{1}B=0} lies within the unit circle.[3]

Theaugmented Dickey–Fuller test can assesses the stability of anintrinsic mode function and trend components. For stationary time series, the ARMA models can be used, while for non-stationary series,Long short-term memory models can be used to derive abstract features. The final value is obtained by reconstructing the predicted outcomes of each time series.[citation needed]

Moving average model

[edit]
Main article:Moving-average model

The notation MA(q) refers to the moving average model of orderq:

Xt=μ+εt+i=1qθiεti{\displaystyle X_{t}=\mu +\varepsilon _{t}+\sum _{i=1}^{q}\theta _{i}\varepsilon _{t-i}\,}

where theθ1,...,θq{\displaystyle \theta _{1},...,\theta _{q}} are the parameters of the model,μ{\displaystyle \mu } is the expectation ofXt{\displaystyle X_{t}} (often assumed to equal 0), andε1{\displaystyle \varepsilon _{1}}, ...,εt{\displaystyle \varepsilon _{t}} are i.i.d. white noise error terms that are commonly normal random variables.[4]

ARMA model

[edit]

The notation ARMA(p,q) refers to the model withp autoregressive terms andq moving-average terms. This model contains the AR(p) and MA(q) models,[5]

Xt=εt+i=1pφiXti+i=1qθiεti.{\displaystyle X_{t}=\varepsilon _{t}+\sum _{i=1}^{p}\varphi _{i}X_{t-i}+\sum _{i=1}^{q}\theta _{i}\varepsilon _{t-i}.\,}

In terms of lag operator

[edit]

In some texts, the models is specified using thelag operatorL. In these terms, the AR(p) model is given by

εt=(1i=1pφiLi)Xt=φ(L)Xt{\displaystyle \varepsilon _{t}=\left(1-\sum _{i=1}^{p}\varphi _{i}L^{i}\right)X_{t}=\varphi (L)X_{t}\,}

whereφ{\displaystyle \varphi } represents the polynomial

φ(L)=1i=1pφiLi.{\displaystyle \varphi (L)=1-\sum _{i=1}^{p}\varphi _{i}L^{i}.\,}

The MA(q) model is given by

Xtμ=(1+i=1qθiLi)εt=θ(L)εt,{\displaystyle X_{t}-\mu =\left(1+\sum _{i=1}^{q}\theta _{i}L^{i}\right)\varepsilon _{t}=\theta (L)\varepsilon _{t},\,}

whereθ{\displaystyle \theta } represents the polynomial

θ(L)=1+i=1qθiLi.{\displaystyle \theta (L)=1+\sum _{i=1}^{q}\theta _{i}L^{i}.\,}

Finally, the combined ARMA(p,q) model is given by

(1i=1pφiLi)Xt=(1+i=1qθiLi)εt,{\displaystyle \left(1-\sum _{i=1}^{p}\varphi _{i}L^{i}\right)X_{t}=\left(1+\sum _{i=1}^{q}\theta _{i}L^{i}\right)\varepsilon _{t}\,,}

or more concisely,

φ(L)Xt=θ(L)εt{\displaystyle \varphi (L)X_{t}=\theta (L)\varepsilon _{t}\,}

or

φ(L)θ(L)Xt=εt.{\displaystyle {\frac {\varphi (L)}{\theta (L)}}X_{t}=\varepsilon _{t}\,.}

This is the form used inBox,Jenkins & Reinsel.[6]

Moreover, starting summations fromi=0{\displaystyle i=0} and settingϕ0=1{\displaystyle \phi _{0}=-1} andθ0=1{\displaystyle \theta _{0}=1}, then we get an even more elegant formulation:i=0pϕiLiXt=i=0qθiLiεt.{\displaystyle -\sum _{i=0}^{p}\phi _{i}L^{i}\;X_{t}=\sum _{i=0}^{q}\theta _{i}L^{i}\;\varepsilon _{t}\,.}

Spectrum

[edit]

Thespectral density of an ARMA process isS(f)=σ22π|θ(eif)ϕ(eif)|2{\displaystyle S(f)={\frac {\sigma ^{2}}{2\pi }}\left\vert {\frac {\theta (e^{-if})}{\phi (e^{-if})}}\right\vert ^{2}}whereσ2{\displaystyle \sigma ^{2}} is thevariance of the white noise,θ{\displaystyle \theta } is the characteristic polynomial of the moving average part of the ARMA model, andϕ{\displaystyle \phi } is the characteristic polynomial of the autoregressive part of the ARMA model.[7][8]

Fitting models

[edit]

Choosingp andq

[edit]

An appropriate value ofp in the ARMA(p,q) model can be found by plotting thepartial autocorrelation functions. Similarly,q can be estimated by using theautocorrelation functions. Bothp andq can be determined simultaneously using extended autocorrelation functions (EACF).[9] Further information can be gleaned by considering the same functions for the residuals of a model fitted with an initial selection ofp andq.

Brockwell & Davis recommend usingAkaike information criterion (AIC) for findingp andq.[10] Another option is theBayesian information criterion (BIC).

Estimating coefficients

[edit]

After choosingp andq, ARMA models can be fitted byleast squares regression to find the values of the parameters which minimize the error term. It is good practice to find the smallest values ofp andq which provide an acceptable fit to the data. For a pure AR model, theYule-Walker equations may be used to provide a fit.

ARMA outputs are used primarily to forecast (predict), and not to infer causation as in other areas of econometrics and regression methods such as OLS and 2SLS.

Software implementations

[edit]
  • InR, standard packagestats has functionarima, documented inARIMA Modelling of Time Series. Packageastsa has an improved script calledsarima for fitting ARMA models (seasonal and nonseasonal) andsarima.sim to simulate data from these models. Extension packages contain related and extended functionality: packagetseries includes the functionarma(), documented in"Fit ARMA Models to Time Series"; packagefracdiff containsfracdiff() for fractionally integrated ARMA processes; and packageforecast includesauto.arima for selecting a parsimonious set ofp, q. The CRAN task view onTime Series contains links to most of these.
  • Mathematica has a complete library of time series functions including ARMA.[11]
  • MATLAB includes functions such asarma,ar andarx to estimate autoregressive, exogenous autoregressive and ARMAX models. SeeSystem Identification Toolbox andEconometrics Toolbox for details.
  • Julia has community-driven packages that implement fitting with an ARMA model such asarma.jl.
  • Python has thestatsmodelsS package which includes many models and functions for time series analysis, including ARMA. Formerly part of thescikit-learn library, it is now stand-alone and integrates well withPandas.
  • PyFlux has a Python-based implementation of ARIMAX models, including Bayesian ARIMAX models.
  • IMSL Numerical Libraries are libraries of numerical analysis functionality including ARMA and ARIMA procedures implemented in standard programming languages like C, Java, C# .NET, and Fortran.
  • gretl can estimate ARMA models, as mentionedhere
  • GNU Octave extra packageoctave-forge supports AR models.
  • Stata includes the functionarima. for ARMA andARIMA models.
  • SuanShu is a Java library of numerical methods that implements univariate/multivariate ARMA, ARIMA, ARMAX, etc models, documented in"SuanShu, a Java numerical and statistical library".
  • SAS has an econometric package, ETS, that estimates ARIMA models.See details.

History and interpretations

[edit]

The general ARMA model was described in the 1951 thesis ofPeter Whittle, who used mathematical analysis (Laurent series andFourier analysis) and statistical inference.[12][13] ARMA models were popularized by a 1970 book byGeorge E. P. Box and Jenkins, who expounded an iterative (Box–Jenkins) method for choosing and estimating them. This method was useful for low-order polynomials (of degree three or less).[14]

ARMA is essentially aninfinite impulse response filter applied to white noise, with some additional interpretation placed on it.

Indigital signal processing, ARMA is represented as a digital filter with white noise at the input and the ARMA process at the output.

Applications

[edit]

ARMA is appropriate when a system is a function of a series of unobserved shocks (the MA or moving average part) as well as its own behavior. For example, stock prices may be shocked by fundamental information as well as exhibiting technical trending andmean-reversion effects due to market participants.[citation needed]

Generalizations

[edit]

There are various generalizations of ARMA.Nonlinear AR (NAR), nonlinear MA (NMA) and nonlinear ARMA (NARMA) model nonlinear dependence on past values and error terms.Vector AR (VAR) and vector ARMA (VARMA) modelmultivariate time series.Autoregressive integrated moving average (ARIMA) models non-stationary time series (that is, whose mean changes over time).Autoregressive conditional heteroskedasticity (ARCH) models time series where the variance changes. Seasonal ARIMA (SARIMA or periodic ARMA) modelsperiodic variation.Autoregressive fractionally integrated moving average (ARFIMA, or Fractional ARIMA, FARIMA) model time-series that exhibitslong memory. Multiscale AR (MAR) is indexed by the nodes of atree instead of integers.

Autoregressive–moving-average model with exogenous inputs (ARMAX)

[edit]

The notation ARMAX(p,q,b) refers to a model withp autoregressive terms,q moving average terms andb exogenous inputs terms. The last term is a linear combination of the lastb terms of a known and external time seriesdt{\displaystyle d_{t}}. It is given by:

Xt=εt+i=1pφiXti+i=1qθiεti+i=1bηidti.{\displaystyle X_{t}=\varepsilon _{t}+\sum _{i=1}^{p}\varphi _{i}X_{t-i}+\sum _{i=1}^{q}\theta _{i}\varepsilon _{t-i}+\sum _{i=1}^{b}\eta _{i}d_{t-i}.\,}

whereη1,,ηb{\displaystyle \eta _{1},\ldots ,\eta _{b}} are theparameters of the exogenous inputdt{\displaystyle d_{t}}.

Some nonlinear variants of models with exogenous variables have been defined: see for exampleNonlinear autoregressive exogenous model.

Statistical packages implement the ARMAX model through the use of "exogenous" (that is, independent) variables. Care must be taken when interpreting the output of those packages, because the estimated parameters usually (for example, inR[15] andgretl) refer to the regression:

Xtmt=εt+i=1pφi(Xtimti)+i=1qθiεti.{\displaystyle X_{t}-m_{t}=\varepsilon _{t}+\sum _{i=1}^{p}\varphi _{i}(X_{t-i}-m_{t-i})+\sum _{i=1}^{q}\theta _{i}\varepsilon _{t-i}.\,}

wheremt{\displaystyle m_{t}} incorporates all exogenous (or independent) variables:

mt=c+i=0bηidti.{\displaystyle m_{t}=c+\sum _{i=0}^{b}\eta _{i}d_{t-i}.\,}

See also

[edit]
This article includes a list ofgeneral references, butit lacks sufficient correspondinginline citations. Please help toimprove this article byintroducing more precise citations.(August 2010) (Learn how and when to remove this message)

References

[edit]
  1. ^Box, George E. P. (1994).Time series analysis : forecasting and control. Gwilym M. Jenkins, Gregory C. Reinsel (3rd ed.). Englewood Cliffs, N.J.: Prentice Hall. p. 54.ISBN 0-13-060774-6.OCLC 28888762.
  2. ^Shumway, Robert H. (2000).Time series analysis and its applications. David S. Stoffer. New York: Springer. pp. 90–91.ISBN 0-387-98950-1.OCLC 42392178.
  3. ^Box, George E. P.; Jenkins, Gwilym M.; Reinsel, Gregory C. (1994).Time series analysis : forecasting and control (3rd ed.). Englewood Cliffs, N.J.: Prentice Hall. pp. 54–55.ISBN 0-13-060774-6.OCLC 28888762.
  4. ^Box, George E. P.; Jenkins, Gwilym M.; Reinsel, Gregory C.; Ljung, Greta M. (2016).Time series analysis : forecasting and control (5th ed.). Hoboken, New Jersey: John Wiley & Sons, Incorporated. p. 53.ISBN 978-1-118-67492-5.OCLC 908107438.
  5. ^Shumway, Robert H. (2000).Time series analysis and its applications. David S. Stoffer. New York: Springer. p. 98.ISBN 0-387-98950-1.OCLC 42392178.
  6. ^Box, George; Jenkins, Gwilym M.; Reinsel, Gregory C. (1994).Time Series Analysis: Forecasting and Control (Third ed.). Prentice-Hall.ISBN 0130607746.
  7. ^Rosenblatt, Murray (2000).Gaussian and non-Gaussian linear time series and random fields. New York: Springer. p. 10.ISBN 0-387-98917-X.OCLC 42061096.
  8. ^Wei, William W. S. (1990).Time series analysis : univariate and multivariate methods. Redwood City, Calif.: Addison-Wesley Pub. pp. 242–243.ISBN 0-201-15911-2.OCLC 18166355.
  9. ^Missouri State University."Model Specification, Time Series Analysis"(PDF).
  10. ^Brockwell, P. J.; Davis, R. A. (2009).Time Series: Theory and Methods (2nd ed.). New York: Springer. p. 273.ISBN 9781441903198.
  11. ^Time series features in MathematicaArchived November 24, 2011, at theWayback Machine
  12. ^Hannan, Edward James (1970).Multiple time series. Wiley series in probability and mathematical statistics. New York: John Wiley and Sons.
  13. ^Whittle, P. (1951).Hypothesis Testing in Time Series Analysis. Almquist and Wicksell.Whittle, P. (1963).Prediction and Regulation. English Universities Press.ISBN 0-8166-1147-5.{{cite book}}:ISBN / Date incompatibility (help)
    Republished as:Whittle, P. (1983).Prediction and Regulation by Linear Least-Square Methods. University of Minnesota Press.ISBN 0-8166-1148-3.
  14. ^Hannan & Deistler (1988, p. 227):Hannan, E. J.; Deistler, Manfred (1988).Statistical theory of linear systems. Wiley series in probability and mathematical statistics. New York: John Wiley and Sons.
  15. ^ARIMA Modelling of Time Series, R documentation

Further reading

[edit]
Discrete time
Continuous time
Both
Fields and other
Time series models
Financial models
Actuarial models
Queueing models
Properties
Limit theorems
Inequalities
Tools
Disciplines
Continuous data
Center
Dispersion
Shape
Count data
Summary tables
Dependence
Graphics
Study design
Survey methodology
Controlled experiments
Adaptive designs
Observational studies
Statistical theory
Frequentist inference
Point estimation
Interval estimation
Testing hypotheses
Parametric tests
Specific tests
Goodness of fit
Rank statistics
Bayesian inference
Correlation
Regression analysis
Linear regression
Non-standard predictors
Generalized linear model
Partition of variance
Categorical
Multivariate
Time-series
General
Specific tests
Time domain
Frequency domain
Survival
Survival function
Hazard function
Test
Biostatistics
Engineering statistics
Social statistics
Spatial statistics
Retrieved from "https://en.wikipedia.org/w/index.php?title=Autoregressive_moving-average_model&oldid=1314428852"
Category:
Hidden categories:

[8]ページ先頭

©2009-2026 Movatter.jp