Movatterモバイル変換


[0]ホーム

URL:


Jump to content
WikipediaThe Free Encyclopedia
Search

Exponential smoothing

From Wikipedia, the free encyclopedia
Generates a forecast of future values of a time series

Exponential smoothing orexponential moving average (EMA) is arule of thumb technique forsmoothingtime series data using the exponentialwindow function. Whereas in thesimple moving average the past observations areweighted equally, exponential functions are used to assign exponentially decreasing weights over time. It is an easily learned and easily applied procedure for making some determination based on prior assumptions by the user, such as seasonality. Exponential smoothing is often used for analysis of time-series data.

Exponential smoothing is one of manywindow functions commonly applied to smooth data insignal processing, acting aslow-pass filters to remove high-frequencynoise. This method is preceded byPoisson's use of recursive exponential window functions in convolutions from the 19th century, as well asKolmogorov and Zurbenko's use of recursive moving averages from their studies of turbulence in the 1940s.

The raw data sequence is often represented by{xt}{\textstyle \{x_{t}\}} beginning at timet=0{\textstyle t=0}, and the output of the exponential smoothing algorithm is commonly written as{st}{\textstyle \{s_{t}\}}, which may be regarded as a best estimate of what the next value ofx{\textstyle x} will be. When the sequence of observations begins at timet=0{\textstyle t=0}, the simplest form of exponential smoothing is given by the following formulas:[1]

s0=x0st=αxt+(1α)st1,t>0{\displaystyle {\begin{aligned}s_{0}&=x_{0}\\s_{t}&=\alpha x_{t}+(1-\alpha )s_{t-1},\quad t>0\end{aligned}}}

whereα{\textstyle \alpha } is thesmoothing factor, and0<α<1{\textstyle 0<\alpha <1}. Ifst1{\textstyle s_{t-1}} is substituted intost{\textstyle s_{t}} continuously so that the formula ofst{\textstyle s_{t}} is fully expressed in terms of{xt}{\textstyle \{x_{t}\}}, thenexponentially decaying weighting factors on each raw dataxt{\textstyle x_{t}} is revealed, showing how exponential smoothing is named.

The simple exponential smoothing is not able to predict what would be observed att+m{\textstyle t+m} based on the raw data up tot{\textstyle t}, while thedouble exponential smoothing andtriple exponential smoothing can be used for the prediction due to the presence ofbt{\displaystyle b_{t}} as the sequence of best estimates of the linear trend.

Basic (simple) exponential smoothing

[edit]

The use of the exponential window function is first attributed toPoisson[2] as an extension of a numerical analysis technique from the 17th century, and later adopted by thesignal processing community in the 1940s. Here, exponential smoothing is the application of the exponential, or Poisson,window function. Exponential smoothing was suggested in the statistical literature without citation to previous work byRobert Goodell Brown in 1956,[3] and expanded byCharles C. Holt in 1957.[4] The formulation below, which is the one commonly used, is attributed to Brown and is known as "Brown’s simple exponential smoothing".[5] All the methods of Holt, Winters, and Brown may be seen as a simple application ofrecursive filtering, first found in the 1940s[2] to convertfinite impulse response (FIR) filters toinfinite impulse response filters.

The simplest form of exponential smoothing is given by the formula:

st=αxt+(1α)st1,{\displaystyle s_{t}=\alpha x_{t}+(1-\alpha )s_{t-1}\,,}

whereα{\displaystyle \alpha } is thesmoothing factor, with0α1{\displaystyle 0\leq \alpha \leq 1}. In other words, the smoothed statisticst{\displaystyle s_{t}} is a simple weighted average of the currentobservationxt{\displaystyle x_{t}} and the previous smoothed statisticst1{\displaystyle s_{t-1}}. Simple exponential smoothing is easily applied, and it produces a smoothed statistic as soon as two observations are available. The termsmoothing factor applied toα{\displaystyle \alpha } here is something of a misnomer, as larger values ofα{\displaystyle \alpha } actually reduce the level of smoothing, and in the limiting case withα{\displaystyle \alpha } = 1 the smoothing output series is just the current observation. Values ofα{\displaystyle \alpha } close to 1 have less of a smoothing effect and give greater weight to recent changes in the data, while values ofα{\displaystyle \alpha } closer to 0 have a greater smoothing effect and are less responsive to recent changes. In the limiting case withα{\displaystyle \alpha } = 0, the output series is just flat or a constant as the observationx0{\textstyle x_{0}} at the beginning of the smoothening processt=0{\textstyle t=0}.

The method for choosingα{\displaystyle \alpha } must be decided by the modeler. Sometimes the statistician's judgment is used to choose an appropriate factor. Alternatively, a statistical technique may be used tooptimize the value ofα{\displaystyle \alpha }. For example, themethod of least squares might be used to determine the value ofα{\displaystyle \alpha } for which the sum of the quantities(stxt+1)2{\displaystyle (s_{t}-x_{t+1})^{2}} is minimized.[6]

Unlike some other smoothing methods, such as the simple moving average, this technique does not require any minimum number of observations to be made before it begins to produce results. In practice, however, a "good average" will not be achieved until several samples have been averaged together; for example, a constant signal will take approximately3/α{\displaystyle 3/\alpha } stages to reach 95% of the actual value. To accurately reconstruct the original signal without information loss, all stages of the exponential moving average must also be available, because older samples decay in weight exponentially. This is in contrast to a simple moving average, in which some samples can be skipped without as much loss of information due to the constant weighting of samples within the average. If a known number of samples will be missed, one can adjust a weighted average for this as well, by giving equal weight to the new sample and all those to be skipped.

This simple form of exponential smoothing is also known as anexponentially weighted moving average (EWMA). Technically it can also be classified as anautoregressive integrated moving average (ARIMA) (0,1,1) model with no constant term.[7]

Time constant

[edit]

Thetime constant of an exponential moving average is the amount of time for the smoothed response of aunit step function to reach11/e63.2%{\displaystyle 1-1/e\approx 63.2\,\%} of the original signal. The relationship between this time constant,τ{\displaystyle \tau }, and the smoothing factor,α{\displaystyle \alpha }, is given by the following formula:

α=1eΔT/τ{\displaystyle \alpha =1-e^{-\Delta T/\tau }}, thusτ=ΔTln(1α){\displaystyle \tau =-{\frac {\Delta T}{\ln(1-\alpha )}}}

whereΔT{\displaystyle \Delta T} is the sampling time interval of the discrete time implementation. If the sampling time is fast compared to the time constant (ΔTτ{\displaystyle \Delta T\ll \tau }) then, by usingthe Taylor expansion of the exponential function,

αΔTτ{\displaystyle \alpha \approx {\frac {\Delta T}{\tau }}}, thusτΔTα{\displaystyle \tau \approx {\frac {\Delta T}{\alpha }}}

Choosing the initial smoothed value

[edit]

Note that in the definition above,s0{\displaystyle s_{0}} (the initial output of the exponential smoothing algorithm) is being initialized tox0{\displaystyle x_{0}} (the initial raw data or observation). Because exponential smoothing requires that, at each stage, we have the previous forecastst1{\displaystyle s_{t-1}}, it is not obvious how to get the method started. We could assume that the initial forecast is equal to the initial value of demand; however, this approach has a serious drawback. Exponential smoothing puts substantial weight on past observations, so the initial value of demand will have an unreasonably large effect on early forecasts. This problem can be overcome by allowing the process to evolve for a reasonable number of periods (10 or more) and using the average of the demand during those periods as the initial forecast. There are many other ways of setting this initial value, but it is important to note that the smaller the value ofα{\displaystyle \alpha }, the more sensitive your forecast will be on the selection of this initial smoother values0{\displaystyle s_{0}}.[8][9]

Optimization

[edit]

For every exponential smoothing method, we also need to choose the value for the smoothing parameters. For simple exponential smoothing, there is only one smoothing parameter (α), but for the methods that follow there are usually more than one smoothing parameter.

There are cases where the smoothing parameters may be chosen in a subjective manner – the forecaster specifies the value of the smoothing parameters based on previous experience. However, a more robust and objective way to obtain values of the unknown parameters included in any exponential smoothing method is to estimate them from the observed data.

The unknown parameters and the initial values for any exponential smoothing method can be estimated by minimizing thesum of squared errors (SSE). The errors are specified aset=yty^tt1{\textstyle e_{t}=y_{t}-{\hat {y}}_{t\mid t-1}} fort=1,,T{\textstyle t=1,\ldots ,T} (the one-step-ahead within-sample forecast errors) whereyt{\textstyle y_{t}} andy^tt1{\textstyle {\hat {y}}_{t\mid t-1}} are a variable to be predicted att{\displaystyle t} and a variable as the prediction result att{\displaystyle t} (based on the previous data or prediction), respectively. Hence, we find the values of the unknown parameters and the initial values that minimize

SSE=t=1T(yty^tt1)2=t=1Tet2{\displaystyle {\text{SSE}}=\sum _{t=1}^{T}(y_{t}-{\hat {y}}_{t\mid t-1})^{2}=\sum _{t=1}^{T}e_{t}^{2}}[10]

Unlike the regression case (where we have formulae to directly compute the regression coefficients which minimize the SSE) this involves a non-linear minimization problem, and we need to use anoptimization tool to perform this.

"Exponential" naming

[edit]

The nameexponential smoothing is attributed to the use of the exponential function as the filterimpulse response in theconvolution.

By direct substitution of the defining equation for simple exponential smoothing back into itself we find that

st=αxt+(1α)st1=αxt+α(1α)xt1+(1α)2st2=α[xt+(1α)xt1+(1α)2xt2+(1α)3xt3++(1α)t1x1]+(1α)tx0.{\displaystyle {\begin{aligned}s_{t}&=\alpha x_{t}+(1-\alpha )s_{t-1}\\[3pt]&=\alpha x_{t}+\alpha (1-\alpha )x_{t-1}+(1-\alpha )^{2}s_{t-2}\\[3pt]&=\alpha \left[x_{t}+(1-\alpha )x_{t-1}+(1-\alpha )^{2}x_{t-2}+(1-\alpha )^{3}x_{t-3}+\cdots +(1-\alpha )^{t-1}x_{1}\right]+(1-\alpha )^{t}x_{0}.\end{aligned}}}

In other words, as time passes the smoothed statisticst{\displaystyle s_{t}} becomes the weighted average of a greater and greater number of the past observationsst1,,stn,{\displaystyle s_{t-1},\ldots ,s_{t-n},\ldots }, and the weights assigned to previous observations are proportional to the terms of the geometric progression

1,(1α),(1α)2,,(1α)n,{\displaystyle 1,(1-\alpha ),(1-\alpha )^{2},\ldots ,(1-\alpha )^{n},\ldots }

Ageometric progression is the discrete version of anexponential function, so this is where the name for this smoothing method originated according toStatistics lore.

Comparison with moving average

[edit]

Exponential smoothing and moving average have similar defects of introducing a lag relative to the input data. While this can be corrected by shifting the result by half the window length for a symmetrical kernel, such as a moving average or gaussian, this approach is not possible for exponential smoothing since it is anIIR filter and therefore has an asymmetric kernel and frequency-dependentgroup delay. This means each constituent frequency is shifted by a different amount and therefore, there is no single number of samples that can be used to shift the output signal to account for the lag.

Both filters also both have roughly the same distribution of forecast error whenα = 2/(k + 1) wherek is the number of past data points in consideration of moving average. They differ in that exponential smoothing takes into account all past data, whereas moving average only takes into accountk past data points. Computationally speaking, they also differ in that moving average requires that the pastk data points, or the data point at lagk + 1 plus the most recent forecast value, to be kept, whereas exponential smoothing only needs the most recent forecast value to be kept.[11]

In thesignal processing literature, the use of non-causal (symmetric) filters is commonplace, and the exponentialwindow function is broadly used in this fashion, but a different terminology is used: exponential smoothing is equivalent to a first-orderinfinite-impulse response (IIR) filter and moving average is equivalent to afinite impulse response filter with equal weighting factors.

Double exponential smoothing (Holt linear)

[edit]

Simple exponential smoothing does not do well when there is atrend in the data.[1] In such situations, several methods were devised under the name "double exponential smoothing" or "second-order exponential smoothing," which is the recursive application of an exponential filter twice, thus being termed "double exponential smoothing". The basic idea behind double exponential smoothing is to introduce a term to take into account the possibility of a series exhibiting some form of trend. This slope component is itself updated via exponential smoothing.

One method works as follows:[12]

Again, the raw data sequence of observations is represented byxt{\displaystyle x_{t}}, beginning at timet=0{\displaystyle t=0}. We usest{\displaystyle s_{t}} to represent the smoothed value for timet{\displaystyle t}, andbt{\displaystyle b_{t}} is our best estimate of the trend at timet{\displaystyle t}. The output of the algorithm is now written asFt+m{\displaystyle F_{t+m}}, an estimate of the value ofxt+m{\displaystyle x_{t+m}} at timem>0{\displaystyle m>0} based on the raw data up to timet{\displaystyle t}. Double exponential smoothing is given by the formulas

s0=x0b0=x1x0{\displaystyle {\begin{aligned}s_{0}&=x_{0}\\b_{0}&=x_{1}-x_{0}\\\end{aligned}}}

and fort>0{\displaystyle t>0} by

st=αxt+(1α)(st1+bt1)bt=β(stst1)+(1β)bt1{\displaystyle {\begin{aligned}s_{t}&=\alpha x_{t}+(1-\alpha )(s_{t-1}+b_{t-1})\\b_{t}&=\beta (s_{t}-s_{t-1})+(1-\beta )b_{t-1}\\\end{aligned}}}

whereα{\displaystyle \alpha } (0α1{\displaystyle 0\leq \alpha \leq 1}) is thedata smoothing factor, andβ{\displaystyle \beta } (0β1{\displaystyle 0\leq \beta \leq 1}) is thetrend smoothing factor.

To forecast beyondxt{\displaystyle x_{t}} is given by the following approximation:

Ft+m=st+mbt{\displaystyle F_{t+m}=s_{t}+m\cdot b_{t}}.

Setting the initial valueb{\displaystyle b} is a matter of preference. An option other than the one listed above isxnx0n{\textstyle {\frac {x_{n}-x_{0}}{n}}} for somen{\displaystyle n}.

Note thatF0 is undefined (there is no estimation for time 0), and according to the definitionF1=s0+b0, which is well defined, thus further values can be evaluated.

A second method, referred to as either Brown's linear exponential smoothing (LES) or Brown's double exponential smoothing, has only one smoothing factor,α{\displaystyle \alpha }:[13]

s0=x0s0=x0st=αxt+(1α)st1st=αst+(1α)st1Ft+m=at+mbt,{\displaystyle {\begin{aligned}s'_{0}&=x_{0}\\s''_{0}&=x_{0}\\s'_{t}&=\alpha x_{t}+(1-\alpha )s'_{t-1}\\s''_{t}&=\alpha s'_{t}+(1-\alpha )s''_{t-1}\\F_{t+m}&=a_{t}+mb_{t},\end{aligned}}}

whereat, the estimated level at timet, andbt, the estimated trend at timet, are given by

at=2ststbt=α1α(stst).{\displaystyle {\begin{aligned}a_{t}&=2s'_{t}-s''_{t}\\[5pt]b_{t}&={\frac {\alpha }{1-\alpha }}(s'_{t}-s''_{t}).\end{aligned}}}

Triple exponential smoothing (Holt–Winters)

[edit]

Triple exponential smoothing applies exponential smoothing three times, which is commonly used when there are three high frequency signals to be removed from atime series under study. There are different types of seasonality: 'multiplicative' and 'additive' in nature, much like addition and multiplication are basic operations in mathematics.

If every month of December we sell 10,000 more apartments than we do in November the seasonality isadditive in nature. However, if we sell 10% more apartments in the summer months than we do in the winter months the seasonality ismultiplicative in nature. Multiplicative seasonality can be represented as a constant factor, not an absolute amount.[14]

Triple exponential smoothing was first suggested by Holt's student, Peter Winters, in 1960 after reading a signal processing book from the 1940s on exponential smoothing.[15] Holt's novel idea was to repeat filtering an odd number of times greater than 1 and less than 5, which was popular with scholars of previous eras.[15] While recursive filtering had been used previously, it was applied twice and four times to coincide with theHadamard conjecture, while triple application required more than double the operations of singular convolution. The use of a triple application is considered arule of thumb technique, rather than one based on theoretical foundations and has often been over-emphasized by practitioners. Suppose we have a sequence of observationsxt,{\displaystyle x_{t},} beginning at timet=0{\displaystyle t=0} with a cycle of seasonal change of lengthL{\displaystyle L}.

The method calculates a trend line for the data as well as seasonal indices that weight the values in the trend line based on where that time point falls in the cycle of lengthL{\displaystyle L}.

Letst{\displaystyle s_{t}} represent the smoothed value of the constant part for timet{\displaystyle t},bt{\displaystyle b_{t}} is the sequence of best estimates of the linear trend that are superimposed on the seasonal changes, andct{\displaystyle c_{t}} is the sequence of seasonal correction factors. We wish to estimatect{\displaystyle c_{t}} at every timet{\displaystyle t}modL{\displaystyle L} in the cycle that the observations take on. As a rule of thumb, a minimum of two full seasons (or2L{\displaystyle 2L} periods) of historical data is needed to initialize a set of seasonal factors.

The output of the algorithm is again written asFt+m{\displaystyle F_{t+m}}, an estimate of the value ofxt+m{\displaystyle x_{t+m}} at timet+m>0{\displaystyle t+m>0} based on the raw data up to timet{\displaystyle t}. Triple exponential smoothing with multiplicative seasonality is given by the formulas[1]

s0=x0st=αxtctL+(1α)(st1+bt1)bt=β(stst1)+(1β)bt1ct=γxtst+(1γ)ctLFt+m=(st+mbt)ctL+1+(m1)modL,{\displaystyle {\begin{aligned}s_{0}&=x_{0}\\[5pt]s_{t}&=\alpha {\frac {x_{t}}{c_{t-L}}}+(1-\alpha )(s_{t-1}+b_{t-1})\\[5pt]b_{t}&=\beta (s_{t}-s_{t-1})+(1-\beta )b_{t-1}\\[5pt]c_{t}&=\gamma {\frac {x_{t}}{s_{t}}}+(1-\gamma )c_{t-L}\\[5pt]F_{t+m}&=(s_{t}+mb_{t})c_{t-L+1+(m-1){\bmod {L}}},\end{aligned}}}

whereα{\displaystyle \alpha } (0α1{\displaystyle 0\leq \alpha \leq 1}) is thedata smoothing factor,β{\displaystyle \beta } (0β1{\displaystyle 0\leq \beta \leq 1}) is thetrend smoothing factor, andγ{\displaystyle \gamma } (0γ1{\displaystyle 0\leq \gamma \leq 1}) is theseasonal change smoothing factor.

The general formula for the initial trend estimateb{\displaystyle b} is

b0=1L(xL+1x1L+xL+2x2L++xL+LxLL){\displaystyle {\begin{aligned}b_{0}&={\frac {1}{L}}\left({\frac {x_{L+1}-x_{1}}{L}}+{\frac {x_{L+2}-x_{2}}{L}}+\cdots +{\frac {x_{L+L}-x_{L}}{L}}\right)\end{aligned}}}.

Setting the initial estimates for the seasonal indicesci{\displaystyle c_{i}} fori=1,2,,L{\displaystyle i=1,2,\ldots ,L} is a bit more involved. IfN{\displaystyle N} is the number of complete cycles present in your data, then

ci=1Nj=1NxL(j1)+iAjfor i=1,2,,L{\displaystyle c_{i}={\frac {1}{N}}\sum _{j=1}^{N}{\frac {x_{L(j-1)+i}}{A_{j}}}\quad {\text{for }}i=1,2,\ldots ,L}

where

Aj=k=1LxL(j1)+kLfor j=1,2,,N{\displaystyle A_{j}={\frac {\sum _{k=1}^{L}x_{L(j-1)+k}}{L}}\quad {\text{for }}j=1,2,\ldots ,N}.

Note thatAj{\displaystyle A_{j}} is the average value ofx{\displaystyle x} in thejth{\displaystyle j^{\text{th}}} cycle of your data.

This results in

ci=1Nj=1NxL(j1)+i1Lk=1LxL(j1)+k{\displaystyle c_{i}={\frac {1}{N}}\sum _{j=1}^{N}{\frac {x_{L(j-1)+i}}{{\frac {1}{L}}\sum _{k=1}^{L}x_{L(j-1)+k}}}}

Triple exponential smoothing with additive seasonality is given by[citation needed]

s0=x0st=α(xtctL)+(1α)(st1+bt1)bt=β(stst1)+(1β)bt1ct=γ(xtst1bt1)+(1γ)ctLFt+m=st+mbt+ctL+1+(m1)modL.{\displaystyle {\begin{aligned}s_{0}&=x_{0}\\s_{t}&=\alpha (x_{t}-c_{t-L})+(1-\alpha )(s_{t-1}+b_{t-1})\\b_{t}&=\beta (s_{t}-s_{t-1})+(1-\beta )b_{t-1}\\c_{t}&=\gamma (x_{t}-s_{t-1}-b_{t-1})+(1-\gamma )c_{t-L}\\F_{t+m}&=s_{t}+mb_{t}+c_{t-L+1+(m-1){\bmod {L}}}.\\\end{aligned}}}

Implementations in statistics packages

[edit]
  • R: the HoltWinters function in the stats package[16] and ets function in the forecast package[17] (a more complete implementation, generally resulting in a better performance[18]).
  • Python: the holtwinters module of the statsmodels package allow for simple, double and triple exponential smoothing.
  • IBMSPSS includes Simple, Simple Seasonal, Holt's Linear Trend, Brown's Linear Trend, Damped Trend, Winters' Additive, and Winters' Multiplicative in the Time-Series modeling procedure within its Statistics and Modeler statistical packages. The default Expert Modeler feature evaluates all seven exponential smoothing models and ARIMA models with a range of nonseasonal and seasonalp,d, andq values, and selects the model with the lowestBayesian Information Criterion statistic.
  • Stata: tssmooth command[19]
  • LibreOffice 5.2[20]
  • Microsoft Excel 2016[21]
  • Julia: TrendDecomposition.jl package[22] implements simple and double exponential smoothing and Holts-Winters forecasting procedure.

See also

[edit]

Notes

[edit]
  1. ^abc"NIST/SEMATECH e-Handbook of Statistical Methods". NIST. Retrieved23 May 2010.
  2. ^abOppenheim, Alan V.; Schafer, Ronald W. (1975).Digital Signal Processing.Prentice Hall. p. 5.ISBN 0-13-214635-5.
  3. ^Brown, Robert G. (1956).Exponential Smoothing for Predicting Demand. Cambridge, Massachusetts: Arthur D. Little Inc. p. 15.
  4. ^Holt, Charles C. (1957). "Forecasting Trends and Seasonal by Exponentially Weighted Averages".Office of Naval Research Memorandum.52. reprinted inHolt, Charles C. (January–March 2004). "Forecasting Trends and Seasonal by Exponentially Weighted Averages".International Journal of Forecasting.20 (1):5–10.doi:10.1016/j.ijforecast.2003.09.015.
  5. ^Brown, Robert Goodell (1963).Smoothing Forecasting and Prediction of Discrete Time Series. Englewood Cliffs, NJ: Prentice-Hall.
  6. ^"NIST/SEMATECH e-Handbook of Statistical Methods, 6.4.3.1. Single Exponential Smoothing". NIST. Retrieved5 July 2017.
  7. ^Nau, Robert."Averaging and Exponential Smoothing Models". Retrieved26 July 2010.
  8. ^"Production and Operations Analysis" Nahmias. 2009.
  9. ^Čisar, P., & Čisar, S. M. (2011). "Optimization methods of EWMA statistics."Acta Polytechnica Hungarica, 8(5), 73–87. Page 78.
  10. ^7.1 Simple exponential smoothing | Forecasting: Principles and Practice.
  11. ^Nahmias, Steven; Olsen, Tava Lennon.Production and Operations Analysis (7th ed.). Waveland Press. p. 53.ISBN 9781478628248.
  12. ^"6.4.3.3. Double Exponential Smoothing".itl.nist.gov. Retrieved25 September 2011.
  13. ^"Averaging and Exponential Smoothing Models".duke.edu. Retrieved25 September 2011.
  14. ^Kalehar, Prajakta S."Time series Forecasting using Holt–Winters Exponential Smoothing"(PDF). Retrieved23 June 2014.
  15. ^abWinters, P. R. (April 1960). "Forecasting Sales by Exponentially Weighted Moving Averages".Management Science.6 (3):324–342.doi:10.1287/mnsc.6.3.324.
  16. ^"R: Holt–Winters Filtering".stat.ethz.ch. Retrieved5 June 2016.
  17. ^"ets {forecast} | inside-R | A Community Site for R".inside-r.org. Archived fromthe original on 16 July 2016. Retrieved5 June 2016.
  18. ^"Comparing HoltWinters() and ets()".Hyndsight. 29 May 2011. Retrieved5 June 2016.
  19. ^tssmooth in Stata manual
  20. ^"LibreOffice 5.2: Release Notes – the Document Foundation Wiki".
  21. ^"Excel 2016 Forecasting Functions | Real Statistics Using Excel".
  22. ^TrendDecomposition.jl Julia implementation of exponential smoothing and Holt-Winters forecasting procedure

External links

[edit]
Continuous data
Center
Dispersion
Shape
Count data
Summary tables
Dependence
Graphics
Study design
Survey methodology
Controlled experiments
Adaptive designs
Observational studies
Statistical theory
Frequentist inference
Point estimation
Interval estimation
Testing hypotheses
Parametric tests
Specific tests
Goodness of fit
Rank statistics
Bayesian inference
Correlation
Regression analysis (see alsoTemplate:Least squares and regression analysis
Linear regression
Non-standard predictors
Generalized linear model
Partition of variance
Categorical
Multivariate
Time-series
General
Specific tests
Time domain
Frequency domain
Survival
Survival function
Hazard function
Test
Biostatistics
Engineering statistics
Social statistics
Spatial statistics
Quantitativeforecasting methods
Retrieved from "https://en.wikipedia.org/w/index.php?title=Exponential_smoothing&oldid=1311619208"
Category:
Hidden categories:

[8]ページ先頭

©2009-2025 Movatter.jp