Skip to main content

Exponential smoothing Contents Background[edit] Basic exponential smoothing[edit] Double exponential smoothing[edit] Triple exponential smoothing[edit] Implementations in statistics packages[edit] See also[edit] Notes[edit] External links[edit] Navigation menu"NIST/SEMATECH e-Handbook of Statistical Methods"Exponential Smoothing for Predicting Demand"Forecasting Trends and Seasonal by Exponentially Weighted Averages"Smoothing Forecasting and Prediction of Discrete Time Series"NIST/SEMATECH e-Handbook of Statistical Methods, 6.4.3.1. Single Exponential Smoothing""Averaging and Exponential Smoothing Models""Model: Second-Order Exponential Smoothing""Time series Forecasting using Holt-Winters Exponential Smoothing""6.4.3.3. Double Exponential Smoothing""Averaging and Exponential Smoothing Models""Time series Forecasting using Holt-Winters Exponential Smoothing""Forecasting Sales by Exponentially Weighted Moving Averages"10.1287/mnsc.6.3.324"R: Holt-Winters Filtering""ets forecast | inside-R | A Community Site for R"the original"Comparing HoltWinters() and ets()"tssmoothLecture notes on exponential smoothing (Robert Nau, Duke University)Data SmoothingThe Holt-Winters Approach to Exponential Smoothing: 50 Years Old and Going StrongAlgorithms for Unevenly Spaced Time Series: Moving Averages and Other Rolling Operatorsee

Time series


rule of thumbtime serieswindow functionsimple moving averagewindow functionssignal processinglow-pass filtersnoisePoissonKolmogorov and Zurbenko's use of recursive moving averageswindow functionsPoissonsignal processingwindow functionRobert Goodell BrownCharles C. HoltFIR filtersIIR filtersmethod of least squaresexponentially weighted moving averageautoregressive integrated moving averagetime constantsum of squared errorsoptimizationgeometric progressionexponential functionStatisticssignal processingwindow functionIIR filterFIR filtertrendHadamard conjecturerule of thumb












Exponential smoothing




From Wikipedia, the free encyclopedia






Jump to navigation
Jump to search


Generates a forecast of future values of a time series

Exponential smoothing is a rule of thumb technique for smoothing time series data using the exponential window function. Whereas in the simple moving average the past observations are weighted equally, exponential functions are used to assign exponentially decreasing weights over time. It is an easily learned and easily applied procedure for making some determination based on prior assumptions by the user, such as seasonality. Exponential smoothing is often used for analysis of time-series data.


Exponential smoothing is one of many window functions commonly applied to smooth data in signal processing, acting as low-pass filters to remove high frequency noise. This method is preceded by Poisson's use of recursive exponential window functions in convolutions from the 19th century, as well as Kolmogorov and Zurbenko's use of recursive moving averages from their studies of turbulence in the 1940s.


The raw data sequence is often represented by xtdisplaystyle x_t beginning at time t=0displaystyle t=0, and the output of the exponential smoothing algorithm is commonly written as stdisplaystyle s_t, which may be regarded as a best estimate of what the next value of xdisplaystyle x will be. When the sequence of observations begins at time t=0displaystyle t=0, the simplest form of exponential smoothing is given by the formulas:[1]


s0=x0st=αxt+(1−α)st−1, t>0displaystyle beginaligneds_0&=x_0\s_t&=alpha x_t+(1-alpha )s_t-1, t>0endaligned


where αdisplaystyle alpha is the smoothing factor, and 0<α<1displaystyle 0<alpha <1.




Contents





  • 1 Background


  • 2 Basic exponential smoothing

    • 2.1 Time Constant


    • 2.2 Choosing the initial smoothed value


    • 2.3 Optimization


    • 2.4 “Exponential” naming


    • 2.5 Comparison with moving average



  • 3 Double exponential smoothing


  • 4 Triple exponential smoothing


  • 5 Implementations in statistics packages


  • 6 See also


  • 7 Notes


  • 8 External links




Background[edit]


Exponential smoothing is based on the use of window functions to smooth time series data.



Basic exponential smoothing[edit]


The use of the exponential window function is first attributed to Poisson[2] as an extension of a numerical analysis technique from the 17th century, and later adopted by the signal processing community in the 1940s. Here, exponential smoothing is the application of the exponential, or Poisson, window function. Exponential smoothing was first suggested in the statistical literature without citation to previous work by Robert Goodell Brown in 1956,[3] and then expanded by Charles C. Holt in 1957.[4] The formulation below, which is the one commonly used, is attributed to Brown and is known as “Brown’s simple exponential smoothing”.[5] All the methods of Holt, Winters and Brown may be seen as a simple application of recursive filtering, first found in the 1940s[2] to convert FIR filters to IIR filters.


The simplest form of exponential smoothing is given by the formula:



st=α⋅xt+(1−α)⋅st−1=st−1+α⋅(xt−st−1)displaystyle s_t=alpha cdot x_t+(1-alpha )cdot s_t-1=s_t-1+alpha cdot (x_t-s_t-1).

where α is the smoothing factor, and 0 < α < 1. In other words, the smoothed statistic st' is a simple weighted average of the current observation xt and the previous smoothed statistic st−1. The term smoothing factor applied to α here is something of a misnomer, as larger values of α actually reduce the level of smoothing, and in the limiting case with α = 1 the output series is just the current observation.
Simple exponential smoothing is easily applied, and it produces a smoothed statistic as soon as two observations are available.


Values of α close to one have less of a smoothing effect and give greater weight to recent changes in the data, while values of α closer to zero have a greater smoothing effect and are less responsive to recent changes. There is no formally correct procedure for choosing α. Sometimes the statistician's judgment is used to choose an appropriate factor. Alternatively, a statistical technique may be used to optimize the value of α. For example, the method of least squares might be used to determine the value of α for which the sum of the quantities (st−xt+1)2displaystyle (s_t-x_t+1)^2 is minimized.[6]


Unlike some other smoothing methods, such as the simple moving average, this technique does not require any minimum number of observations to be made before it begins to produce results. In practice, however, a “good average” will not be achieved until several samples have been averaged together; for example, a constant signal will take approximately 3/α stages to reach 95% of the actual value. To accurately reconstruct the original signal without information loss all stages of the exponential moving average must also be available, because older samples decay in weight exponentially. This is in contrast to a simple moving average, in which some samples can be skipped without as much loss of information due to the constant weighting of samples within the average. If a known number of samples will be missed, one can adjust a weighted average for this as well, by giving equal weight to the new sample and all those to be skipped.


This simple form of exponential smoothing is also known as an exponentially weighted moving average (EWMA). Technically it can also be classified as an autoregressive integrated moving average (ARIMA) (0,1,1) model with no constant term.[7]



Time Constant[edit]


The time constant of an exponential moving average is the amount of time for the smoothed response of a unit set function to reach 1−1/e≈63.2%displaystyle 1-1/eapprox 63.2,% of the original signal. The relationship between this time constant, τdisplaystyle tau , and the smoothing factor, αdisplaystyle alpha , is given by the formula:


α=1−e−ΔTτdisplaystyle alpha =1-e^-Delta T over tau

Where ΔTdisplaystyle Delta T is the sampling time interval of the discrete time implementation. If the sampling time is fast compared to the time constant (ΔT≪τdisplaystyle Delta Tll tau ) then


α≈ΔTτdisplaystyle alpha approx Delta T over tau


Choosing the initial smoothed value[edit]


Note that in the definition above, s0 is being initialized to x0. Because exponential smoothing requires that at each stage we have the previous forecast, it is not obvious how to get the method started. We could assume that the initial forecast is equal to the initial value of demand; however, this approach has a serious drawback. Exponential smoothing puts substantial weight on past observations, so the initial value of demand will have an unreasonably large effect on early forecasts. This problem can be overcome by allowing the process to evolve for a reasonable number of periods (10 or more) and using the average of the demand during those periods as the initial forecast. There are many other ways of setting this initial value, but it is important to note that the smaller the value of α, the more sensitive your forecast will be on the selection of this initial smoother value s1.[8]



Optimization[edit]


For every exponential smoothing method we also need to choose the value for the smoothing parameters. For simple exponential smoothing, there is only one smoothing parameter (α), but for the methods that follow there is usually more than one smoothing parameter.


There are cases where the smoothing parameters may be chosen in a subjective manner — the forecaster specifies the value of the smoothing parameters based on previous experience. However, a more robust and objective way to obtain values for the unknown parameters included in any exponential smoothing method is to estimate them from the observed data.


The unknown parameters and the initial values for any exponential smoothing method can be estimated by minimizing the sum of squared errors (SSE). The errors are specified as et=yt−y^t|t−1displaystyle e_t=y_t-hat y_t for t=1,...,T (the one-step-ahead within-sample forecast errors). Hence we find the values of the unknown parameters and the initial values that minimize


SSE=∑t=1T(yt−y^t|t−1)2=∑t=1Tet2displaystyle SSE=sum _t=1^T(y_t-hat y_t)^2=sum _t=1^Te_t^2 [9]


Unlike the regression case (where we have formulae to directly compute the regression coefficients which minimize the SSE) this involves a non-linear minimization problem and we need to use an optimization tool to perform this.



“Exponential” naming[edit]


The name 'exponential smoothing' is attributed to the use of the exponential window function during convolution. It is no longer attributed to Holt, Winters & Brown.


By direct substitution of the defining equation for simple exponential smoothing back into itself we find that


st=αxt+(1−α)st−1=αxt+α(1−α)xt−1+(1−α)2st−2=α[xt+(1−α)xt−1+(1−α)2xt−2+(1−α)3xt−3+⋯+(1−α)t−1x1]+(1−α)tx0.displaystyle beginaligneds_t&=alpha x_t+(1-alpha )s_t-1\[3pt]&=alpha x_t+alpha (1-alpha )x_t-1+(1-alpha )^2s_t-2\[3pt]&=alpha left[x_t+(1-alpha )x_t-1+(1-alpha )^2x_t-2+(1-alpha )^3x_t-3+cdots +(1-alpha )^t-1x_1right]+(1-alpha )^tx_0.endaligned

In other words, as time passes the smoothed statistic st becomes the weighted average of a greater and greater number of the past observations xtn, and the weights assigned to previous observations are in general proportional to the terms of the geometric progression 1, (1 − α), (1 − α)2, (1 − α)3, .... A geometric progression is the discrete version of an exponential function, so this is where the name for this smoothing method originated according to Statistics lore.



Comparison with moving average[edit]


Exponential smoothing and moving average have similar defects of introducing a lag relative to the input data. While this can be corrected by shifting the result by half the window length for a symmetrical kernel, such as a moving average or gaussian, it is unclear how appropriate this would be for exponential smoothing. They also both have roughly the same distribution of forecast error when α = 2/(k+1). They differ in that exponential smoothing takes into account all past data, whereas moving average only takes into account k past data points. Computationally speaking, they also differ in that moving average requires that the past k data points, or the data point at lag k+1 plus the most recent forecast value, to be kept, whereas exponential smoothing only needs the most recent forecast value to be kept.[10]


In the signal processing literature, the use of non-causal (symmetric) filters is commonplace, and the exponential window function is broadly used in this fashion, but a different terminology is used: exponential smoothing is equivalent to a first-order Infinite Impulse Response or IIR filter and moving average is equivalent to a Finite Impulse Response or FIR filter with equal weighting factors.



Double exponential smoothing[edit]


Simple exponential smoothing does not do well when there is a trend in the data, which is inconvenient.[1] In such situations, several methods were devised under the name "double exponential smoothing" or "second-order exponential smoothing.", which is the recursive application of an exponential filter twice, thus being termed "double exponential smoothing". This nomenclature is similar to quadruple exponential smoothing, which also references its recursion depth.[11]
The basic idea behind double exponential smoothing is to introduce a term to take into account the possibility of a series
exhibiting some form of trend. This slope component is itself updated via exponential smoothing.


One method, sometimes referred to as "Holt-Winters double exponential smoothing"[12] works as follows:[13]


Again, the raw data sequence of observations is represented by xt, beginning at time t = 0. We use st to represent the smoothed value for time t, and bt is our best estimate of the trend at time t. The output of the algorithm is now written as Ft+m, an estimate of the value of x at time t + m for m > 0 based on the raw data up to time t. Double exponential smoothing is given by the formulas


s1=x1b1=x1−x0displaystyle beginaligneds_1&=x_1\b_1&=x_1-x_0\endaligned

And for t > 1 by


st=αxt+(1−α)(st−1+bt−1)bt=β(st−st−1)+(1−β)bt−1displaystyle beginaligneds_t&=alpha x_t+(1-alpha )(s_t-1+b_t-1)\b_t&=beta (s_t-s_t-1)+(1-beta )b_t-1\endaligned

where α is the data smoothing factor, 0 < α < 1, and β is the trend smoothing factor, 0 < β < 1.


To forecast beyond xt


Ft+m=st+mbtdisplaystyle beginalignedF_t+m&=s_t+mb_tendaligned

Setting the initial value b0 is a matter of preference. An option other than the one listed above is (xn - x0)/n for some n > 1.


Note that F0 is undefined (there is no estimation for time 0), and according to the definition F1=s0+b0, which is well defined, thus further values can be evaluated.


A second method, referred to as either Brown's linear exponential smoothing (LES) or Brown's double exponential smoothing works as follows.[14]


s0′=x0s0″=x0st′=αxt+(1−α)st−1′st″=αst′+(1−α)st−1″Ft+m=at+mbt,displaystyle beginaligneds'_0&=x_0\s''_0&=x_0\s'_t&=alpha x_t+(1-alpha )s'_t-1\s''_t&=alpha s'_t+(1-alpha )s''_t-1\F_t+m&=a_t+mb_t,endaligned

where at, the estimated level at time t and bt, the estimated trend at time t are:


at=2st′−st″bt=α1−α(st′−st″).displaystyle beginaligneda_t&=2s'_t-s''_t\b_t&=frac alpha 1-alpha (s'_t-s''_t).endaligned


Triple exponential smoothing[edit]


Triple exponential smoothing applies exponential smoothing three times, which is commonly used when there are three high frequency signals to be removed from a time series under study. There are different types of seasonality: 'multiplicative' and 'additive' in nature, much like addition and multiplication are basic operations in mathematics.


If every month of December we sell 10,000 more apartments than we do in November the seasonality is additive in nature. However, if we sell 10% more apartments in the summer months than we do in the winter months the seasonality is multiplicative in nature. Multiplicative seasonality can be represented as a constant factor, not an absolute amount.
[15]


Triple exponential smoothing was first suggested by Holt's student, Peter Winters, in 1960 after reading a signal processing book from the 1940s on exponential smoothing.[16] Holt's novel idea was to repeat filtering an odd number of times greater than 1 and less than 5, which was popular with scholars of previous eras.[16] While recursive filtering had been used previously, it was applied twice and four times to coincide with the Hadamard conjecture, while triple application required more than double the operations of singular convolution. The use of a triple application is considered a rule of thumb technique, rather than one based on theoretical foundations and has often been over-emphasized by practitioners.


Suppose we have a sequence of observations xt, beginning at time t = 0 with a cycle of seasonal change of length L.


The method calculates a trend line for the data as well as seasonal indices that weight the values in the trend line based on where that time point falls in the cycle of length L.


st represents the smoothed value of the constant part for time t. bt represents the sequence of best estimates of the linear trend that are superimposed on the seasonal changes. ct is the sequence of seasonal correction factors. ct is the expected proportion of the predicted trend at any time t mod L in the cycle that the observations take on. As a rule of thumb, a minimum of two full seasons (or 2L periods) of historical data is needed to initialize a set of seasonal factors.


The output of the algorithm is again written as Ft+m, an estimate of the value of x at time t+m, m>0 based on the raw data up to time t. Triple exponential smoothing with multiplicative seasonality is given by the formulas[1]


s0=x0st=αxtct−L+(1−α)(st−1+bt−1)bt=β(st−st−1)+(1−β)bt−1ct=γxtst+(1−γ)ct−LFt+m=(st+mbt)ct−L+1+(m−1)modL,displaystyle beginaligneds_0&=x_0\s_t&=alpha frac x_tc_t-L+(1-alpha )(s_t-1+b_t-1)\b_t&=beta (s_t-s_t-1)+(1-beta )b_t-1\c_t&=gamma frac x_ts_t+(1-gamma )c_t-L\F_t+m&=(s_t+mb_t)c_t-L+1+(m-1)mod L,endaligned

where α is the data smoothing factor, 0 < α < 1, β is the trend smoothing factor, 0 < β < 1, and γ is the seasonal change smoothing factor, 0 < γ < 1.


The general formula for the initial trend estimate b0 is:


b0=1L(xL+1−x1L+xL+2−x2L+…+xL+L−xLL)displaystyle beginalignedb_0&=frac 1Lleft(frac x_L+1-x_1L+frac x_L+2-x_2L+ldots +frac x_L+L-x_LLright)endaligned

Setting the initial estimates for the seasonal indices ci for i = 1,2,...,L is a bit more involved. If N is the number of complete cycles present in your data, then:


ci=1N∑j=1NxL(j−1)+iAj∀i=1,2,…,Ldisplaystyle beginaligned\c_i&=frac 1Nsum _j=1^Nfrac x_L(j-1)+iA_jquad forall i&=1,2,ldots ,L\endaligned

where


Aj=∑i=1LxL(j−1)+iL∀j=1,2,…,Ndisplaystyle beginalignedA_j&=frac sum _i=1^Lx_L(j-1)+iLquad forall j&=1,2,ldots ,Nendaligned

Note that Aj is the average value of x in the jth cycle of your data.


Triple exponential smoothing with additive seasonality is given by:


s0=x0st=α(xt−ct−L)+(1−α)(st−1+bt−1)bt=β(st−st−1)+(1−β)bt−1ct=γ(xt−st−1−bt−1)+(1−γ)ct−LFt+m=st+mbt+ct−L+1+(m−1)modL,displaystyle beginaligneds_0&=x_0\s_t&=alpha (x_t-c_t-L)+(1-alpha )(s_t-1+b_t-1)\b_t&=beta (s_t-s_t-1)+(1-beta )b_t-1\c_t&=gamma (x_t-s_t-1-b_t-1)+(1-gamma )c_t-L\F_t+m&=s_t+mb_t+c_t-L+1+(m-1)mod L,endaligned



Implementations in statistics packages[edit]



  • R: the HoltWinters function in the stats package[17] and ets function in the forecast package[18] (a more complete implementation, generally resulting in a better performance[19]).

  • IBM SPSS includes Simple, Simple Seasonal, Holt's Linear Trend, Brown's Linear Trend, Damped Trend, Winters' Additive, and Winters' Multiplicative in the Time-Series modeling procedure within its Statistics and Modeler statistical packages. The default Expert Modeler feature evaluates all seven exponential smoothing models and ARIMA models with a range of nonseasonal and seasonal p, d, and q values, and selects the model with the lowest Bayesian Information Criterion statistic.


  • Stata: tssmooth command[20]


  • LibreOffice 5.2[21]


  • Microsoft Excel 2016[22]


See also[edit]



  • Autoregressive moving average model (ARMA)

  • Errors and residuals in statistics

  • Moving average

  • Continued fraction


Notes[edit]




  1. ^ abc "NIST/SEMATECH e-Handbook of Statistical Methods". NIST. Retrieved 23 May 2010..mw-parser-output cite.citationfont-style:inherit.mw-parser-output .citation qquotes:"""""""'""'".mw-parser-output .citation .cs1-lock-free abackground:url("//upload.wikimedia.org/wikipedia/commons/thumb/6/65/Lock-green.svg/9px-Lock-green.svg.png")no-repeat;background-position:right .1em center.mw-parser-output .citation .cs1-lock-limited a,.mw-parser-output .citation .cs1-lock-registration abackground:url("//upload.wikimedia.org/wikipedia/commons/thumb/d/d6/Lock-gray-alt-2.svg/9px-Lock-gray-alt-2.svg.png")no-repeat;background-position:right .1em center.mw-parser-output .citation .cs1-lock-subscription abackground:url("//upload.wikimedia.org/wikipedia/commons/thumb/a/aa/Lock-red-alt-2.svg/9px-Lock-red-alt-2.svg.png")no-repeat;background-position:right .1em center.mw-parser-output .cs1-subscription,.mw-parser-output .cs1-registrationcolor:#555.mw-parser-output .cs1-subscription span,.mw-parser-output .cs1-registration spanborder-bottom:1px dotted;cursor:help.mw-parser-output .cs1-ws-icon abackground:url("//upload.wikimedia.org/wikipedia/commons/thumb/4/4c/Wikisource-logo.svg/12px-Wikisource-logo.svg.png")no-repeat;background-position:right .1em center.mw-parser-output code.cs1-codecolor:inherit;background:inherit;border:inherit;padding:inherit.mw-parser-output .cs1-hidden-errordisplay:none;font-size:100%.mw-parser-output .cs1-visible-errorfont-size:100%.mw-parser-output .cs1-maintdisplay:none;color:#33aa33;margin-left:0.3em.mw-parser-output .cs1-subscription,.mw-parser-output .cs1-registration,.mw-parser-output .cs1-formatfont-size:95%.mw-parser-output .cs1-kern-left,.mw-parser-output .cs1-kern-wl-leftpadding-left:0.2em.mw-parser-output .cs1-kern-right,.mw-parser-output .cs1-kern-wl-rightpadding-right:0.2em


  2. ^ ab Oppenheim, Alan V.; Schafer, Ronald W. (1975). Digital Signal Processing. Prentice Hall. p. 5. ISBN 0-13-214635-5.


  3. ^ Brown, Robert G. (1956). Exponential Smoothing for Predicting Demand. Cambridge, Massachusetts: Arthur D. Little Inc. p. 15.


  4. ^ Holt, Charles C. (1957). "Forecasting Trends and Seasonal by Exponentially Weighted Averages". Office of Naval Research Memorandum. 52. reprinted in Holt, Charles C. (January – March 2004). "Forecasting Trends and Seasonal by Exponentially Weighted Averages". International Journal of Forecasting. 20 (1): 5–10.


  5. ^ Brown, Robert Goodell (1963). Smoothing Forecasting and Prediction of Discrete Time Series. Englewood Cliffs, NJ: Prentice-Hall.


  6. ^ "NIST/SEMATECH e-Handbook of Statistical Methods, 6.4.3.1. Single Exponential Smoothing". NIST. Retrieved 5 July 2017.


  7. ^ Nau, Robert. "Averaging and Exponential Smoothing Models". Retrieved 26 July 2010.


  8. ^ "Production and Operations Analysis" Nahmias. 2009.


  9. ^ https://www.otexts.org/fpp/7/1


  10. ^ Nahmias, Steven. Production and Operations Analysis (6th ed.). ISBN 0-07-337785-6.
    [page needed]



  11. ^ "Model: Second-Order Exponential Smoothing". SAP AG. Retrieved 23 January 2013.


  12. ^ Prajakta S. Kalekar. "Time series Forecasting using Holt-Winters Exponential Smoothing" (PDF).


  13. ^ "6.4.3.3. Double Exponential Smoothing". itl.nist.gov. Retrieved 25 September 2011.


  14. ^ "Averaging and Exponential Smoothing Models". duke.edu. Retrieved 25 September 2011.


  15. ^ Kalehar, Prajakta S. "Time series Forecasting using Holt-Winters Exponential Smoothing" (PDF). Retrieved 23 June 2014.


  16. ^ ab Winters, P. R. (April 1960). "Forecasting Sales by Exponentially Weighted Moving Averages". Management Science. 6 (3): 324–342. doi:10.1287/mnsc.6.3.324.


  17. ^ "R: Holt-Winters Filtering". stat.ethz.ch. Retrieved 5 June 2016.


  18. ^ "ets forecast | inside-R | A Community Site for R". www.inside-r.org. Archived from the original on 16 July 2016. Retrieved 5 June 2016.


  19. ^ "Comparing HoltWinters() and ets()". Hyndsight. 29 May 2011. Retrieved 5 June 2016.


  20. ^ tssmooth in Stata manual


  21. ^ https://wiki.documentfoundation.org/ReleaseNotes/5.2#New_spreadsheet_functions


  22. ^ http://www.real-statistics.com/time-series-analysis/basic-time-series-forecasting/excel-2016-forecasting-functions/



External links[edit]


  • Lecture notes on exponential smoothing (Robert Nau, Duke University)


  • Data Smoothing by Jon McLoone, The Wolfram Demonstrations Project


  • The Holt-Winters Approach to Exponential Smoothing: 50 Years Old and Going Strong by Paul Goodwin (2010) Foresight: The International Journal of Applied Forecasting


  • Algorithms for Unevenly Spaced Time Series: Moving Averages and Other Rolling Operators by Andreas Eckner











Retrieved from "https://en.wikipedia.org/w/index.php?title=Exponential_smoothing&oldid=892831597"










Navigation menu


























(window.RLQ=window.RLQ||[]).push(function()mw.config.set("wgPageParseReport":"limitreport":"cputime":"0.636","walltime":"0.997","ppvisitednodes":"value":1963,"limit":1000000,"ppgeneratednodes":"value":0,"limit":1500000,"postexpandincludesize":"value":184079,"limit":2097152,"templateargumentsize":"value":1394,"limit":2097152,"expansiondepth":"value":12,"limit":40,"expensivefunctioncount":"value":3,"limit":500,"unstrip-depth":"value":1,"limit":20,"unstrip-size":"value":55471,"limit":5000000,"entityaccesscount":"value":1,"limit":400,"timingprofile":["100.00% 752.018 1 -total"," 38.83% 291.985 1 Template:Reflist"," 27.79% 208.999 1 Template:Use_dmy_dates"," 27.24% 204.818 1 Template:DMCA"," 26.77% 201.287 1 Template:Dated_maintenance_category"," 23.89% 179.692 1 Template:Statistics"," 21.48% 161.565 1 Template:Navbox_with_collapsible_groups"," 17.55% 131.965 11 Template:Cite_web"," 16.12% 121.257 12 Template:Navbox"," 8.06% 60.641 1 Template:Page_needed"],"scribunto":"limitreport-timeusage":"value":"0.288","limit":"10.000","limitreport-memusage":"value":4877380,"limit":52428800,"cachereport":"origin":"mw1272","timestamp":"20190423145234","ttl":2592000,"transientcontent":false);mw.config.set("wgBackendResponseTime":1116,"wgHostname":"mw1272"););

Popular posts from this blog

کانن (شرکت) محتویات تاریخچه[ویرایش] بخشی از تولیدات موفق این شرکت[ویرایش] در رده APS-C[ویرایش] گزارش محیط زیست[ویرایش] رده‌بندی محصولات[ویرایش] منابع[ویرایش] پانویس[ویرایش] پیوند به بیرون[ویرایش] منوی ناوبریwww.canon.comموزه آنلاین دوربین‌های کانننمودار تاریخچه سهام کاننوبگاه رسمی شرکت کاننوووووIDC Worldwide Hardcopy 2013

Rest API with Magento using PHP with example. Planned maintenance scheduled April 17/18, 2019 at 00:00UTC (8:00pm US/Eastern) Announcing the arrival of Valued Associate #679: Cesar Manara Unicorn Meta Zoo #1: Why another podcast?How to update product using magento client library for PHP?Oauth Error while extending Magento Rest APINot showing my custom api in wsdl(url) and web service list?Using Magento API(REST) via IXMLHTTPRequest COM ObjectHow to login in Magento website using REST APIREST api call for Guest userMagento API calling using HTML and javascriptUse API rest media management by storeView code (admin)Magento REST API Example ErrorsHow to log all rest api calls in magento2?How to update product using magento client library for PHP?

Magento 2 - Auto login with specific URL Planned maintenance scheduled April 23, 2019 at 23:30 UTC (7:30pm US/Eastern) Announcing the arrival of Valued Associate #679: Cesar Manara Unicorn Meta Zoo #1: Why another podcast?Customer can't login - Page refreshes but nothing happensCustom Login page redirectURL to login with redirect URL after completionCustomer login is case sensitiveLogin with phone number or email address - Magento 1.9Magento 2: Set Customer Account Confirmation StatusCustomer auto connect from URLHow to call customer login form in the custom module action magento 2?Change of customer login error message magento2Referrer URL in modal login form