|
by Romain Berry This article is the eighth in a series of articles exploring risk management for institutional investors. What is volatility? As we demonstrated in our previous articles, volatility is one of the main drivers of Value-at-Risk (VaR) models for market risk. Therefore, it is important to select the right volatility model to capture various properties such as the lag in time, the clustering and so forth. Volatility is simply the standard deviation of the returns of a distribution. If returns are independently and individually distributed (i.i.d.)1, then the volatility is called the sample standard error2 of the return distribution. However, since returns of most financial assets are not normal, we need to estimate their volatility with some statistical analysis. Financial time series exhibit volatility clustering, as can clearly be seen over various periods of time for the S&P 500 Index (1950 - 2010) in Exhibit 1. Exhibit 1 – Returns of the S&P 500
would raise two questions regarding the optimal size of N (and its perenniality in the future to keep consistent measures over time) and the fact that each observation carries the same weight (which is in contradiction with the volatility clustering assumption). Unlike the mean, higher (central) moments4 of a return distribution exhibit recognizable patterns and can be helpful when forecasting future prices. Indeed, when prices increase sharply one day, it can be reasonably assumed that they will rise or fall sharply the following day (high volatility regime). Therefore, it seems feasible to forecast the magnitude of the price changes, but not the direction. We introduce in this paper the two most common conditional volatility models used to estimate a univariate volatility financial time series: the Generalized Auto-Regressive Conditional Heteroscedasticity5 (GARCH) and the Exponentially Weighted Moving Average (EWMA) models. We also demonstrate how to estimate univariate volatility with either model and how to assess their prediction power. Stochastic Volatility Models The simplest and most commonly used GARCH model designed by Bollerslev6 is the GARCH (1,1) and is defined as
This model estimates the volatility on a given day based on a linear combination of the squared returns and volatilities of the previous days plus a constant. Indeed the “(1,1)” term in GARCH (1,1) indicates that the current variance is based on the squared return and variance of the previous day (1 lag for each). We use the squared returns because they also exhibit strong recognizable patterns. The EWMA model is also widely used to estimate and forecast volatility and is defined by
It is easy to see that the EWMA is a particular case of a GARCH model where ω = 0, α = 1 – λ and β = λ. One of the main differences between these two models is that the GARCH model incorporates mean reversion8 while the EWMA does not. Estimating The Parameters In order to illustrate how the parameters are estimated, let us analyze the S&P 500 between January 1950 and January 2010. The most common method to estimate the parameters is to take the Log Likelihood. It is the logarithm of the Maximum Likelihood (ML). ML employs trails and errors to determine the optimal parameters that maximize the likelihood of the data to occur. It requires first to estimate the distribution of returns. For illustration (only), let us assume it follows a normal distribution.
Since the EWMA and GARCH models contain lagged squared returns and lagged volatilities, we need to estimate σ12. Theoretically, it is sounder to include it in the parameters (α, β, ω) to be estimated. In practice, if the sample size is large enough (few hundred observations), we can specify σ12asthe unconditional variance9 of all observations of ri. Since we have more than 15,000 observations, we go with the latter methodology. Exhibit 2 and 3 show how to construct these two models for the S&P 500. I used the Solver functionality of Excel to estimate the parameters ω, α and β. The cell in yellow corresponds to σ12 (since we can only start from January 5th, 1950) obtained by setting it to the unconditional variance of the sample (i.e., the long run variance). Exhibit 2 – Methodology to estimate the parameters of a GARCH(1,1) model10
The estimation of the EWMA model is simpler since we set ω =0, α = 1 – λ, and β = λ. Exhibit 3 – Methodology to estimate the parameters of an EWMA model10
The log-likelihood of both models are very close from each other (with an advantage to GARCH (1,1)) but diverge on the maximum volatility (6.46% for the GARCH (1,1) model compared to 5.45% with the EWMA model). Therefore, the next questions are which model to choose and why? Selecting A Volatility Model We noticed that the volatility calculated with the GARCH (1,1) model was higher than the one calculated with the EWMA model for 67% of the data. Most of these differences were quite reasonable – between -5% and 0% in 85% of the occurrences. Thus, the GARCH (1,1) model seemed a bit more reactive to new shocks than the EWMA model, which is a good feature for a model to have in a volatility clustering environment. Exhibit 4 – GARCH(1,1) and EWMA for the S&P 500
A quick visual test would be to draw the correlogram12 of the S&P 500 with each model and check how much autocorrelation has been effectively removed by each model (difference between a red bar and a green bar at the same time lag in Exhibit 5 and 6). Exhibit 5 – S&P 500 correlogram with GARCH(1,1)
This is confirmed after running a Ljung-Box test13 over the first 30 lags at 99% confidence level.
According to this test, both models have removed more than 99% of the autocorrelation but the GARCH (1,1) model did a slightly better job with a removal of 99.56% of the autocorrelation versus 99.17% for the EWMA model. We note though that we cannot reject at 99% confidence the hypothesis that the EWMA model has completely removed all the autocorrelation from the time series a contrario of the GARCH(1,1) model. Other autocorrelation tests like the Box-Pierce could be applied to confirm that result. Conclusion We presented in this article two of the most commonly used stochastic models to estimate the volatility of financial returns. The GARCH model is to be preferred for short term horizons because it is mean reverted. Many different GARCH models have been designed since Engle and Bollerslev which are well worth considering because they take account of leverage, asymmetry, and other properties of financial time series not well captured by the two models presented here. They are IGARCH, GARCH-M, EGARCH, TGARCH, PGARCH, NGARCH, QGARCH, GJR-GARCH, AGARCH.
To view the next article, Multiple Asset Class Return Comparison, click here.
1independently and individually distributed means that the observations of a random variable (the returns of an asset over a period of time, for instance) are independent of each other and have the same distribution. Thus, they all have the same mean and variance. mk = E(x – μ)k
The second, third and fourth central moments are the variance, skewness and kurtosis of the distribution.
The Ljung-Box statistic is asymptotically distributed as chi-squared with p degrees of freedom. Here, for k=30, the hypothesis of complete removal of the autocorrelation in the time series can be rejected with 99% confidence when the Ljung-Box statistic is greater than 50.9.
|