Keywords

1 Introduction

Volatility modelling has been extensively studied: more than 30 years have gone by since the seminal contributions by [9, 14]. As they have about 25 K citations each (and some pertinent papers do not even mention them), it is clear that GARCH-type models are the standard among academicians and practitioners alike. These models build upon stylized facts of persistence in the conditional second moments (volatility clustering), an analysis made easier by the direct measurement of volatility, starting from the availability of ultra-high-frequency data (cf. [7]). Looking directly at the series of the Standard and Poor’s (S&P) 500 Index and of its realized volatility, as illustrated in Fig. 1, one encounters two of such stylized facts in need of adequate modelling: the first is that volatility has a slow-moving/state-dependent average local level of volatility to be accounted for, and hence its dynamic evolution is driven by two components: a high-frequency and a low-frequency one. Another is that peaks of volatility are recorded in correspondence with streaks of downturns in the index, a sign of well-documented asymmetry in the dynamics.

Fig. 1
figure 1

S&P 500 Index and its realized volatility

Various suggestions exist in the literature to model the first of these two stylized facts: in a Markov Switching approach, GARCH parameters are state-dependent ( [10, 13, 19], among others). The resulting high-frequency dynamics varies across states and evolves around a constant average level of volatility within states as a low-frequency component. In other contributions, the two components are additive; [12, 15] specify a model in which higher persistence is an identifying condition for the long-run component. The most popular GARCH specification is one in which long- and short-run components combine multiplicatively with the error term. Amado et al.  [4] survey the contributions in this field: the common feature is that long run is a term which smoothly amplifies or dampens the short-run GARCH dynamics. The long-run term can be a deterministic function of time as in the Spline GARCH [16]; a logistic function of a forcing variable as in the Smooth Transition approach ( [1,2,3], for instance); an exponential function of a one-sided filter of past values of a variable sampled at a lower frequency than the daily series of interest, as in the GARCH-MIDAS of [17]. In this paper, we take a modification of this latter model, called the Double Asymmetric GARCH-MIDAS (DAGM) introduced by [5], where a rate of change at a low frequency is allowed to have differentiated effects according to its sign, determining a local trend around which an asymmetric GARCH that describes the short-run dynamics. A market volatility variable (in our context, we choose the VIX index which is based on implied volatilities from at-the-money option prices) is inserted as a daily lagged variable, and monthly variations represent an additional channel through which market volatility can influence individual stocks.

The issue at stake in this empirically motivated paper is how this information about market-based volatility can help in shaping the MIDAS-GARCH dynamics. The idea we are pursuing is to illustrate

  1. 1.

    how a predetermined daily variable (in lagged levels) adds some significant influence to the short-run component (an asymmetric GARCH in the form of the GJR [18] model—this would be the first asymmetry considered); and, most importantly,

  2. 2.

    how the same variable observed at a lower frequency (in lagged first differences) can determine a useful combination (in the MIDAS sense seen in detail below) for the low-frequency component in the slowly moving level of local average volatility. In particular, it is of interest to explore what frequency (weekly or monthly), works best in this context, and what horizon is informative. In so doing, we maintain that positive changes (an increase in volatility) and negative ones should be treated differently in the model (this is the second asymmetry considered).

The results show that in characterizing the volatility dynamics, our model with monthly data and six months of lagged information works best, together with the contribution of the lagged VIX in the short-run component. Out-of-sample, the model behaves well, with a performance which is dependent on the sub-period considered.

The rest of the paper is organized as follows: Sect. 2 addresses the empirical question, illustrating first how the DAGM works and then we report the results of an application of various GARCH, GARCH-MIDAS and DAGM models on the S&P 500 volatility, both in- and out-of-sample perspectives. Finally, Sect. 3 concludes.

2 Modelling Volatility with the DAGM Model

Let us focus on the GARCH-MIDAS model, here synthetically labelled GM: the paper by [17] defines GARCH dynamics in the presence of mixed frequency variables. The short-run component varies with the same frequency as the dependent variable while the long-run component filters the lower frequency macro-Variable(s) (MV) observations. Recent contributions on (univariate) GARCH-MIDAS model are [6, 8, 11].

The paper by [5] proposes a DAGM where asymmetry in the short run is captured by a GJR-type [18] reaction to the sign of past returns, and positive and negative MV values have different impacts on the long-run.

2.1 The DAGM Framework

The DAGM-X model is defined as

$$\begin{aligned} r_{i,t} = \sqrt{\tau _t \times g_{i,t} } \varepsilon _{i,t}, \quad \text {with } i = 1,\ldots ,N_t, \end{aligned}$$
(1)

where

  • \(r_{i,t}\) represents the log-return for day i of the period t;

  • \(N_t\) is the number of days for period t, with \(t=1,\ldots ,T\);

  • \(\varepsilon _{i,t} | \Phi _{i-1,t} \sim N(0,1)\), where \(\Phi _{i-1,t}\) denotes the information set up to day \(i-1\) of period t;

  • \(g_{i,t}\) follows a unit-mean reverting GARCH(1,1) process (short-run component);

  • \(\tau _t\) provides the slow-moving average level of volatility (long-run component).

The short-run component of the DAGM-X is given by

$$\begin{aligned} g_{i,t} = \left( 1-\alpha -\beta -\gamma /2\right) + \left( \alpha + \gamma \cdot \mathbbm {1}_{\left( r_{i-1,t} < 0 \right) }\right) \frac{\left( r_{i-1,t}\right) ^2}{\tau _t} + \beta g_{i-1,t} + z V_{i-1,t}, \end{aligned}$$
(2)

where \(\mathbbm {1}_{(.)}\) is an indicator function and \(V_{i,t}\) is an additional, positive volatility determinant, observed daily, whose importance on \(g_{i,t}\) is given by z. In order to assure the positivity of \(g_{i,t}\), we impose the constraint \(z \ge 0\). In absence of \(V_{i,t}\), the DAGM-X model becomes the DAGM specification.

The long-run component of the DAGM-X and DAGM is defined as

$$\begin{aligned} \tau _t = \mathrm{exp} \left( m + \theta ^{+} \sum _{k=1}^K \delta _k(\omega )^{+} X_{t-k} \mathbbm {1}_{\left( X_{t-k} \ge 0 \right) } + \theta ^{-} \sum _{k=1}^K \delta _k(\omega )^{-} X_{t-k} \mathbbm {1}_{\left( X_{t-k} < 0 \right) } \right) , \end{aligned}$$
(3)

where

  • m plays the role of an intercept;

  • \(\theta ^+\), \(\theta ^-\) represent the asymmetric responses to the one-sided filter;

  • \(\delta _k(\omega )^+\) and \(\delta _k(\omega )^-\) are suitable functions weighing the past K realizations of the additional stationary variable \(X_t\). As in the related literature, we opt for the Beta function, that is

    $$\begin{aligned} \delta _k(\omega )=\frac{(k/K)^{\omega _1-1} (1-k/K)^{\omega _2-1}}{\sum _{j=1}^K (j/K)^{\omega _1-1}(1-j/K)^{\omega _2-1}}. \end{aligned}$$
    (4)

Given that we are only interested in the case of larger weights put on the most recent observations, we set \(\omega _1=1\) and \(\omega _2 \ge 1\). Note that the Beta function represented in (4) is readily applicable for both the GM and the DAGM. In this latter case, it is sufficient to replace \(\delta _k(\omega )\) with \(\delta _k(\omega )^+\) and \(\delta _k(\omega )^-\).

Thus, the short-run component includes a term related to negative returns (“bad news” increasing volatility, the well-known leverage effect) and potentially a term associated with an additional MV observed with the same frequency of the dependent variable. The long-run component avoids positive and negative compensations within the one-sided filter, separating the positive MV variations from the negative ones.

Typically, MVs can only be observed at low frequency, but here we move out of the classic MV framework where observations are available only at low frequency. Thus we take a variable which is observable daily, but can be sampled at lower frequencies, e.g. weekly or monthly. We take the DAGM to the empirical evaluation of how different frequencies of observations in the MV may change the results both in estimation and forecasting. Besides that, we include the same variable at high frequency in the short-run component (“–X” specifications).

Assuming a conditional normal distribution for the error term \(\varepsilon _{i,t}\) allows to apply the standard statistical inference (for details on the asymptotic properties of the GARCH-MIDAS class of models, see [22]) according to the maximization of the following log-likelihood:

$$\begin{aligned} \ln {\mathscr {L}} = -\frac{1}{2}\sum _{t=1}^T \left[ \sum _{i=1}^{N_t} \left[ \log (2\pi ) + \log (g_{i,t} \tau _t) + \frac{\left( r_{i,t}\right) ^2}{g_{i,t} \tau _t}\right] \right] . \end{aligned}$$
(5)

2.2 The Role of VIX in the S&P 500 Volatility Dynamics

The returns of interest are daily log-differences of the S&P 500 Index (also examined on a different sample and context in [5]), annualized on a sample period: 7 January 2000–7 September 2018 (number of daily observations: 4686, collected from Yahoo Finance).

The MV in this paper is VIX (an implied volatility-based index built on the same index, cf. [23]) which in our setup will appear: (i) lagged daily as a predetermined variable in the short-run component \(g_i\) of the GARCH-X; (ii) lagged variations—end-of-month or end-of-week (with various choices of K) in the long-run component. All the observations concerning VIX have been collected from the Thomson Reuters Eikon provider. The distance between the estimated volatility, labelled as \(\hat{h}_i\), and the chosen volatility proxy, the realized volatility at five minutes, labelled as \(\sigma _i\) and collected from the realized library of the Oxford-Man Institute, are investigated through three loss functionsFootnote 1: QLIKE, Root Mean Squared Error (RMSE) and Mean Absolute Errors (MAE), defined as follows:

$$\begin{aligned} \begin{aligned} \text {QLIKE}&: E\left( \sigma _i/\hat{h}_i + \log (\hat{h}_i)\right) ;\\ \text {RMSE}&: \sqrt{E\left( \left( \sigma _i - \hat{h}_i\right) ^2\right) };\\ \text {MAE}&: E\left( | \sigma _i - \hat{h}_i|\right) . \end{aligned} \end{aligned}$$
(6)

2.2.1 Estimation and Diagnostics

The estimation and diagnostics results are shown in Table 1, where we report the coefficients with their standard errors in parenthesis and their significance. GARCH is the standard (1, 1) model; the GJR allows for an asymmetric response to the lagged negative returns; the GARCH-X and GJR-GARCH-X and DAGM-X contain an extra predetermined variable, the lagged daily VIX. The GM and DAGM are built on a one-sided filter for the monthly VIX, while in the DAGM-W we consider the weekly VIX. The last six months of VIX have been used in GM, DAGM, DAGM-X, and DAGM-W, i.e. \(K = 6\) and \(K_{\mathrm{week}}=24\). The choice of the adopted MIDAS lags derives from some preliminary estimations aiming at finding the best values according to the Bayesian information criterion (BIC). The number of “\(^{*}\)” indicates the significance (\(10\%,5\%, 1\%\), respectively) of the estimated coefficients heteroscedasticity and autocorrelation consistent ( [21], HAC) standard errors in parenthesis). LB\(_{l}\) and LM\(_{l}\) report the p-value of the Ljung-Box and ARCH-LM tests on the squared standardized residuals at lag l, respectively. RMSE is in percentage terms.

A few comments are in order: the GARCH models (first four columns) exhibit customary results, with the possible surprise of the non-significance of the lagged VIX in the X specifications. The GM has non-significant coefficient on the one-sided filter and the wrong sign: as a matter of fact, the information criteria and the QLIKE signal a worse fit of this model, relative to the standard models. When we introduce our DAGM, the signs of the impact coefficients \(\theta ^+\) and \(\theta ^-\) are the right ones (positive, negative, respectively), and significant. The information criteria and the QLIKE report a marked improvement over the models previously considered, with the best model being the DAGM-X model where the significant coefficients on the low-frequency component are, besides the constant, those pertaining to the positive changes (inducing an increase in volatility). Generally, the residual diagnostics show a good fit of the models. In particular, almost all the models fail to reject the null hypotheses of the Ljung-Box and ARCH-LM tests, independently of the order of lags considered. The only model whose p-values are below the significance level of 5% is the DAGM-X, for \(l=12\), for both the Ljung-Box and ARCH-LM tests. Despite this, the conclusion is that the DAGM-X provides the most convincing performance with VIX contributing to a marked improvement over other models. The result can be appraised graphically as in Fig. 2 where we show the close proximity of the fitted values to the realized volatility.

Table 1 DAGM and GARCH estimates
Fig. 2
figure 2

Notes The figure plots the DAGM-X volatility (solid black line) and the S&P 500 realized volatility (dashed grey line). Shaded areas represent NBER recession periods. Annualized scale

Realized and DAGM-X volatilities.

2.2.2 Forecasting

Further insights can be had moving to an out-of-sample exercise where we estimate the model over a 10-year period and project one-step ahead for one year and then move forward the estimation and forecasting window. The results are summarized in Fig. 3 where we report the presence in a Model Confidence Set as proposed by [20]. The results (at \(\alpha =10\%\)) show that while the DAGM-X has a satisfactory performance, at the same time the standard GARCH or GJR models enter the set.

Fig. 3
figure 3

Notes The figure plots the composition of the Model Confidence Set (MCS). For different loss functions, dark (QLIKE), medium-dark (MSE) and light (MAE) shades of grey indicate that a given model is included in the MCS, at a significance level of \(\alpha =0.1\).

MCS composition.

3 Wrapping Up

The slow-moving feature of conditional volatility can be addressed within a Double Asymmetric GARCH-MIDAS framework [5] where the low-frequency variable here is a volatility measure (variations in VIX). The main novelty in this approach is that the same variable can be inserted as a forcing variable (-X in levels) in the short-run component, and we can explore which frequency is the most suitable for the long-run component (in first differences). The fitting capabilities of this approach are comforting, with monthly movements in volatility providing the best in-sample results. In out-of-sample forecasting, though, the model is less satisfactory, in that it gives a performance very similar to a standard GARCH model.