1 Introduction

Volatility modelling is the key to the theory and practice of pricing financial products. Asset allocation and portfolio as well as risk management depend heavily on a correct modelling of the underlying(s). This insight has spurred extensive research in financial econometrics and mathematical finance. Stochastic volatility models with separate dynamic structure for the volatility process have been in the focus of the mathematical finance literature, see Heston (1993) and Bates (2000), while parametric GARCH-type models for the returns of the underlying(s) have been intensively analyzed in financial econometrics.

The validity of these models in practice though depends upon specific distributional properties or the knowledge of the exact (parametric) form of the volatility dynamics. Moreover, the evaluation of the predictive ability of volatility models is quite important in empirical applications. However, the latent character of the volatility poses a problem. To what measure should the volatility forecasts be compared to? Conventionally, the forecasts of daily volatility models, such as GARCH-type or stochastic volatility models, have been evaluated with respect to absolute or squared daily returns. In view of the excellent in-sample performance of these models, the forecasting performance, however, seems to be disappointing.

The availability of ultra-high-frequency data opens the door for a refined measurement of volatility and model evaluation. An often used and very flexible model for logarithmic prices of speculative assets is the (continuous time) stochastic volatility model:

$$\begin{aligned} dY_t = (\mu + \beta \sigma _t) dt + \sigma _t dW _t, \end{aligned}$$
(14.1)

where \(\sigma ^2_t\) is the instantaneous (spot) variance, \(\mu \) denotes the drift, \(\beta \) is the risk premium, and \(W_t\) defines the standard Wiener process. The object of interest is the amount of variation accumulated in a time interval \(\varDelta \) (e.g., a day, week, month etc.). If \(n = 1, 2,\ldots \) denotes a counter for the time intervals of interest, then the term

$$\begin{aligned} \sigma ^2_n=\int ^{n\varDelta }_{(n-1)\varDelta }\sigma ^2_t dt \end{aligned}$$
(14.2)

is called the actual volatility , see Barndorff-Nielsen and Shephard (2002b). The actual volatility is the quantity that reflects the market risk structure (scaled in \(\varDelta \)) and is the key element in pricing and portfolio allocation. Actual volatility (measured in scale \(\varDelta \)) is of course related to the integrated volatility:

$$\begin{aligned} V(t)=\int ^t_0\sigma ^2_s ds \end{aligned}$$
(14.3)

It is worth noting that there is a small notational confusion here: the mathematical finance literature would denote \(\sigma _t\) as “volatility” and \(\sigma ^2_t\) as “variance”, see Nelson and Foster (1994). For example, an important result is that V(t) can be estimated from \(Y_t\) via the quadratic variation:

$$\begin{aligned}{}[Y_t]_M=\sum (Y_{t_j}-Y_{t_{j-1}})^2, \end{aligned}$$
(14.4)

where \(t_0 = 0< t_1< \ldots < t_M = t\) is a sequence of partition points and \(\sup _j| t_{j+1} - t_j| \rightarrow 0\). Andersen and Bollerslev (1998) have shown that

$$\begin{aligned}{}[Y_t]_M\mathop {\rightarrow }^{p}V(t),\ M\rightarrow \infty . \end{aligned}$$
(14.5)

This observation leads us to consider in an interval \(\varDelta \) with M observations

$$\begin{aligned} RV_n=\sum \limits _{j=1}^M(Y_{t_j}-Y_{t_{j-1}})^2 \end{aligned}$$
(14.6)

with \(t_j = \varDelta \{(n - 1) + j/M\}\). Note that \(RV_n\) is a consistent estimator of \(\sigma ^2_n\) and is called realized volatility. Barndorff-Nielsen and Shephard (2002b) point out that \(RV_n - \sigma ^2_n\) is approximately mixed Gaussian and provide the asymptotic law of

$$\begin{aligned} \sqrt{M}(RV_n-\sigma ^2_n). \end{aligned}$$
(14.7)

The realized volatility turns out to be very useful in the assessment of the validity of volatility models. For instance, reconciling evidence in favor of the forecast accuracy of GARCH-type models is observed when using realized volatility as a benchmark rather than daily squared returns. Moreover, the availability of the realized volatility measure initiated the development of a new and quite accurate class of volatility models. In particular, based on the ex-post observability of the realized volatility measure, volatility is now treated as an observed rather than a latent variable to which standard time series procedures can be applied.

The remainder of this chapter is structured as follows. We first discuss the practical problems encountered in the empirical construction of realized volatility which are due to the existence of market microstructure noise. Section 14.3 presents the stylized facts of realized volatility, while Sect. 14.4 reviews the most popular realized volatility models. Section 14.5 illustrates the usefulness of the realized volatility concept for measuring time-varying systematic risk within a conditional asset pricing model (CAPM).

2 Market Microstructure Effects

The consistency of the realized volatility estimator builds on the notion that prices are observed in continuous time and without measurement error. In practice, however, the sampling frequency is inevitably limited by the actual quotation or transaction frequency. Since high-frequency prices are subject to market microstructure noise, such as price-discreteness, bid-and-ask bounce effects, transaction costs etc., the true price is unobservable. Market microstructure effects induce a bias in the realized volatility measure, which can straightforwardly be illustrated in the following simple discrete-time setup. Assume that the logarithmic high-frequency prices are observed with noise, i.e.,

$$\begin{aligned} Y_{t_j}=Y_{t_j}^{*}+\varepsilon _{t_j}, \end{aligned}$$
(14.8)

where \(Y_{t_j}^{*}\) denotes the latent true price. Moreover, the microstructure noise \(\varepsilon _{t_j}\) is assumed to be iid distributed with mean zero and variance \(\eta ^2\), and is independent of the true return. Let \(r^{*}_{t_j}\) denote the efficient return, then the high-frequency continuously compounded returns

$$\begin{aligned} r_{t_j}=r^{*}_{t_j}+\varepsilon _{t_j}-\varepsilon _{t_{j-1}} \end{aligned}$$
(14.9)

follow an MA(1) process. Such a return specification is well established in the market microstructure literature and is usually justified by the existence of the bid-ask bounce effect, see, e.g., Roll (1984). In this model, the realized volatility isgiven by

$$\begin{aligned} RV_n=\sum \limits _{i=1}^M(r_{t_j}^{*})^2+2\sum \limits _{j=1}^Mr^{*}_{t_j}(\varepsilon _{t_j}-\varepsilon _{t_{j-1}})+\sum \limits _{j=1}^M(\varepsilon _{t_j}-\varepsilon _{t_{j-1}})^2. \end{aligned}$$
(14.10)

with

$$\begin{aligned} \textsf {E}[RV_n]=\textsf {E}[RV_n^{*}]+2M\eta ^2. \end{aligned}$$
(14.11)

If the sampling frequency goes to infinity, we know from the previous section that \(RV^{*}_n\) consistently estimates \(\sigma ^2_n\) and, thus, the realized volatility based on the observed price process is a biased estimator of the actual volatility with bias term \(2M\eta ^2\). Obviously, for \(M \rightarrow \infty ,\ RV_n\) diverges.

Fig. 14.1
figure 1

Volatility signature plot for IBM, 2001–2006. Average time between trades: 6.78 s. XFGsignature

This diverging behavior can also be observed empirically in so called volatility signature plots . Figure 14.1 shows the volatility signature for one stock of the IBM incorporation over the period ranging from January 2, 2001 to December 29, 2006. The plot depicts the average annualized realized volatility over the full sample period constructed at different frequencies measured in number of ticks (depicted in log scale). Obviously, the realized volatility is large at the very high frequency, but decays for lower frequencies and stabilizes around a sampling frequency of 300 ticks, which corresponds approximately to a 30 min sampling frequency, given that the average duration between two consecutive trades is around 6.78 s.

Thus, sampling at a lower frequency, such as every 10, 15 or 30 min, seems to alleviate the problem of market microstructure noise and has thus frequently been applied in the literature. This so-called sparse sampling, however, comes at the cost of a less precise estimate of the actual volatility. Alternative methods have been proposed to solve this bias-variance trade-off for the above simple noise assumption as well as for more general noise processes, allowing also for serial dependence in the noise and/or for dependence between the noise and the true price process, which is sometimes referred to as endogenous noise. A natural approach to reduce the market microstructure noise effect is to construct the realized volatility measure based on prefiltered high-frequency returns, using, e.g., an MA(1) model.

In the following we briefly present two more elaborate and under specific noise assumptions consistent procedures for estimating actual volatility. Both have been theoretically considered in several papers. The subsampling approach originally suggested by Zhang et al. (2005) builds on the idea of averaging over various realized volatilities constructed from different high-frequency subsamples. For the ease of exposition we focus again on one time period, e.g., one day, and denote the full grid of time points at which the M intradaily prices are observed by \(\mathcal {G}_t = \{t_0, \ldots , t_M\}\). The realized volatility that makes use of all observations in the full grid is denoted by \(RV^{(all)}_n\). Moreover, the grid is partitioned into L nonoverlapping subgrids \(\mathcal {G}^{(l)},\,l \,{=}\, 1, \ldots , L\). A simple way for selecting such a subgrid may be the socalled regular allocation, in which the l-th subgrid is given by \(\mathcal {G}^{(l)} \,{=}\, \{t_{l-1}, t_{l-1+L}, \ldots , t_{l-1+M_lL}\}\) for \(l \,{=}\, 1, \ldots , L\), and \(M_l\) denoting the number of observations in each subgrid. E.g., consider 5-min returns that can be measured at the time points 9:30, 9:35, 9:40, ..., and at the time points 9:31, 9:36, 9:41, ...and so forth. In analogy to the full grid, the realized volatility for subgrid l, denoted by \(RV^{(l)}_n\), is constructed from all data points in subgrid l. Thus, \(RV^{(l)}_n\) is based on sparsely sampled data.

The actual volatility is then estimated by:

$$\begin{aligned} RV_n^{ (ZMA)}=\frac{1}{L}\sum \limits _{l=1}^L RV ^{(l)}_n-\frac{\overline{M}}{M} RV _n^{(all)}, \end{aligned}$$
(14.12)

where \(\overline{M} = \frac{1}{L}\sum ^L_{l=1}M_l\). The latter term on the right-hand side is included to bias-correct the averaging estimator \(\frac{1}{L}\sum ^L_{l=1} RV ^{(l)}_n\) . As the estimator (14.12) consists of a component based on sparsely sampled data and one based on the full grid of price observations, the estimator is also called the two-timescales estimator.

Given the similarity to the problem of estimating the long-run variance of a stationary time series in the presence of autocorrelation, it is not surprising that kernel-based methods have been developed for estimating the realized volatility. Most recently, Barndorff-Nielsen et al. (2008) proposed the flat-top realized kernel estimator

$$\begin{aligned} RV _n^{ (BHLS)}= RV _n+\sum \limits _{h=1}^{H^{*}}K\left( \frac{h-1}{H^{*}}\right) (\widehat{\gamma }_h+\widehat{\gamma }_{-h}) \end{aligned}$$
(14.13)

with

$$\begin{aligned} \widehat{\gamma }_h=\frac{M}{M-h}\sum \limits _{j=1}^{M}r_{t_j}r_{t_{j-h}}, \end{aligned}$$
(14.14)

and \(K(0) = 1,\ K(1) = 0\). Obviously, the summation term on the righthand side is the realized kernel correction of the market microstructure noise. Zhou (1996), who was the first to consider realized kernels, proposed (14.13) with \(H = 1\), while Hansen and Lunde (2006) allowed for general H but restricted \(K(x) = 1\). Both of these estimators, however, have been shown to be inconsistent. Barndorff-Nielsen et al. (2008) instead propose several consistent realized kernel estimators with an optimally chosen \(H^{*}\), such as the Tukey-Hanning kernel, i.e. \(K(x) = \{1-\cos \pi (1-x)^2\}/2\), which performs also very well in terms of efficiency as illustrated in a Monte Carlo analysis. They further show, that these realized kernel estimators are robust to market microstructure frictions that may induce endogenous and dependent noise terms.

Fig. 14.2
figure 2

Kernel density estimates of the (logarithmic) realized volatility and of correspondingly standardized returns for IBM, 2001–2006. The dotted line depicts the density of the correspondingly fitted normal distribution. The left column depicts the kernel density estimates based on a log scale. XFGkernelcom

3 Stylized Facts of Realized Volatility

Figure 14.2 shows kernel density estimates of the plain and logarithmic daily realized volatility in comparison to plots of a correspondingly fitted (log) normal distribution based on the IBM data, 2001–2006. The pictures in the top of Fig. 14.2 show the unconditional distribution of the (plain) realized volatility in contrast to a fitted normal distribution. As also confirmed by the corresponding descriptive statistics displayed by Table 14.1, we observe that realized volatility reveals severe right-skewness and excess kurtosis. This result might be surprising given that the realized volatility consists of the sum of squared intra-day returns and thus central limit theorems should apply. However, it is a common finding that intra-day returns are strongly serially dependent requiring significantly higher intra-day sampling frequencies to observe convergence to normality. In contrast, the unconditional distribution of the logarithmic realized volatility is well approximated by a normal distribution. The sample kurtosis is strongly reduced and is close to 3. Though slight right-skewness and deviations from normality in the tails of the distribution are still observed, the underlying distribution is remarkably close to that of a Gaussian distribution.

A common finding is that financial returns have fatter tails than the normal distribution and reveal significant excess kurtosis. Though GARCH models can explain excess kurtosis, they cannot completely capture these properties in real data. Consequently, (daily) returns standardized by GARCH-induced volatility, typically still show clear deviations from normality. However, a striking result in recent literature is that return series standardized by the square root of realized volatility, \(r_n/\sqrt{RV}_n\), are quite close to normality. This result is illustrated by the plots in the bottom of Fig. 14.2 and the descriptive statistics in Table 14.1. Though we observe deviations from normality for returns close to zero resulting in a kurtosis which is even below 3, the fit in the tails of the distribution is significantly better than that for plain log returns. Summarizing the empirical findings from Fig. 14.2, we can conclude that the unconditional distribution of daily returns is well described by a lognormal-normal mixture. This confirms the mixture-of-distribution hypothesis by Clark (1973) as well as the idea of the basic stochastic volatility model, where the log variance is modelled in terms of a Gaussian AR(1) process.

Figure 14.3 shows the evolvement of daily realized volatility over the analyzed sample period and the implied sample autocorrelation functions (ACFs). As also shown by the corresponding Ljung-Box statistics in Table 14.1, the realized volatility is strongly positively autocorrelated with high persistence. This is particularly true for the logarithmic realized volatility. The plot shows that the ACF decays relatively slowly providing hints on the existence of long range dependence. Indeed, a common finding is that the realized volatility processes reveal long range dependence which is well captured by fractionally integrated processes. In particular, if \(RV_n\) is integrated of the order \(d \in \) (0, 0.5), it can be shown that

Table 14.1 Descriptive statistics of the realized volatility, log realized volatility and standardized returns, IBM stock, 2001–2006. LB (40) denotes the Ljung-Box statistic based on 40 lags. The last row gives an estimate of the order of fractional integration based on the Geweke and Porter-Hudak estimator
Fig. 14.3
figure 3

Time evolvement and sample autocorrelation function of the realized volatility for IBM, 2001–2006. XFGrvtsacf

$$\begin{aligned} \textsf {Var}\left[ \sum \limits _{j=1}^h RV _{n+j}\right] \approx ch ^{2d+1}, \end{aligned}$$
(14.15)

with c denoting a constant. Then, plotting \(\ln \textsf {Var} \left[ \sum ^h_{j=1} RV _{n+j}\right] \) against ln h should result in a straight line with slope \(2d + 1\). Most empirical studies strongly confirm this relationship and find values for d between 0.35 and 0.4 providing clear evidence for long range dependence. Estimating d using the Geweke and Porter-Hudak estimator, we obtain \(\widehat{d} = 0.38\) for the series of realized volatilities and \(\widehat{d} = 0.62\) for its logarithmic counterpart. Hence, for both series we find clear evidence for long range dependence. However, the persistence in logarithmic realized volatilities is remarkably high providing even hints on non-stationarity of the process.

Summarizing the most important empirical findings, we can conclude that the unconditional distributions of logarithmic realized volatility and of correspondingly standardized log returns are well approximated by normal distributions and that realized volatility itself follows a long memory process. These results suggest (Gaussian) ARFIMA models as valuable tools to model and to predict (log) realized volatility.

4 Realized Volatility Models

As illustrated above, realized volatility models should be able to capture the strong persistence in the sample autocorrelation function. While this seemingly long-memory pattern is widely acknowledged, there is still no consensus on the mechanism generating it. One approach is to assume that the long memory is generated by a fractionally integrated process as originally introduced by Granger and Joyeux (1980) and Hosking (1981). In the GARCH literature this has lead to the development of the fractionally integrated GARCH model as, e.g., proposed by Baillie et al. (1996). For realized volatility the use of a fractionally integrated autoregressive moving average (ARFIMA) process was advocated, for example, by Andersen et al. (2003). The ARFIMA \((p,\, q)\) model is given by

$$\begin{aligned} \phi (L)(1-L)^d(y_n-\mu )=\psi (L)u_n, \end{aligned}$$
(14.16)

with \(\phi (L) = 1-\phi _1L- \ldots -\phi _pL^p,\ \psi (L) = 1+\psi _1L+\ldots \psi _qL^q\), and d denoting the fractional difference parameter. Moreover, \(u_n\) is usually assumed to be a Gaussian white noise process, and yn denotes either the realized volatility (see Koopman et al. 2005) or its logarithmic transformation. Several extensions of the realized volatility ARFIMA model have been proposed, accounting, for example, for leverage effects (see Martens et al. 2004), for non-Gaussianity of (log) realized volatility or for time-variation in the volatility of realized volatility (see Corsi et al. 2008). Generally the empirical results show significant improvements in the point forecasts of volatility when using ARFIMA rather than GARCH-type models.

An alternative model for realized volatility has been suggested by Corsi (2009). The so-called heterogeneous autoregressive (HAR) model of realized volatility approximates the long-memory pattern by a sum of multi-period volatility components. The simulation results in Corsi (2009) show, that the HAR model can quite adequately reproduce the hyperbolic decay in the sample autocorrelation function of realized volatility even if the number of volatility components is small. For the HAR model, let the kperiod realized volatility component be defined by the average of the single-period realized volatilities, i.e.,

$$\begin{aligned} RV _{n+1-k:n}=\frac{1}{k}\sum \limits _{j=1}^k RV _{n-j}. \end{aligned}$$
(14.17)

The HAR model with the so-defined daily, weekly and monthly realizedvolatility components, is given by

$$\begin{aligned} \log RV _n= & {} \alpha _0+\alpha _d\log RV _{n-1}+\alpha _w\log RV _{n-5:n-1}\nonumber \\&+\alpha _m\log RV _{n-21:n-1}+u_n, \end{aligned}$$
(14.18)

with \(u_n\) typically being a Gaussian white noise. The HAR model has become very popular due to its simplicity in estimation and its excellent in-sample fit and predictive ability (see e.g. Andersen et al. 2003; Corsi et al. 2008). Several extensions exist and deal, for example, with the inclusion of jump measures (see Andersen et al. 2003) or non-linear specifications based on neural networks (see Hillebrand and Medeiros 2007).

Alternative realized volatility models have been proposed in, e.g., Barndorff-Nielsen and Shephard (2002a), who consider a superposition of Ornstein Uhlenbeck processes, and in Deo et al. (2006), who specify a long-memory stochastic volatility model. A recent and comprehensive review on realized volatility models can also be found in McAleer and Medeiros (2008b).

5 Time-Varying Betas

So far, our discussion focused on the measurement and modeling of the volatility of a financial asset using high-frequency transaction data. From a pricing perspective, however, systematic risk is most important. In this section, we therefore discuss, how high-frequency information can be used for the evaluation and modeling of systematic risk. A common measure for the systematic risk is given by the so-called (market) beta, which represents the sensitivity of a financial asset to movements of the overall market. As the beta plays a crucial role in asset pricing, investment decisions, and the evaluation of the performance of asset managers, a precise estimate and forecast of betas is indispensable. While the unconditional capital asset pricing model implies a linear and stable relationship between the asset’s return and the systematic risk factor, i.e., the return of the market, empirical results suggest that the beta is time-varying, see, for example, Bos and Newbold (1984), and Fabozzi and Francis (1978). Similar evidence has been found for multi-factor asset pricing models, where the factor loadings seem to be time-varying rather than constant. A large amount of research has therefore been devoted to conditional CAPM and APT models, which allow for time-varying factor loadings, see, for example, Dumas and Solnik (1995), Ferson and Harvey (1991), Ferson and Harvey (1993), and Ferson and Korajczyk (1995).

5.1 The Conditional CAPM

Below we consider the general form of the conditional CAPM. A similar discussion for multi-factor models can be found in Bollerslev and Zhang (2003). Assume that the continuously compounded return of a financial asset i from period n to \(n + 1\) is generated by the following process

$$\begin{aligned} r_{i;n+1}=\alpha _{i;n+1|n}+\beta _{i;n+1|n}r_{m;n+1}+u_{i;n+1}, \end{aligned}$$
(14.19)

with \(r_{m;n+1}\) denoting the excess market return and \(\alpha _{n+1|n}\) denoting the intercept that may be time-varying conditional on the information set available at time n, as indicated by the subscript. The idiosyncratic risk \(u_{n+1}\) is serially uncorrelated, \(\textsf {E}_n(u_{n+1}) = 0\), but may exhibit conditionally time-varying variance. Note that \(\textsf {E}_n(\cdot )\) denotes the expectation conditional on the information set available at time n. Moreover, we assume that \(\textsf {E}(r_{m;n+1}u_{n+1}) = 0\) for all n. The conditional beta coefficient of the CAPM regression (14.19) is defined as

$$\begin{aligned} \beta _{i;n+1|n}=\frac{\mathrm{Cov}(r_{i;n+1},r_{m;n+1})}{\textsf {Var}(r_{i;n+1})}. \end{aligned}$$
(14.20)

Now, assume that lending and borrowing at a one-period risk-free rate \(r_{f;n}\) is possible. Then, the arbitrage-pricing theory implies that the conditional expectation of the next period’s return at time n is given by

$$\begin{aligned} \textsf {E}_\textsf {n}(r_{i;n+1})=r_{f;n}+\beta _{i;n+1|n}\textsf {E}_\textsf {n}(r_{m;n+1}). \end{aligned}$$
(14.21)

Thus, the computation of the future return of asset i requires to specify how the beta coefficient evolves over time.

The most common approach to allow for time-varying betas is to re-run the CAPM regression in each period based on a sample of 3 or 5 years. We refer to this as the rolling regression (RR) method. More elaborate estimates of the beta can be obtained using the Kalman-filter, which builds on a statespace representation of the conditional CAPM or by specifying a dynamic model for the covariance matrix between the return of asset i and the market return.

5.2 Realized Betas

The evaluation of the in-sample fit and predictive ability of various beta models is also complicated by the unobservability of the true beta. Consequently, model comparisons are usually conducted in terms of implied pricing errors, i.e., \(e_{i,n+1} = \widehat{r}_{i,n+1} - r_{i,n+1}\), with \(\widehat{r}_{i,n+1} = r_{f;n} + \widehat{\beta }_{i;n+1|n} \textsf {E}_n(r_{m;n+1})\). Owing to the discussion on the evaluation of volatility models, the question arises, whether high-frequency data may also be useful for the evaluation of competing beta estimates. The answer is a clear “yes”. In fact, high-frequency based estimates of betas are quite informative for the dynamic behavior of systematic risk. The construction of so-called realized betas is straightforward and builds on realized covariance and realized volatility measures. In particular, denote the realized volatility of the market by \(RV_{m;n}\) and the realized covariance between the market and asset i by \( RC ov_{m,i;n} = \sum ^M_{j=1} r_{i,t_j}r_{m,t_j}\), where \(r_{i,t_j}\) and \(r_{m,t_j}\) denote the j-th high-frequency return of the asset and the market, respectively, during day n. The realized beta is then defined as

$$\begin{aligned} \widehat{\beta }_{HF;i;n}=\frac{ RCov _{m,i;n}}{ RV _{m;n}}. \end{aligned}$$
(14.22)

Barndorff-Nielsen and Shephard (2004) show that the realized beta converges almost surely for all n to the integrated beta over the time period from \(n-1\) to n, i.e., the daily systematic risk associated with the market index. Note that the realized beta can also be obtained from a simple regression of the highfrequency returns of asset i on the high-frequency returns of the market, see, e.g., Andersen et al. (2006). The preciseness of the realized beta estimator can easily be assessed by constructing the \((1-\alpha )\)-percent confidence intervals, which have been derived in Barndorff-Nielsen and Shephard (2004) and are given by

$$\begin{aligned} \widehat{\beta }_{HF;i;n}\pm z_{\alpha /2}\sqrt{\left( \sum \limits _{j=1}^Mr^2_{m,t_j}\right) ^{-2}\widehat{g}_{i;n},} \end{aligned}$$
(14.23)

where \(z_{\alpha /2}\) denotes the \((\alpha /2)\)-quantile of the standard normal distribution,

$$\begin{aligned} \widehat{g}_{i;n}=\sum \limits _{j=1}^Mx_{i;j}^2-\sum \limits _{j=1}^{M-1}x_{i;j}x_{i;j+1}, \end{aligned}$$
(14.24)

and

$$\begin{aligned} x_{i;j}=r_{i,t_j}r_{m,t_j}-{\widehat{\beta }_{HF;i;n}}r^2_{m,t_j}. \end{aligned}$$
(14.25)

The upper panel in Fig. 14.4 presents the time-evolvement of the monthly realized beta for IBM incorporation over the period ranging from 2001 to 2006. We use the Dow Jones Industrial Average Index as the market index and construct the realized betas using 30 min returns. The graph also shows the 95%-confidence intervals of the realized beta estimator. The time-varying nature of systematic risk emerges strikingly from the figure and provides once more evidence for the relevance of its inclusion in asset pricing models.

Fig. 14.4
figure 4

Time evolvement and sample autocorrelation function of the realized volatility for IBM, 2001–2006. XFGbetatsacf

Interestingly, the sample autocorrelation function of the realized betas depicted in the lower panel of Fig. 14.4 indicates significant serial correlation over the short horizon. This dependency can be explored for the prediction of systematic risk. Bollerslev and Zhang (2003), for example, find that an autoregressive model for the realized betas outperforms the RR approach both in terms of forecast accuracy as well as in terms of pricing errors.

6 Summary

We review the usefulness of high-frequency data for measuring and modeling actual volatility at a lower frequency, such as a day. We present the realized volatility as an estimator of the actual volatility along with the practical problems arising in the implementation of this estimator. We show that market microstructure effects induce a bias to the realized volatility and we discuss several approaches for the alleviation of this problem. The realized volatility is a more precise estimator of the actual volatility than the conventionally used daily squared returns, and thus provides more accurate information on the distributional and dynamic properties of volatility. This is important for many financial applications, such as asset pricing, portfolio allocation or risk management. As a consequence, several modeling approaches for realized volatility exist and have been shown to usually outperform traditional GARCH or stochastic volatility models, both in terms of in-sample as well as out-of-sample performance. We further demonstrate the usefulness of the realized variance and covariance estimator for measuring and modeling systematic risk. For the empirical examples provided in this chapter we use tick-by-tick transaction data of one stock of the IBM incorporation and of the DJIA index.