Keywords

1 Introduction

Volatility modelling plays an important role in the analysis of financial time series. The persistence of volatility phenomenon is the most well-established effect exhibited by financial time series. Indeed, the variance of returns exhibits high serial autocorrelation, which becomes evident by looking at periods of high volatility, with large changes in assets returns being followed by large ones as well, and at periods of low volatility in which small changes are followed by small ones. As this observation obviously is of great interest, capturing this effect could be challenging. This is the reason why stochastic volatility (SV) models have bee introduced and undergone a lot of research during the last two decades. Since the seminal papers by [14, 15], the univariate SV model has been widely used and several estimation methods have been introduced (see e.g. [3, 10, 13]). Nevertheless, as pointed out by e.g. [2], assets are linked together or influenced by common unobserved factors, which render the univariate approach too restrictive. It is then crucial to extend the univariate SV model to the multivariate case in order to capture the covariation effect. Several alternatives can be considered to describe the time evolution of the joint distribution of different assets (see e.g. [1, 7, 11]).

In this paper we aim at providing a multivariate SV model, based on a latent structure, in which covariation is accounted for via a factor model. Appropriately accounting for covariation is crucial in terms of portfolio diversification and asset allocation. Indeed, the ultimate goal of this paper is to provide indications on portfolio diversification with minimum risk, in a Markovitz framework.

Parameters estimation could be cumbersome and inference becomes therefore hard. This has lead to a substantial development of sampling-based methods in order to obtain parameter estimates, such as rejection sampling, Markov Chain Monte Carlo and Monte Carlo integration (see e.g. [3, 8, 12]). To reduce the computational burden often involved in sampling procedures, we adopt a nested Griddy Gibbs as sampler. In this way, we avoid the need of simulated density that mimic the shape of the preceding one and of guaranteeing the dominance (as imposed by [10]).

The proposed SV model is then applied to a subset of 5 banks among the best capitalized banks’ stocks traded on the Italian stock market (BIT) between 1 January 1986 and 31 August 2011.

The paper is organized as follows. In Sect. 2 we introduce the model, while its computational aspects are discussed in Sect. 3. Section 4 briefly introduces the data, and the obtained results. Section “Conclusions” concludes.

2 Model Summary

In the following, we consider the SV parameterization introduced in [8, 9]. Let \(\mathbf{r}_{t} = (r_{t}^{(1)},\ldots,r_{t}^{(n)})\) an array of n stock returns, in which \(r_{t}^{(i)}\) is the return for the i-th asset at time t and \(\mathbf{a}_{t} = a(_{1t},\ldots,a_{\mathit{Kt}})\) the vector of K common factors at time t, t = 1, , T. The SV parameterization we consider sets

$$\displaystyle\begin{array}{rcl} \mathbf{r}_{t} =\boldsymbol{\mu } _{t} + \varTheta \mathbf{a}_{t} + \mathbf{e}_{t}& &{}\end{array}$$
(1)
$$\displaystyle\begin{array}{rcl} \mu _{t}^{(i)} =\beta _{ 0}^{(i)} +\beta _{ 1}^{(i)}r_{ t-1}^{(i)} +\ldots +\beta _{ p}^{(i)}r_{ t-p}^{(i)}& &{}\end{array}$$
(2)
$$\displaystyle\begin{array}{rcl} e_{t}^{(i)} = \sqrt{h_{ t}^{(i)}}\epsilon _{t}^{(i)}& &{}\end{array}$$
(3)
$$\displaystyle\begin{array}{rcl} \log (h_{t}^{(i)}) =\alpha _{ 0}^{(i)} +\alpha _{ 1}^{(i)}\log (h_{ t-1}^{(i)}) + v_{ t}^{(i)}.& &{}\end{array}$$
(4)

In this factor model, the matrix Θ is a constant (n × K) matrix of factor loading with K < n, \(\boldsymbol{\mu _{t}} = (\mu _{t}^{(1)},\ldots,\mu _{t}^{(n)})'\), the errors \(\mathbf{e}_{t} \sim N(\mathbf{0},\mathbf{E})\) are serially and mutually independent of all other error terms. The errors ε t (i) and v t (i), i = 1, , n are serially and mutually independent N(0, 1) and N(0, σ v 2, (i)), E(a t ) = 0, Cov(a t ) = I. We have also | α 1 (i) | < 1, so that the factor log-volatility processes are stationary. Furthermore, \(\beta ^{(i)} = (\beta _{1}^{(i)},\ldots,\beta _{p}^{(i)})\) are fixed regression parameters.

It follows from this model that the marginal distribution of the returns is multivariate Gaussian with mean μ t and covariance Θ H t Θ′ + E, where \(\mathbf{H}_{t} = \mathit{diag}(h_{t}^{(1)},\ldots,h_{t}^{(n)})\). Of course, it should also be noted that more complicated dynamics could be introduced in the latent SV process and for our purpose there is no need to estimate a t .

Jacquier et al. [9] proposed to use MCMC methods to estimate model parameters. However, the method has not been implemented for a multivariate financial portfolio. In the following we provide computational details to obtain parameter estimates.

3 Computational Details

Let us denote with \(\omega ^{(i)} = (\sigma _{v}^{2,(i)},\alpha _{0}^{(i)},\alpha _{1}^{(i)})'\) and \(h^{(i)} = (h_{1}^{(i)},\ldots,h_{T}^{(i)})\), \(r^{(i)} = (r_{1}^{(i)},\ldots,r_{T}^{(i)})\), i = 1, , n. The likelihood function can be written as:

$$\displaystyle{ f(r^{(i)}\mid \beta ^{(i)},\omega ^{(i)}) =\int _{\mathbf{ R}^{T}}f(r^{(i)}\mid X^{(i)},\beta ^{(i)},h^{(i)})f(h^{(i)}\mid \omega ^{(i)})\mathit{dh}^{(i)}. }$$
(5)

To tackle this estimation problem we use the MCMC method as in [8].

Let’s describe the steps of the algorithm that implements the MCMC method. First step: for each return i the optimal lag p of the Eq. (2) is selected, independently from the other returns, according to the AIC criterion, fixing \(v_{t}^{(i)} = 0\).

Second step: for each return i the following likelihood function, related to the Eqs. (1)–(4) with \(v_{t}^{(i)} = 0\), is maximized to find the MLE parameters as in [6]:

$$\displaystyle{ -L(\beta ^{(i)},\omega ^{(i)}) = T^{-1}\sum _{ t=1}^{n}a_{ t}^{(i)}/\tilde{h}_{ t}^{(i)} +\log (\tilde{h}_{ t}^{(i)}) }$$
(6)

where

$$\displaystyle{ \begin{array}{ll} a_{t}^{(i)} = r_{t}^{(i)} -\mu _{t}^{(i)} \\ \tilde{h}_{t}^{(i)} =\log (h_{t}^{(i)}) =\alpha _{ 0}^{(i)} +\alpha _{ 1}^{(i)}\tilde{h}_{t-1}^{(i)} \end{array} }$$
(7)

These parameter estimations are chosen as the initial values for the conditional posterior distributions used by the MCMC iterations.

Third step: Steps 1,2 are repeated for each intrinsic values i corresponding to the return i, given by PE (i) ×EPS (i), where PE (i) is the Price to Earnings ratio and EPS (i) is the Earning per share.

As the intrinsic value is considered to give indication on the return, the obtained parameter estimates are chosen as the prior distribution parameters used to define the conditional posterior distributions needed by the MCMC iterations.

Let’s call these prior parameters as follow: \(\beta ^{0,(i)} = (\beta _{1}^{0,(i)},\ldots,\beta _{p}^{0,(i)})'\) and their variances \(A^{0,(i)} = \mathrm{diag}(\sigma _{\beta _{1}^{0,(i)}}^{2},\ldots,\sigma _{\gamma ^{0,(i)}}^{2})\) for the mean parameters, \(\omega ^{0,(i)} = (\sigma _{v}^{2,0,(i)},\alpha _{0}^{0,(i)},\alpha _{1}^{0,(i)})'\) and \(C^{0,(i)} = \mathrm{diag}(\sigma _{\alpha _{0}^{0,(i)}}^{2},\sigma _{\alpha _{1}^{0,(i)}}^{2})\) for the volatility parameters.

Moreover the prior distributions are hypothesized multivariate normal for \(\beta ^{(i)} \sim N(\beta ^{0,(i)},A^{0,(i)})\) and α (i) ∼ N(α 0, (i), C 0, (i)), inverted Chi squared for σ v 2, (i), that’s to say \(T\lambda /\sigma _{v}^{2,(i)} \sim \chi _{T}^{2}\), with λ a scale parameter.

Fourth step: We consider simulation-based methods. The MCMC Gibbs sampling estimation of the model (1)–(4), after combining the prior distributions with the likelihood using the Bayes’ rule, consists in drawing random samples from the conditional posterior distributions

$$\displaystyle{ \begin{array}{ll} f(\beta ^{(i)}\vert r^{(i)},x^{(i)},h^{(i)}) \sim N(\beta ^{{\ast},(i)},A^{{\ast},(i)}) \\ f(h_{t}^{(i)}\vert r^{(i)},x^{(i)},h^{(i)},\beta ^{(i)},\omega ^{(i)}) \\ f(\alpha ^{(i)}\vert h^{(i)},\sigma _{v}^{2,(i)}) \sim N(\alpha ^{{\ast},(i)},C^{{\ast},(i)}) \\ f(d/\sigma _{v}^{2,(i)}\vert h^{(i)},\alpha ^{(i)}) \sim \chi _{2T-1}^{2}\end{array} }$$
(8)

in a sequence from initial value of the conditioning variables, with step by step substitutions of the new sampled values to the previous ones, until a number of iteration g is reached.

That’s is to say, at the MCMC Gibbs iteration j with j = 1, , g:

  1. 1.

    we draw a random sample[j] β (i) from: \(f(\beta ^{(i)}\vert r^{(i)},x^{(i)},_{[j-1]}h^{(i)})\)

  2. 2.

    we draw a random sample \(_{[j]}h_{t}^{(i)}\) from: \(f(h_{t}^{(i)}\vert r^{(i)},x^{(i)},_{[j-1]}h_{-t}^{(i)},_{[j]}\beta ^{(i)},_{[j-1]}\omega ^{(i)})\)

  3. 3.

    we draw a random sample[j] α (i) from: \(f(\alpha ^{(i)}\vert _{[j]}h^{(i)},_{[j-1]}\sigma _{v}^{2,(i)})\)

  4. 4.

    we draw a random sample \(_{[j]}\sigma _{v}^{2,(i)}\) from: \(f(d/\sigma _{v}^{2,(i)}\vert _{[j]}h^{(i)},_{[j]}\alpha ^{(i)})\)

This completes a MCMC Gibbs iteration and current parameters values are

$$\displaystyle{(_{[j]}\beta ^{(i)},_{ [j]}h_{1,\ldots,t}^{(i)},_{ [j]}\alpha ^{(i)},_{ [j]}\sigma _{v}^{2,(i)})}$$

In this way we obtain random samples \(\{_{[j]}\beta ^{(i)}\}_{j=g_{0},\ldots,g}\),\(\{_{[j]}h^{(i)}\}_{j=g_{0},\ldots,g}\) \(\{_{[j]}\alpha ^{(i)}\}_{j=g_{0},\ldots,g}\), \(\{_{[j]}\sigma _{v}^{2,(i)}\}_{j=g_{0},\ldots,g}\) that can be used to make inference.

Our estimations are the point estimation sample means of the previous random samples after eliminating the first g 0 − 1 values, that’s to say

$$\displaystyle{ \begin{array}{ll} \hat{\beta }^{(i)} & = \frac{1} {g-g_{0}} \sum \nolimits _{j=g_{0},\ldots,g\ [j]}\beta ^{(i)} \\ \hat{h}_{t}^{(i)} & = \frac{1} {g-g_{0}} \sum \nolimits _{j=g_{0},\ldots,g\ [j]}h_{t}^{(i)} \\ \hat{\alpha }^{(i)} & = \frac{1} {g-g_{0}} \sum \nolimits _{j=g_{0},\ldots,g\ [j]}\alpha ^{(i)} \\ \hat{\sigma }_{v}^{2,(i)} & = \frac{1} {g-g_{0}} \sum _{j=g_{0},\ldots,g\ [j]}\sigma _{v}^{2,(i)} \end{array} }$$

The value of g 0 is chosen so that the estimation \((\hat{\beta }^{(i)},\hat{h}^{(i)},\hat{\alpha }^{(i)},\hat{\sigma }_{v}^{2,(i)})\) of the parameters \((\beta ^{(i)},h^{(i)},\alpha ^{(i)},\sigma _{v}^{2,(i)})\) is stable in the sense that after g 0 the means obtained by adding one by one the successive random sample of the Gibbs are almost identical.

In the Eq. (8) the Bayes’ rule gives \(A^{{\ast},(i)} = \left (\sum _{t=1}^{T}x_{0,t}^{(i)}x_{0,t}^{(i)'} + (A^{0,(i)})^{-1}\right )^{-1}\), \(x_{0,t}^{(i)} = x_{t}^{(i)}/\sqrt{h_{t }^{(i)}}\) \(\beta ^{{\ast},(i)} = A^{{\ast},(i)}\left (\sum _{t=1}^{T}x_{0,t}^{(i)}r_{0,t}^{(i)} + (A^{0,(i)})^{-1}\beta ^{0,(i)}\right )\), \(r_{0,t}^{(i)} = r_{t}^{(i)}/\sqrt{h_{t }^{(i)}}\),\(C^{{\ast},(i)} = \left (\sum _{t=2}^{T}y_{t}^{(i)}y_{t}^{(i)'}/\sigma _{v}^{2,(i)} + (C^{0,(i)})^{-1}\right )^{-1}\), \(y_{t}^{(i)} = (1,\mathit{ln}(h_{t}^{(i)}))'\) \(\alpha ^{{\ast},(i)} = C^{{\ast},(i)}\left (\sum _{t=2}^{T}y_{t}^{(i)}\mathit{ln}(h_{t}^{(i)})/\sigma _{v}^{2,(i)} + (C^{0,(i)})^{-1}\alpha ^{0,(i)}\right )\) \(d = T\lambda +\sum _{ t=2}^{T}v_{t}^{2,(i)}\), r t (i) is the compound return,x j, t (i) is its lagged value.

The posterior distribution \(f(h_{t}^{(i)}\vert r^{(i)},x^{(i)},h^{(i)},\beta ^{(i)},\omega ^{(i)})\) is a non standard one even if its density is known up to a normalizing constant [8].

Therefore a nested Gibbs sampler of type Griddy is implemented in the following way:

  1. 1.

    a grid of values for h t (i) is selected, say, \(h_{t,1}^{(i)} \leq h_{t,2}^{(i)} \leq \ldots \leq h_{t,m}^{(i)}\); its posterior distribution is evaluated on this values to obtain \(w_{s} = f(h_{t,s}^{(i)}\vert r^{(i)},x^{(i)},h_{-t}^{(i)},\beta ^{(i)},\omega ^{(i)})\), s = 1, , m.

  2. 2.

    The w s are used to obtain an approximation of the inverse cumulative distribution function of \(f(h_{t,s}^{(i)}\vert r^{(i)},x^{(i)},h_{-t}^{(i)},\beta ^{(i)},\omega ^{(i)})\).

  3. 3.

    A uniform random variable between 0 and 1 is drawn and transformed via the preceding step 2 to obtain a random drawn for h t (i).

Fifth step: the estimation \((\hat{v}_{i,i}(T + 1\vert \varOmega _{T}))_{i\in \{1,\ldots,n\}}\) of the volatility matrix \((v_{i,i}(T + 1\vert \varOmega _{T}))_{i\in \{1,\ldots,n\}}\), that’s to say \(v_{i,i}(T + 1) = h_{T+1}^{(i)}\),will be obtained in the following way, at the iteration j of the Gibbs sampler, j = g 0, , g:

  1. 1.

    we draw a random sample \(v_{T+1}^{(i)}\) from \(N(0,_{[j]}\sigma _{v}^{2,(i)})\) and the Eq. (4) with[j] β (i) is used to compute \(_{[j]}h_{T+1}^{(i)}\);

  2. 2.

    we draw a random sample ε T+1 (i) from N(0, 1) to obtain \(e_{T+1}^{(i)} = \sqrt{_{[j] } h_{T+1 }^{(i)}}\epsilon _{T+1}^{(i)}\) and the Eq. (2) with[j] α (i) is used to compute \(_{[j]}r_{T+1}^{(i)}\);

    In this way we obtain a random sample \(\{_{[j]}h_{T+1}^{(i)}\}_{j=g_{0},\ldots,g}\) and a random sample \(\{_{[j]}r_{T+1}^{(i)}\}_{j=g_{0},\ldots,g}\) that can be used to make inference.

    Our estimation is the point forecast of the previous two random samples, that’s to say \(\hat{h}_{T+1}^{(i)} = \frac{1} {g-g_{0}} \sum _{j=g_{0},\ldots,g}\) \(_{[j]}h_{T+1}^{(i)}\) and \(\hat{r}_{T+1}^{(i)} = \frac{1} {g-g_{0}} \sum _{j=g_{0},\ldots,g}\) \(_{[j]}r_{T+1}^{(i)}\).

Sixth step: To estimate the off-diagonal elements of V, we consider the multivariate Latent Factor model (see [4]).

Seventh step: The Markowitz problem can be foretasted at time T + 1 using the preceding estimation of v i, j and r (i) that we called \(\hat{v}_{i,j}(T + 1)\) and \(\hat{r}_{T+1}^{(i)}\), by solving through a quadratic programming method, the following:

$$\displaystyle{ \mathit{min}_{\gamma \in \mathbf{R}}\{\gamma '\hat{V }(T + 1)\gamma \:\gamma ' \cdot \underline{ 1} = 1,\gamma ' \cdot \hat{ r}_{T+1} = R_{p},\gamma \geq 0\} }$$
(9)

where γ = (γ 1, , γ n ) and \(R_{p} \in [\mathit{min}_{\{i:i=1,\ldots,n\}}\hat{r}_{T+1}^{(i)},\mathit{max}_{\{i:i=1,\ldots,n\}}\hat{r}_{T+1}^{(i)}]\).

4 Data and Results

The model is applied to a subset of the entire stocks’ universe among the series of data regarding the best capitalized 5 banks’ stocks traded on the Italian stock market (BIT) between 1 January 1986 and 31 August 2011. Data are shown in Fig. 1.

Fig. 1
figure 1

Data description

This figure shows the histograms and the fitted normal density. Even if the empirical distributions of the returns are symmetry and uni-modal, the fitted normal curves are not enough similar to the empirical counterparts in the tails. Tests of Jarque–Bera for normality reject the null hypothesis H 0 of normality at 95 %.

Moreover tests Ljung–Box for the squared of residuals e t (i) of Eqs. (1)–(2), reject the null hypothesis H 0 of no ARCH effects at 95 %. Therefore it is necessary to include the Eqs. (3)–(4), that’s to say the SV model part, in order to obtain an unconditional distribution of the residuals e t (i) that has heavier tails, an excess of kurtosis with respect to a gaussian distribution. This is in agreement with the financial data in hand.With respect to the model (1)–(2) setting v t (i) ≡ 0,where r t (i) is the intrinsic value, the explanatory variables x j, t (i) are the lagged values r tj (i) and the exogenous variable z t (i) is the market index, we give the optimal lags in Table 1, which minimize the AIC.

Table 1 Optimal AIC lags

The posterior distribution parameters, calculated as means of the last \(g - g_{0} =\) 100 iterations of sampling from the MCMC posterior conditional distribution are as follows (Tables 2 and 3):

Table 2 \(\hat{\beta }^{{\ast},(i)} = (\hat{\beta }_{0}^{(i)},\ldots,\hat{\beta }_{p}^{(i)})'\) posterior distribution parameters (stock i)
Table 3 \(\hat{\omega }^{(i)} = (\hat{\sigma }_{v}^{2,(i)},\hat{\alpha }_{0}^{(i)},\hat{\alpha }_{1}^{(i)})'\) posterior distribution parameters (stock i)

From Table 2 it can be seen that \(\hat{\beta }_{0}^{(i)} \simeq 0\) so there is no constant term in the mean Eqs. (1)–(2). As \(\hat{\beta }_{k}^{(i)}\neq 0,k> 0\) it can be said that return is dependent on its past values. Moreover, from Table 3 it can be seen that the volatility is dependent from its past as each \(\hat{\alpha }_{1}^{(i)}\neq 0\) in the Eq. (4). Lastly \(\hat{\sigma }_{v}^{2,(i)}> 0\) is not negligible so the SV model, which introduces v t (i), is capable to improve a pure ARCH model for the volatility.

The estimated variance-covariance (risk) matrix is given in Table 4.

Table 4 Covariance risk matrix estimation

As expected, it can be seen in Table 4 that \(0.47 <\hat{ v}_{i,j} <0.70,i\neq j\), so there are positive correlations among the series. Indeed the series belong to the same sector.

It can be seen that the variances dominate the covariances which are all positive. Thus, it is possible to find portfolios that have lower risk than either single asset. Among those ones we choose the minimum risk portfolio.

In order to provide further insights on the association structure we may look at the estimated latent factors (Table 5). The first latent factor can be associated with Unicredit, BPM and Intesa San Paolo, whilst the second one is mainly related to Credito Emiliano. This can be seen in Fig. 2, where we call Component i the \(\hat{\varTheta }_{i},i = 1,2\) and the numbers are the stocks in the same order of Table 5.

Table 5 Latent factor loadings estimates
Fig. 2
figure 2

Latent factor loadings

The possible double clustering suggested by the Fig. 2 could drive another portfolio diversification in order to take into consideration the different dependency (loading) each stock’s return has of the common latent factors.

Of course, as a central issue, we have to solve the quadratic programming of Markowitz in order to optimize our portfolio. We propose the following optimal (with minimum risk) fractions to invest in each stock as the numerical solution of the optimization problem given in the Eq. (9) (Table 6).

Table 6 \(\hat{\gamma }_{(i)}\) optimal (with minimum risk) fractions

By depicting the optimal fractions in Fig. 3, it can be seen that a good diversification among the stocks is obtained.

Fig. 3
figure 3

Optimal fractions

The optimal fractions \(\hat{\gamma }_{(i)}\) give a risk of 0.90724 and a monthly portfolio return of 0.035194. Different portfolio choice are possible that can get greater return or better latent shocks warranty at the expense of a greater risk. Thus in a risk averse view, an investor should choose the minimum risk portfolio suggested.

Conclusions

We have conducted an empirical investigation of stochastic volatility of major Italian banks, by introducing a computational feasible algorithm based on simulation techniques. The proposed estimation methodology is easily implementable and this is an important step forward on multivariate volatility estimation, since the likelihood function of stochastic volatility models is not easily calculable. The procedure proposed in this work attempts to combine the simplicity of the factor model with the sophistication of stochastic volatility procedures. Open problems remain, primarily in the modelling of multivariate heavy-tailed or skewed error distributions, as well as the computational burden required in the estimation step in the modelling of high dimensional data. In time, further significant developments can be achieved by introducing a time-varying latent structure such as parsimonious hidden Markov models, which are able to reduce the curse of dimensionality of the considered problem and account for well-know stylized facts arising in the stock returns modelling.