Academia.eduAcademia.edu

Bayesian unit-root tests for Stochastic Volatility models

2009, Statistical Methodology

Abstract

In this article, we consider Bayesian inference procedures to test for a unit root in Stochastic Volatility (SV) models. Unit-root tests for the persistence parameter of the SV models, based on the Bayes Factor (BF), have been recently introduced in the literature. In contrast, we propose a flexible class of priors that is noninformative over the entire support of the persistence parameter (including the non-stationarity region). In addition, we show that our model fitting procedure is computationally efficient (using the software WinBUGS). Finally, we show that our proposed test procedures have good frequentist properties in terms of achieving high statistical power, while maintaining low total error rates. We illustrate the above features of our method by extensive simulation studies, followed by an application to a real data set on exchange rates.

Statistical Methodology 6 (2009) 189–201 Contents lists available at ScienceDirect Statistical Methodology journal homepage: www.elsevier.com/locate/stamet Bayesian unit-root tests for Stochastic Volatility models Zeynep I. Kalaylıoğlu a , Sujit K. Ghosh b,∗ a Department of Statistics, Middle East Technical University, Ankara, Turkey b Department of Statistics, North Carolina State University, Raleigh, NC 27695-8203, USA article info Article history: Received 23 April 2007 Received in revised form 10 May 2008 Accepted 25 July 2008 Keywords: BUGS software Gibbs sampling Markov Chain Monte Carlo Stochastic Volatility models Unit-root test a b s t r a c t In this article, we consider Bayesian inference procedures to test for a unit root in Stochastic Volatility (SV) models. Unit-root tests for the persistence parameter of the SV models, based on the Bayes Factor (BF), have been recently introduced in the literature. In contrast, we propose a flexible class of priors that is noninformative over the entire support of the persistence parameter (including the non-stationarity region). In addition, we show that our model fitting procedure is computationally efficient (using the software WinBUGS). Finally, we show that our proposed test procedures have good frequentist properties in terms of achieving high statistical power, while maintaining low total error rates. We illustrate the above features of our method by extensive simulation studies, followed by an application to a real data set on exchange rates. © 2008 Elsevier B.V. All rights reserved. 1. Introduction In finance, it is very common to see that the mean corrected return on holding an asset has time dependent variation (i.e. volatility). Researchers have been interested in modeling the time dependent feature of unobserved volatility. A model that is commonly used to model such features, is known as the Autoregressive Conditional Heteroscedastic (ARCH) model (see [4]). An extension is to consider Generalized ARCH (GARCH) models, which have been found very popular to model the volatility over time. There is an extensive literature on estimation and testing procedures for GARCH models (e.g., see [2]). However, in ARCH and GARCH models, given the past observations (returns), volatility is a deterministic function of the past observations. This feature may not be appropriate for some real ∗ Corresponding author. Tel.: +1 9195151950; fax: +1 9195157591. E-mail addresses: [email protected] (Z.I. Kalaylıoğlu), [email protected] (S.K. Ghosh). 1572-3127/$ – see front matter © 2008 Elsevier B.V. All rights reserved. doi:10.1016/j.stamet.2008.07.002 190 Z.I. Kalaylıoğlu, S.K. Ghosh / Statistical Methodology 6 (2009) 189–201 data sets such as stock exchange rates, where we expect the volatilities to vary stochastically as a function of past observations. An alternative approach is to use Stochastic Volatility (SV) models. In the SV models, the logarithm of the unobserved volatility is modeled as a stochastic process. A popular choice is to model the log-volatility as an AR(1) process. Extensions to higher order AR process may also be considered. However, for illustrative purpose only, we restrict our attention to only AR(1) processes for log-volatilities within the SV models. Jacquier, Polson and Rossi [10] developed Bayesian and frequentist methods for estimating the parameters of SV models. The posterior distributions of the parameters (and hence the Bayes estimates) are obtained using a Markov Chain Monte Carlo (MCMC) algorithm that is based on treating the unobserved volatilities as missing observations. The maximum likelihood estimation of the parameters as well as the method of moment estimation are also considered by Jacquier et al. [10]. They compare these estimation procedures for SV models based on the bias and the root mean squared errors of the estimators. Their simulation results show that their Bayesian estimation technique performs better than the other two procedures for SV models. They also illustrate their methodology by fitting four series; equal-weighted (EW) and value-weighted (VW) indices of NYSE, and three decile portfolios corresponding to the 1, 5, and 10 deciles of stock. By using their MCMC algorithm they obtain the Bayes estimators under squared error loss, i.e. the posterior mean of the persistence parameter (henceforth denoted by φ ) of the volatility processes. It turns out that the posterior mean of φ for the four series are 0.95 (s.d. 0.013), 0.93 (s.d. 0.016), 0.91(s.d. 0.046), and 0.93 (s.d. 0.022), respectively, along with posterior standard deviations (s.d.) in parenthesis. The striking feature being that in all four series φ is expected to be close to unity. It is to be noted that the priors used for φ had support only on the stationarity region, i.e. uniform on the interval (0, 1). We suspect that such priors down-weight the posterior mass of φ at unity. In Section 2 we introduce priors that overcome such limitations of previous Bayesian procedures. There are several other real data examples in the literature where this φ parameter of the process for the volatility is estimated to be close to unity. For instance, see [13,14]. This raises the question of the presence of a unit-root in the SV models. Unfortunately this feature of the SV models has not received much attention in the literature. There has been only few attempts to test for the presence of a unit root in SV models. Harvey, Ruiz and Shephard [9] considered a SV model to fit four exchange rates. They obtained the Quasi-Maximum Likelihood (QML) estimates of the φ parameter of the volatility process to be 0.9912 (s.e. 0.007), 0.9646 (s.e. 0.021), 0.9948 (s.e. 0.005), and 0.9575 (s.e. 0.002), respectively along with the standard error (s.e.) of the estimates in parenthesis. They test the null hypothesis of a unit root (i.e. H0 : φ = 1) by applying Augmented Dickey–Fuller (ADF) unit-root test (see [19]) to the log-squared returns for all the four series. The hypothesis of a unit-root is strongly rejected at the 1% level for all the series, despite the fact that the estimates were very close to unity. Based on this result, they conclude that the reliability of the classical ADF test is questionable for SV models. We observe similar problems with some of the frequentist unit-root tests for SV models (not reported in this article, however see Section 7). Bayesian data analysis is becoming more and more appealing because of its flexibility in handling complex models that typically involve many parameters and/or missing observations. This is one of the reasons that we consider a Bayesian perspective for unit-root testing. It may be noted that the data generated from a SV model has the same number of unobserved volatilities as the sample size. So and Li [21] propose a Bayesian unit-root test procedure to test for the unit root in SV models. Their rejection criterion is to reject the null hypothesis if the posterior odds in favor of H0 is less than or equal to one. Based on their simulation results, they argue that one can get reliable results from their Bayesian unitroot testing procedure. However, using a prior density that is defined on an interval that excludes the point one is not suitable for statistical inference where the main goal is testing for a unit root. Also the MCMC computation method proposed by So and Li [21] is based on approximating the error process of log-volatilities by discrete mixture of several (seven) normals. Their method is computationally inefficient as the efficiency is lost in estimating several nuisance parameters of the mixing normals. In contrast, we develop methods that do not require such approximation, use a flexible class of prior distributions and make use of a computation method that is easy to implement in practice. Z.I. Kalaylıoğlu, S.K. Ghosh / Statistical Methodology 6 (2009) 189–201 191 Prior densities in Bayesian inference play a crucial role. Suitably chosen non-informative priors often yields estimators with good frequentist properties. We find the model building procedure to be quite flexible from a Bayesian perspective, but find it more convenient to validate its use from a frequentist perspective. In this article, we address the problems associated with employing different prior densities for the φ parameter of the SV models. In particular, we suggest the use of a noninformative prior that is supported on a space that includes one with positive probability. We also illustrate how our proposed method can be implemented in practice using a public domain software like WinBUGS. The rest of the article is organized as follows; Section 2 describes the SV model, and some alternative assumptions concerning this model. Section 3 presents the motivation for our choice of prior densities for the parameters and Bayesian unit-root testing methods that we suggest for SV models. Section 4 describes the proposed Gibbs sampling algorithm, and how we implement it in WinBUGS. Extensive simulation studies are presented in Section 5 to validate the use of our proposed method from a frequentist perspective. We illustrate our proposed method by applying the testing procedure to a real data set on exchange rates in Section 6. Finally, in Section 7, we indicate few limitations of our approach and some directions for further research in SV models. 2. The Stochastic Volatility (SV) model The widely used SV models can be described in its simplest form as follows. Let rt denote the mean corrected return observed at time t. As an example of return data rt , first consider pt which denotes, say, the Turkish Lira/U.S. Dollar exchangeP rate at time t = 1, 2, . . . , n. Then mean corrected return, n pi p ). We analyze such data sets in Section 6. Let rt , can be computed as rt = log p t − 1n i=1 (log p t −1 i−1 Vt = Var(rt ) be the unobserved volatility at time t. In SV models it is customary to model the logvolatilities, ht = log Vt in contrast to ARCH and GARCH models. The SV model is represented as, 1 rt = ut e 2 ht ht − µ = φ(ht −1 − µ) + σ ηt , t = 2, . . . , n (1) h1 = µ1 + σ1 η1 , where ηt ’s and ut ’s are assumed to be a sequence of iid realizations from a distribution with mean zero and variance one. The parameters µ1 and σ1 of the initial distribution are chosen (as a function of φ and σ ) appropriately, so as make the process ht stationary when |φ| < 1. When φ = 1, we set µ1 = µ and σ1 = σ . Note that stationarity of the process ht implies stationarity of the process rt . In this article the ut ’s and ηt ’s are also assumed to be independent. However several extensions of the above model in (1), are possible. For instance, Sandmann and Koopman [20] consider a SV model where ut ’s and ηt ’s are assumed to be correlated. It is quite straightforward to extend the AR(1) process of the latent ht ’s to a general AR(p) process. The main parameter of interest in this model is the persistence parameter, φ and our focus is to test the unit-root hypothesis, H0 : φ = 1 against the alternative Ha : |φ| < 1. When there is a unit root in the log-volatility process, shocks to volatility do not decay rapidly. That is, the effect of past shocks on current volatility (conditional variance of returns) remain persistent for long periods. Persistence of shocks to volatility has important implications in economics and finance. High persistence of shocks to volatility increases the fluctuation in the volatility (change in the return over time). As a result, the business environment becomes more uncertain. That causes the market to plunge. See Pyndick [16,17], Poterba and Summers [15], Chou [3] and Bollerslev and Engle [2], who studied the volatility and its relationship with market fluctuations. Tests for a unit-root in SV models continue to receive much attention. Jacquier et al. [10] suggested several methods to estimate the parameter, θ = (µ, φ, σ 2 ) of model (1). From a Bayesian perspective, the above model can be viewed as a three-level hierarchical model. In the first level of hierarchy, the conditional distribution of the data (viz., rt ’s) is specified given the unobserved volatility (ht ’s). At the second level, the conditional distribution of the unobserved volatility is specified given the parameter θ . Finally at the last level, prior distributions are specified. 192 Z.I. Kalaylıoğlu, S.K. Ghosh / Statistical Methodology 6 (2009) 189–201 Jacquier et al. [10] obtained the Bayes estimates of θ under conjugate prior specification and more importantly under the assumption that Pr(0 < φ < 1) = 1. We argue that such priors may not be suitable for testing the unit-root hypothesis. [13] also obtained parameter estimates under model (1) using priors that puts a zero mass at φ = 1. Kim et al. [13] provide three efficient MCMC methods to obtain samples from the posterior distribution of θ , using state-space representation of the above model. So and Li [21] proposed a Bayesian unit-root test procedure to test for the unit root in the SV model in (1). They consider four different prior densities for the persistence parameter φ . The first one is the uniform prior on the interval (0, 1) and they claimed this prior to be ‘‘non-informative’’. The second prior that they used is the truncated Normal density on the interval (0, 1). The other two priors were Beta(10, 1) and Beta(20, 2) distributions. They treat the testing problem as comparing two models under the null hypothesis and the alternative hypothesis. They select the model based on the value of the posterior odds ratio. We compare our methods to their approach in Section 5. However to our knowledge, none of the previous methods have used a prior that puts a positive mass at unity, and hence the use of such priors as ‘‘non-informative’’ is questionable. We address this issue specifically in Section 3. 3. Prior distributions We consider several prior densities for the persistence parameter, φ , of the SV model in (1). The main reason to consider more than one prior density is to perform a sensitivity analysis of our Bayesian procedures. In the literature, continuous densities such as uniform, truncated normal, and beta distributions that are defined over the interval (0,1) have been used as prior densities for φ [10,21]. For estimation purpose such densities might be useful, but for testing the unit-root hypothesis H0 : φ = 1, any continuous distribution with support on only (0, 1) interval can produce results biased in favor of stationarity. For this reason, we broaden the support, and consider prior densities that either assign a positive mass at unity or a density that is defined on an interval that includes the unity (say (−a, a), with a > 1). Specifically, we propose to use two new types of prior densities for φ . (i) The mixed priors. Let B be a Bernoulli random variable with success probability p = P (B = 1). In addition, let U be a continuous random variable having support on the open interval (0, 1). U is assumed to be independent of B. We set the persistence parameter φ = B + (1 − B)U. Here, p can be viewed as the mixing probability. The mixing probability can be assumed to be a known constant or an unknown parameter. When p is assumed to be a constant, the marginal prior density of φ is fφ (φ) = pI (φ = 1) + (1 − p)fU (φ)I (0 < φ < 1), (2) where I (A) = 1 if A is true, otherwise I (A) = 0. However when p is assumed to be a parameter with hyper-prior density say fp (·), then, f (φ|p) = pI (φ = 1) + (1 − p)fU (φ)I (0 < φ < 1) and fφ (φ) = Z f (φ|p)fp (p) dp = E [p]I (φ = 1) + (1 − E [p])fU (φ)I (0 < φ < 1). That is, the marginal distribution of φ is also a mixture distribution as given in (2) with p = E [p], where E [p] denotes the expectation of p w.r.t. fp (·). Here, U can be assumed to have a Beta distribution defined over (0, 1) with parameters αU and βU . In this article, we set αU = βU = 1. That is, we take U ∼ U (0, 1), and denote this mixed uniform (MU) density as MU (p). In our simulations (see Section 5) we study the sensitivity of the constant p on the posterior distribution of φ . Z.I. Kalaylıoğlu, S.K. Ghosh / Statistical Methodology 6 (2009) 189–201 193 (ii) The flat priors. In addition to the above mixed priors, we suggest using flat priors that are defined on a region that includes the point one. We suggest the use of priors in the form of U (a, b) where b > 1. For example, U (−2, 2) and U (0, 1 + ǫ). In the latter case, ǫ > 0 can be chosen arbitrarily. We choose ǫ such that the expected value of φ having this density is equal to the expected value of φ having a mixed density with mixing probability p as in MU (p). Another alternative is to choose ǫ such that P (1 ≤ φ < 1 + ǫ) = p, where the probability is taken with respect to the U(0,1+ǫ ) distribution. p In this case, ǫ = 1−p . Finally to complete the full model specification, we assume that apriori the parameters φ, µ and σ are independent. Note that prior independence does not necessarily imply posterior independence and certainly not for our version of SV models. The prior we use for µ, the population mean of the log-volatilities, is N (µ0 , σ02 ) where µ0 and σ02 are assumed to be known quantities. We choose the hyper-parameters µ0 and σ0 so as to make the prior belief have a minimal influence on the posterior distribution. For instance, we can set σ02 = 106 to approximate a diffuse prior. For the parameter σ 2 , we use an Inverse-Gamma (IG) distribution, IG(a0 , b0 ), with mean b0 /(a0 − 1) and variance b20 (a0 −1)2 (a0 −2) , respectively. Again to have minimal prior influence we can set a0 = 2 + 10−10 and b0 = 0.1 (also suggested by So and Li [21]). We study the sensitivity of these priors in Section 5. 4. Testing the unit-root hypothesis We use two approaches to test the unit-root hypothesis H0 : φ = 1 against the alternative Ha : |φ| < 1. First one is based on constructing an interval estimate for φ based on its marginal posterior distribution. The criteria are to reject the null hypothesis if the number one is not included in the posterior interval. For all our simulation studies, we have used a 95% equal tail credible interval for φ . The second approach is based on the popular Bayes Factor (BF) criteria, which reject the null hypothesis if log10 (BF ) < 0. It is to be noted that our proposed method of computing the BF is different from that of [21] in the sense that we use an exact SV model to obtain the predictive distribution, instead of approximating the errors of volatility processes (ηt ’s) by a discrete mixture of normals. In our simulation studies (see Section 5) we find that the posterior interval method is easier to implement and maintains good frequentist power as compared with the BF criteria. We found the BF for SV models can be slightly more intensive to compute, as it requires evaluation of the conditional density of the returns given the volatilities which are of dimension as the sample n and hence grow with the sample size. We briefly explain the model fitting techniques that we have used for our computations. 4.1. The Gibbs sampling algorithm for SV models Deriving the joint posterior density of the SVM parameters θ = (µ, φ, σ 2 ) amounts to integrating out the unobserved log-volatilities h, which is difficult to perform both analytically and numerically, as the number of unobserved log-volatilities is equal to the sample size. This makes obtaining the posterior inference about the parameters very difficult. At this point, sampling based techniques provide an alternative approach to obtain posterior distribution of the parameters. One class of such techniques is known as the Markov Chain Monte Carlo (MCMC) methods. MCMC methods are being increasingly used for cases where marginal distribution of the random parameters cannot be obtained either analytically, or by numerical integration. MCMC methods consist of sampling random variates from a Markov Chain, such that its stationary distribution is the posterior distribution of the parameter of interest (see [7] and [26] for an excellent overview of the theoretical properties of MCMC). That means, under some regularity conditions, the realizations of this Markov Chain (after some ‘‘burn-in’’ time) can be thought of as realizations sampled from the desired posterior distribution. We use a Gibbs sampler [5], a widely used MCMC method, to obtain dependent samples from the posterior distribution p(θ |r ). To implement the Gibbs sampler, one obtains the full conditional densities, i.e. the conditional density of a component of the vector of parameters given the other components and the observed data. Specifically, we derive the conditional densities, 194 Z.I. Kalaylıoğlu, S.K. Ghosh / Statistical Methodology 6 (2009) 189–201 f (µ|φ, σ 2 , h, r ), f (σ 2 |µ, φ, h, r ), f (φ|µ, σ 2 , h, r ), and f (ht |µ, φ, σ 2 , h−t , r ) as the full conditional densities of µ, σ 2 , φ , and ht in model (1) respectively. Here h−t is the vector of unobserved logvolatilities excluding the one at time t. The full conditional densities of µ and σ 2 are Normal and Inverse Gamma densities respectively. The full conditional density of ht is obtained as, f (ht |µ, φ, ση2 , h−t , r ) ∝ f (rt |ht )f (ht |ht −1 , µ, ση2 , φ)f (ht +1 |ht , µ, ση2 , φ). (3) In this expression, f (rt |ht ) and f (ht |ht −1 , µ, σ 2 , φ) are the probability density functions of N (0, eht ) and N (µ(1 − φ) + φ ht , σ 2 ) respectively. When we use a flat prior distribution, which is in the form of U (a, b) with b > 1 for the persistence parameter φ , the full conditional density of φ is given by, f (φ|µ, σ 2 , h, r ) ∝ p − 1 − φ2e 1 2σ 2 " φ 2 {(h1 −µ)2 + n P t =2 (ht −1 −µ)2 }−2φ n P (ht −µ)(ht −1 −µ) t =2 # . (4) When we use a mixed prior for φ with a constant mixing probability p, full conditional density of this parameter is given by, # " n Y  2   f (ht |ht −1 , µ, σ ) f (h1 |µ, σh21 )p   t = 2 # f (φ|µ, σ 2 , h, r , b) ∝ " n  Y  2   f (ht |ht −1 , µ, φ, σ ) f (h1 |µ, φ, σh21 )(1 − p)  if φ = 1 if φ < 1 t =2 where σh21 = σ 2 for φ = 1 and it is σ 2 /(1 − φ 2 ) for φ < 1. For the model under the null hypothesis, the full conditional densities are obtained in a similar manner, along with φ = 1. Given an arbitrary choice of starting values for the parameters at the beginning of the algorithm, say, µ(0) , φ (0) and σ 2(0) , Gibbs Sampling algorithm for SV model is given by Initialize µ(0) , φ (0) , and σ 2(0) . For k = 1, 2, . . . , m + M, (k) 1. Draw ht from f (ht |µ(k−1) , φ (k−1) , σ 2(k−1) , h(−kt−1) , r ) for t = 1, . . . , n. 2. Draw µ(k) from f (µ|φ (k−1) , σ 2(k−1) , h(k) , r ). 3. Draw σ 2(k) from f (σ 2 |µ(k) , φ (k−1) , h(k) , r ). 4. Draw φ (k) from f (φ|µ(k) , σ 2(k) , h(k) , r ). Repeating the above sampling steps, we obtain a discrete-time Markov Chain {(µ(k) , σ 2(k) , φ (k) , h ), k = 1, 2, . . .} whose stationary distribution is the joint posterior density of the parameters i.e., the conditional density of (θ , h) given r, under the corresponding model. It follows that the marginal samples θ (k) = (µ(k) , σ 2(k) , φ (k) ) converge in distribution to our target posterior density of (µ, σ 2 , φ) given r. We carry out the Gibbs sampling algorithm as proposed above by means of a software package known as WinBUGS [25]. For our simulation study, we use the unix version of the software called BUGS [8]. We suggest reading [22–24], to get a satisfactory insight to the use of BUGS. One interesting feature of applying our Gibbs sampling scheme via BUGS or WinBUGS for SV models merits a mention. Although the algorithm samples the log-volatilities one at a time (as univariate updates), the software performs this operation outstandingly fast. In fact, to obtain a sample of M = 5000 samples after a burn-in of m = 5000 from a parallel chain of three starting values, it takes only about 65 s. This makes the use of our proposed method much simpler for end-users. The code we have used for our data analysis is available upon request from the first author. (k) Z.I. Kalaylıoğlu, S.K. Ghosh / Statistical Methodology 6 (2009) 189–201 195 4.2. Posterior interval method to test the unit-root hypothesis In frequentist analysis, one way to perform hypothesis testing is to use confidence intervals as acceptance regions for the corresponding testing problems. Similarly, one can employ the posterior interval of the parameter of interest to conduct hypothesis testing in Bayesian analysis. It may be noted that this method is applicable in our context, only if we use those prior densities for φ that includes 1 in their support. For instance, the usual Beta densities (or any other continuous densities with support (0, 1)) as priors for φ are not suitable for this posterior interval method as the posterior probability Pr(φ ≥ 1|r ) = 0 for all such priors. One can obtain a credible interval based on the posterior distribution of φ from the Markov chain that we construct via the Gibbs sampling algorithm. For instance, a 95% equal-tail posterior interval for φ can be arbitrarily approximated by computing the 2.5th and 97.5th percentiles of the chain for φ (k) , k = m + 1, . . . , m + M, for large values of m and M. In Sections 5 and 6 we provide the specific values of m and M that have been used for our simulation studies and data analysis. In order to test the unit-root hypothesis H0 : φ = 1, we may simply use the rule of rejecting the null hypothesis, if φ = 1 is not included in the posterior interval. From our simulation studies, we find that this simple rule actually performs remarkably well in maintaining good frequentist properties, such as high power or equivalently low total (Type I plus Type II) error rates. Besides maintaining good frequentist properties, this rule is very easy to implement by constructing posterior intervals based on MCMC methods. 4.3. Bayes factor to test the unit-root hypothesis Testing the null hypothesis H0 : φ = 1 versus the alternative hypothesis H1 : |φ| < 1, can be thought of as comparing the two competing models. In this sense the ratio of the posterior Pr (H |r ) probabilities of the models, Pr (H0 |r ) , is considered as a suitable test statistic. The Bayes Factor (BF) is 1 defined as the ratio of posterior odds to prior odds. The BF quantifies the amount of information gained from the data in support of one model with respect to that from prior information. In most of the data analyses, prior odds are usually taken to be 1 to indicate prior ignorance. In calculating the Bayes Factor for SV Models, we use the exact distribution of {log u2t }, instead of using an approximation that is obtained by a mixture of several normal densities. This is in sharp contrast to the method proposed by [21] that writes the model in a linear state space form and approximate the exact distribution of log u2t , by a mixture of seven normal distributions. Note that the Bayes Factor is defined as, BF = m0 (r ) m1 (r ) where mi (r ) is the marginal likelihood of the mean corrected returns under the model i, where i = 0, 1 correspond to H0 and H1 , respectively. Denoting the parameter vector of model i by θ i , one can write mi (r ) as mi (r ) = f (r |h)f (h|θ i )f (θ i ) f (h, θ i |r ) which leads to 1 mi (r ) = Eh|r  1 f (r |h) . Using Monte Carlo integration estimate for the denominator in the above expression, we obtain an estimator of the marginal likelihood of the data. That is, 1 m̂i (r ) = 1 M M P l= 1 1 f (r |h(l) ) (5) 196 Z.I. Kalaylıoğlu, S.K. Ghosh / Statistical Methodology 6 (2009) 189–201 Table 1 Proportion of correct decisions based on posterior interval (PI) method and BF method for small sample size (n = 100, 500 MC replications) Prior for φ True φ PI method BF method MU (p = 0.95) 1 0.98 0.95 0.70 0.39 0.60 0.62 0.37 0.32 MU (p = 0.5) 1 0.98 0.95 0.41 0.78 0.94 0.57 0.32 0.31 U (−2, 2) 1 0.98 0.95 0.87 0.19 0.87 0.26 0.89 0.98 U (0, 1) 1 0.98 0.95 0.00 1.00 1.00 0.75 0.24 0.17 can be used to estimate mi (r ). It is well known that the above form of (harmonic mean) Monte Carlo estimation for BF does not satisfy the the CLT for Monte Carlo sampling (see [12]). However, we did not find any problem with convergence. But we found the method to be extremely inefficient computationally, as it requires the evaluation of the high-dimensional densities. After we obtain the BF using above steps, we use the rule of rejecting the null hypothesis if log10 (BF ) < 0 (also suggested by So and Li [21]). It may be noted that other cut off values than 0 can be used (see [12]). 5. Simulation studies We have conducted extensive simulations to explore the frequentist properties of the posterior interval method, and the BF method to test the unit root hypothesis. We present some of the significant findings of our simulation experiments. In order to perform the simulation experiments, we simulate the data from the model (1), by using the same parameter values as in the power study of So and Li [21]. In other words, for each data set throughout our simulation experiments, we fix the true value of the mean, µ = −9 and variance of the shocks to volatility σ 2 = 0.1. Several test values (e.g. φ = 1, 0.98 and 0.95) have been used for φ to test the null hypothesis H0 : φ = 1 vs. Ha : |φ| < 1. We use µ ∼ N (0, 106 ) and N (0, 10) and σ 2 ∼ IG(2 + 10−10 , 0.1) and IG(5, 0.1) as diffuse and informative priors, respectively, for µ and σ 2 . However we did not find any significant difference between the posterior inference based on these priors. So to save space, we report the results based on only the non-informative priors. For φ , we consider a wide class of prior densities, see Tables 2 and 3 for more details. In order to compare the methods, we compute the proportion of correct decisions under different point null alternatives. The proportion of correct decisions is the ratio of number of correct decisions to the total number of Monte Carlo replications. For the test based on the posterior interval, the number of correct decisions is the number of samples for which the value of 1 is covered by the posterior interval of φ when the true value of φ is 1. When the true value of φ is 0.98 or 0.95, the number of correct decisions is the number of posterior intervals that do not include 1. For the test based on the log10 (BF ), the number of correct decision is the number of the samples for which log10 (BF ) ≥ 0 when the true value of φ is 1. On the other hand, it is the number of samples for which log10 (BF ) < 0 when the true value of φ is a point under the alternative hypothesis. In order to determine suitable values of m and M to be used throughout our simulations, we used some initial plots (e.g. trace and autocorrelation plots) of the chains. We also checked the numerical summary values as suggested by Gelman and Rubin [6] and Raftery and Lewis [18] to diagnose the convergence. All these diagnostics were performed using the R version of CODA (see [1] 197 Z.I. Kalaylıoğlu, S.K. Ghosh / Statistical Methodology 6 (2009) 189–201 Table 2 Proportion of correct decisions for moderate to large samples (based 100 MC replications) True φ Prior density n = 500 n = 1000 1 U (0, 1) TN (0.9, 0.052 ) Beta(10, 1) Beta(20, 2) U(0, 1.95) U(−2, 2) MU(p = 0.95) MU with p ∼ Beta(0.5, 0.5) 0.96 0.88 0.86 0.85 0.93 0.93 0.94 0.79 0.90 0.94 0.75 0.84 0.93 0.94 0.98 0.86 0.98 U (0, 1) TN (0.9, 0.052 ) Beta(10, 1) Beta(20, 2) U(0, 1.95) U(−2, 2) MU(p = 0.95) MU with p ∼ Beta(0.5, 0.5) 0.36 0.44 0.60 0.50 0.43 0.47 0.41 0.77 0.64 0.67 0.66 0.80 0.86 0.87 0.77 0.93 0.95 U (0, 1) TN (0.9, 0.052 ) Beta(10, 1) Beta(20, 2) U(0, 1.95) U(−2, 2) MU(p = 0.95) MU with p ∼ Beta(0.5, 0.5) 0.82 0.87 0.89 0.93 0.94 0.94 0.89 0.97 0.98 0.98 0.97 0.98 0.99 0.99 0.98 0.99 Bold numbers are based on proposed PI method, while other numbers are based on So and Li’s BF method. Table 3 Total Error Rates using PI’s for sample sizes n = 500 and 1000 Method Posterior interval (φ ∼ MU (p = 0.95)) Posterior interval (φ ∼ U (01.95)) Posterior interval (φ ∼ U (−22)) BF of So and Li (φ ∼ U (0, 1)) φ = 0.98 n = 500 n = 1000 φ = 0.95 n = 500 n = 1000 0.65 0.25 0.17 0.04 0.64 0.21 0.13 0.08 0.60 0.19 0.13 0.07 0.68 0.46 0.22 0.12 for convergence diagnostics in CODA). Based on these initial plots and diagnostics we decided to use m = 1000 and M = 5000 for all our simulations. In Table 1, we present the proportion of correct decisions made by the posterior interval (PI) method of testing (in column 3), as well as the test based on the log10 (BF ) (in column 4) for testing φ = 1 using our SV model. For this study, we used a small sample of size n = 100, and repeated the procedures 500 times. In this table, we observe that the posterior interval based test is quite sensitive to the choice of priors. For instance, the prior U (−2, 2) performs very well in detecting the presence of unit-root under null compared to other prior and BF. However, it loses power when φ moves below one. It appears that MU priors perform reasonably well in terms of posterior interval method compared to the BF method. It is obvious that using U (0, 1) as a prior for φ makes the posterior interval test incapable of detecting the unit root. When a mixed density is used, the probability of committing a Type I error by the posterior interval test is quite high but not as severe as the case when U(0,1) is used as a prior. In general, the proportion of correct decisions made by the posterior interval test is higher than that of the test based on the BF. One nice feature of the BF test is that it is not as severely affected by the use of U (0, 1) as compared to interval method. 198 Z.I. Kalaylıoğlu, S.K. Ghosh / Statistical Methodology 6 (2009) 189–201 In Table 2, we compare the power of the Bayesian unit root testing criteria of [21] (based on BF) to that of our test that is based on the posterior interval of φ . In this table the first four priors were used by So and Li [21], and the proportion of correct decisions are reproduced from their article that they obtained based on their BF test. Our results have been indicated by bold numbers. We observe that, under H0 , the proportion of correct decisions based on our tests are higher in most cases than that of So and Li’s test based on the BF. This clearly suggests that one can get improved results from posterior interval test by using the prior densities we suggested for φ when applied to SV models. In addition, we observe that under H0 , proportion of correct decisions made by the test criteria of So and Li decreases as the sample size increases in all the cases except for the Truncated Normal (TN) prior. We do not see such a surprising phenomenon with our testing method. As expected, the mixed priors which have a high probability on the point φ = 1, seem to perform better as compared with the ones that have a lower probability on the point φ = 1 in terms of the probability of committing a Type 1 error. However the other priors that do not put high mass at unity a priori also perform reasonably well. Based on our simulation study, we suggest that practitioners should use both types of priors to see the sensitivity on the posterior inference. From Table 2 we also observe that under the alternative, our proposed testing method maintains a reasonably higher proportion of correct decisions as compared with that of [21]’s method. In order to compare the priors for this testing problem, we present the total (Type I plus Type II) error rates in Table 3. It appears that the U (−2, 2) performs reasonably well as compared with other priors in maintaining low total error rates. So in practice, we would recommend this prior as a default choice for testing unit root for SV models. Table 2 demonstrates the superiority of our test over the existing unit root test criteria for SV models. From Table 2, we see that, in general, So and Li’s test yields high power at the cost of high Type I Error. On the other hand, our proposed testing method maintains reasonably low total error rates as observed in Table 3. The PI methods performs much better in terms of detecting the absence of unit root, when the true value of φ is far away from 1 and when sample size increases. So we did not report such findings here for space limitations. For instance in Table 3 we observe that the total error rates drops to about only 19% even when φ = 0.98 as compared to that of BF method yielding a total error rate of 46%. 6. Application to exchange rate data The data set for our application consist of the following four exchange rates: Pound/Dollar, Deutschmark(DM)/Dollar, Yen/Dollar, and Swiss-Frank(SF)/Dollar. The data are closed exchange rates from October 1st, 1981 to June 28th, 1985 on the weekdays. The sample size for four series is n = 946. Let us denote the exchange rate for the i-th series on t-th day by pit . Then, for each series i, the mean corrected return data is calculated as rit = 1 log pit − n 1X 1 log pit t = 1, . . . , n n i=1 where 1 is the difference operator, i.e., 1 log pit = log pit − log pi,t −1 . Here, i = 1, 2, 3, or 4 corresponds to Pound/Dollar, Deutschmark/Dollar, Yen/Dollar, or Swiss-Frank/Dollar, respectively. The time series plots of the returns are displayed in Fig. 1. Visually, it is hard to tell from the plots whether the affect of the random shocks to volatilities are persistent or not. So we use our testing method to see if φ = 1 assuming that the series of returns can be approximately described by a SV model. For each of the four series, we fitted an ARMA (1, 1) model to the data in the form of log rit2 and checked the autocorrelations of the residual from these fits by using PROC ARIMA in SAS. The test result is that residuals are white noise implying that ARMA(1, 1) is an adequate fit to log rit2 for each series, i = 1, . . . , 4 which in turn implies that a SV model is a reasonable fit to the mean corrected exchange rate returns. Based on our finding from simulation studies, we use only the posterior interval method to test the null hypotheses for these four series on exchange rates. We decided to use same set of flat prior distributions for µ and σ 2 that were used in our simulations. For the parameter φ we use MU (0.95), U (0, 1.95) and U (−2, 2) priors. Z.I. Kalaylıoğlu, S.K. Ghosh / Statistical Methodology 6 (2009) 189–201 199 Fig. 1. Plot of the mean corrected returns rit versus time t. To implement the Gibbs sampling, we used three dispersed initial values for all the parameters. We execute the Gibbs Sampling (in WinBUGS) to obtain 10,000 samples from each of the three parallel chains resulting into 30,000 samples. Using plots and numerical summary measures for convergence diagnostics (available in CODA) we decided to throw out first 5000 samples from each of the three parallel chains. Thus we used 15,000 final samples to obtain summary values of the posterior distribution of the parameters. The results based on different priors for φ are presented in Table 4. When we use a mixed prior with p = 0.95, in addition to the posterior inference given in the table, the posterior median of φ turns out to be one for each series. It appears that, for each series, there is more than 50% probability that φ is at least 1 based on the data. The posterior intervals cover the value of one implying that null hypothesis can not be rejected for any of the four series. However, when we use φ ∼ U (0, 1.95) and φ ∼ U (−2, 2), the test based on the posterior interval does not find sufficient evidence in favor of the unit root for all series, but for the Yen/Dollar series. This makes it harder to conclude, but for the Yen/Dollar series. However given the superior performance of the test based on U (−2, 2) prior from our simulation study, for this data we would conclude that only for the Yen/Dollar series a unit-root hypothesis can be accepted and for the other three series there is not enough evidence to believe the presence of a unit root. 7. Conclusions In this article, we have developed Bayesian unit root tests for Stochastic Volatility (SV) Models, where the log-squared volatilities are assumed to have a first order autoregressive model. We have introduced a mixed prior density that puts a nonzero mass on the null hypothesis value of the autoregressive parameter. We also suggested to use continuous densities that include the nonstationary values of φ as prior densities. By using such densities on φ , one can use the posterior 200 Z.I. Kalaylıoğlu, S.K. Ghosh / Statistical Methodology 6 (2009) 189–201 Table 4 Posterior Inference for φ for three different prior distributions Exchange rate Estimated posterior mean Standard error of the estimate Estimated 95% posterior interval 0.0013 0.0018 0.0016 0.0033 [0.9528, 1.0] [0.9363, 1.0] [0.9437, 1.0] [0.8925, 1.0] 0.0007 0.0013 0.0007 0.0017 [0.9462, 0.9971] [0.9167, 0.9978] [0.9557, 1.0000] [0.8739, 0.9869] 0.0006 0.0009 0.0007 0.0014 [0.9415, 0.9972] [0.9206, 0.9978] [0.9430, 1.0000] [0.8670, 0.9830] MU (p = 0.95) Pound/$ DM/$ Yen/$ SF/$ 0.9920 0.9881 0.9898 0.9791 Pound/$ DM/$ Yen/$ SF/$ 0.9769 0.9642 0.9868 0.9405 Pound/$ DM/$ Yen/$ SF/$ 0.9761 0.9679 0.9796 0.9343 U (0, 1.95) U (−2, 2) interval of φ to test for a unit root in a SV model. We showed that, applying a test based on the Bayes Factor in SV models is not practical as computing the Bayes Factor is a computationally intensive process due to the high-dimension of the missing observations (in this case, volatilities). On the other hand, calculating the posterior interval is very simple, and can be performed easily in WinBUGS. Using a Monte Carlo study, we have shown that the posterior interval unit root test in SV Models is reliable. In addition to the Bayesian methods presented in this article, we also compared the total error rates of posterior interval tests to that of the commonly used frequentist tests (not shown here, however see [11]). The results from several frequentist procedures will be reported in a separated article elsewhere. From our studies, we found that our Bayesian methods based on posterior interval method perform relatively better in terms of having low total error rates as compared with the frequentist counterparts. In conclusion we would recommend users to use the Bayesian unit root test, based on the posterior interval of φ as it is a simple, practical, and reliable solution to the problem of seeking evidence for a unit root in SV models. Acknowledgement We are grateful for the comments and suggestions given by a reviewer, which has improved the presentation of an earlier version of this article. References [1] N.G. Best, M.K. Cowles, S.K. Vines, CODA Convergence diagnosis and output analysis software for Gibbs Sampler output, 1995. [2] T. Bollerslev, R.F. Engle, Common persistence in conditional variance, Econometrica 61 (1993) 167–186. [3] R.Y. Chou, Volatility persistence and stock valuations: Some empirical evidence using GARCH, Journal of Applied Econometrics 3 (1988) 279–294. [4] R.F. Engle, Autoregressive conditional heteroscedasticity with estimates of the variance of United Kingdom inflations, Econometrica 50 (1982) 987–1007. [5] A. Gelfand, A.F.M. Smith, Sampling-based approaches to calculating marginal densities, Journal of the American Statistical Association 85 (1990) 398–409. [6] S. Gelman, D. Rubin, Inference from iterative simulation using multiple sequences (with discussion), Statistical Science 7 (1992) 457–511. [7] W.R. Gilks, S. Richardson, D.J. Spiegelhalter, Markov Chain Monte Carlo in Practice, Chapman and Hall, 1996. [8] W.R. Gilks, A. Thomas, D.J. Spiegelhalter, A language and a program for complex bayesian modeling, The Statistician 43 (1994) 169–177. [9] A. Harvey, E. Ruiz, N. Shephard, Multivariate stochastic models, Review of Economic Studies 61 (1994) 247–264. [10] E. Jacquier, N.G. Polson, P.E. Rossi, Bayesian analysis of stochastic volatility models, Journal of Business and Economic Statistics 12 (1994) 371–388. Z.I. Kalaylıoğlu, S.K. Ghosh / Statistical Methodology 6 (2009) 189–201 201 [11] Z.I. Kalaylioğlu, Frequentist and Bayesian unit root tests in stochastic volatility models, Unpublished Ph.D. Thesis, North Carolina State University, Department of Statistics, Raleigh, NC, 2002. [12] R. Kass, A. Raftery, Bayes factors, Journal of the American Statistical Association 90 (1995) 773–795. [13] S. Kim, N. Shephard, S. Chib, Stochastic volatility: Likelihood inference and comparison with ARCH models, Review of Economic Studies 65 (1998) 361–393. [14] R.J. Mahieu, P.C. Schotman, An empirical application of stochastic volatility models, Journal of Applied Econometrics 13 (1998) 333–360. [15] J.M. Poterba, L.H. Summers, The persistence of volatility and stock market fluctuations, The American Economic Review 76 (1986) 1142–1151. [16] R.S. Pyndick, Risk, inflation, and the stock market, The American Economic Review 74 (1984) 335–351. [17] R.S. Pyndick, Risk aversion and determinants of stock market behavior, The Review of Economics and Statistics LXX (1988) 183–190. [18] A.E. Raftery, S. Lewis, How many iterations in the Gibbs sampler? in: J.M. Bernardo, et al. (Eds.), Bayesian Statistics 4, Oxford University Press, 1992. [19] S.E. Said, D.A. Dickey, Testing for unit roots in autoregressive-moving average models of unknown order, Biometrika 71 (1984) 599–607. [20] G. Sandmann, S.J. Koopman, Estimation of stochastic volatility models via monte carlo maximum likelihood, Journal of Econometrics 87 (1998) 271–301. [21] M.K.P. So, W.K. Li, Bayesian unit-root testing in stochastic volatility models, Journal of Business and Economic Statistics 17 (1999) 491–496. [22] D.J. Spiegelhalter, A. Thomas, N. Best, W. Gilks, BUGS: Bayesian Inference Using Gibbs Sampling, Version 0.5, (version ii), 1996. [23] D.J. Spiegelhalter, A. Thomas, N. Best, W. Gilks, BUGS Examples Volume 1, Version 0.5, (version ii), 1996. [24] D.J. Spiegelhalter, A. Thomas, N. Best, W. Gilks, BUGS Examples Volume 2, Version 0.5, (version ii), 1996. [25] D.J. Spiegelhalter, A. Thomas, N. Best, WinBUGS Version 1.2. User Manual, 1999. [26] L. Tierney, Markov chains for exploring posterior distributions (with discussion), Annals of Statistics 22 (1994) 1701–1762. View publication stats