Fiscal Policy Rules in Practice



2.1 Bayesian Analysis of Markov-Switching Models

The Bayesian analysis of Markov-switching models goes back to McCulloch and Tsay (1994).
They show that Bayesian estimation of Markov-switching models is kept relatively simple when
using the Gibbs sampler, as it solves the problem of drawing samples from a multivariate density
function by drawing successive samples from the corresponding univariate density functions. The
exposition given in the following is based on Harris (1999) and Krolzig (1997).

We consider the following simple univariate model, where the parameters can take on k different
states S,

yt = μ(St)+ XtB(St)+ εt(St),                              (2.2)

where Xt is the vector of explanatory variables and εt(St) a normally distributed i.i.d. error
term with mean zero and regime-dependent covariance matrix
Ω(St). μ(St) denotes a constant
and
B(St) is the vector of coefficients in state St. Furthermore, we define the transition proba-
bilities for a switch from regime
i to regime j as pij = p(St = j|St-1 = i). We summarize these
probabilities in the transition matrix
P with size (k × k).

Let λ denote the set of all unknown parameters, i.e.

λ = [μ(1),..., μ(k), B(1),..., B(k), Ω(1),..., Ω(k), P].

In partitioned notation this boils down to λ = [Θ, P]. Inference on λ depends on the posterior
distribution

p(λ Y) α π(λ)ρ(YIλ),

(2.3)


where Y' = (y1,... ,yT) is the vector of observations and π(λ) the prior for the parameter
vector. As we are in a Markov-regime switching environment, we have additional unknown
parameters given by the unobservable states. Therefore, the posterior density (2.3) is obtained
by the integration of the joint probability distribution with respect to the state vector S, i.e.

p(λIY) =


p(λ, SIY)dS.


(2.4)


The problem arising from (2.4) is that the posterior distribution of λ depends on an unknown
multivariate distribution p(λ, S
IY). The Gibbs sampler offers a solution to this problem, as it
allows to draw successive samples from univariate distributions for λ and S, namely p(S
IY, λ)
and p(λ
IY, S), instead of the multivariate distribution p(λ, SIY). The Gibbs sampler constructs
a Markov chain on (λ, S) such that the limiting distribution of the chain is the joint distribution
of p(λ, S
IY). There are two types of Gibbs sampler, single-move and multi-move, which differ in
the way the states S are generated. We apply multi-move sampling as it - according to Liu et
al. (1994) - will lead to a faster convergence than single-move sampling.

The idea of multi-move Gibbs sampling is to draw all states in S at once conditional on the
observations. The starting point is to make use of the structure of the underlying Markov chain,
i.e.



More intriguing information

1. Consciousness, cognition, and the hierarchy of context: extending the global neuronal workspace model
2. Luce Irigaray and divine matter
3. Developmental changes in the theta response system: a single sweep analysis
4. The Tangible Contribution of R&D Spending Foreign-Owned Plants to a Host Region: a Plant Level Study of the Irish Manufacturing Sector (1980-1996)
5. Sectoral Energy- and Labour-Productivity Convergence
6. Has Competition in the Japanese Banking Sector Improved?
7. The name is absent
8. SOME ISSUES IN LAND TENURE, OWNERSHIP AND CONTROL IN DISPERSED VS. CONCENTRATED AGRICULTURE
9. DEVELOPING COLLABORATION IN RURAL POLICY: LESSONS FROM A STATE RURAL DEVELOPMENT COUNCIL
10. NVESTIGATING LEXICAL ACQUISITION PATTERNS: CONTEXT AND COGNITION