Model exercises
Download
Report
Transcript Model exercises
DSGE Models and Optimal
Monetary Policy
Andrew P. Blake
A framework of analysis
• Typified by Woodford’s Interest and Prices
– Sometimes called DSGE models
– Also known as NNS models
• Strongly micro-founded models
• Prominent role for monetary policy
• Optimising agents and policymakers
What do we assume?
• Model is stochastic, linear, time invariant
• Objective function can be approximated
very well by a quadratic
• That the solutions are certainty equivalent
– Not always clear that they are
• Agents (when they form them) have rational
expectations or fixed coefficient
extrapolative expectations
Linear stochastic model
• We consider a model in state space form:
s t 1 As t Bu t C t 1
• u is a vector of control instruments, s a
vector of endogenous variables, ε is a shock
vector
• The model coefficients are in A, B and C
Quadratic objective function
• Assume the following objective function:
V 0 min
1
ut
s Qs
2
t
t
t
u t Ru t
t0
• Q and R are positive (semi-) definite
symmetric matrices of weights
• 0 < ρ ≤ 1 is the discount factor
• We take the initial time to be 0
How do we solve
for the optimal policy?
• We have two options:
– Dynamic programming
– Pontryagin’s minimum principle
• Both are equivalent with non-anticipatory
behaviour
• Very different with rational expectations
• We will require both to analyse optimal
policy
Dynamic programming
• Approach due to Bellman (1957)
• Formulated the value function:
Vt
1
2
s t Ss t min
1
ut
2
s tQs t u t Ru t s t1 Ss t 1
• Recognised that it must have the structure:
V t min
ut
s tQs t u t Ru t ( s t A u t B ) S ( As t Bu t )
Optimal policy rule
• First order condition (FOC) for u:
Vt
ut
0 Ru t B S ( As t Bu t ) 0
• Use to solve for policy rule:
1
u t ( R B SB ) B SAs t
Fs t
The Riccati equation
• Leaves us with an unknown in S
• Collect terms from the value function:
s t Ss t s tQs t s t F RFs
• Drop z:
t
s t ( A F B ) S ( A BF ) s t
S Q F RF ( A F B ) S ( A BF )
Riccati equation (cont.)
• If we substitute in for F we can obtain:
2
1
S Q A SA A SB ( R B SB ) B SA
• Complicated matrix quadratic in S
• Solved ‘backwards’ by iteration, perhaps
by:
S j Q A S j 1 A A S j 1 B ( R B S j 1 B )
2
1
B S j 1 A
Properties of the solution
•
•
•
•
‘Principle of optimality’
The optimal policy depends on the unknown S
S must satisfy the Riccati equation
Once you solve for S you can define the policy rule
and evaluate the welfare loss
• S does not depend on s or u only on the model and
the objective function
• The initial values do not affect the optimal control
Lagrange multipliers
• Due to Pontryagin (1957)
• Formulated a system using constraints as:
Hk
k
s k Qs k
u k Ru k k 1 ( As k Bu k s k 1 )
• λ is a vector of Lagrange multipliers:
• The constrained objective function is:
Vt
H
k t
k
FOCs
• Differentiate with respect to the three sets of
variables:
H t
u t
H t
st
H t
t
0 Ru t B t 1 0
0 Qs t A t 1 t 0
0 As t Bu t s t 1 0
Hamiltonian system
• Use the FOCs to yield the Hamiltonian system:
I
0
1
B R B s t 1 A
A t 1 Q
0 st
I t
• This system is saddlepath stable
• Need to eliminate the co-states to determine the
solution
• NB: Now in the form of a (singular) rational
expectations model (discussed later)
Solutions are equivalent
• Assume that the solution to the saddlepath
problem is t Ss t
• Substitute into the FOCs to give:
H t
u t
H t
st
0 Ru t B Ss t 1 0
0 Qs t A Ss t 1 Ss t 0
Equivalence (cont.)
• We can combine these with the model and
eliminate s to give:
2
1
S Q A SA A SB ( R B SB ) B SA
• Same solution for S that we had before
• Pontryagin and Bellman give the same answer
• Norman (1974, IER) showed them to be
stochastically equivalent
• Kalman (1961) developed certainty equivalence
What happens with RE?
• Modify the model to:
z t 1 A11 A12 z t B1
e
ut
x t 1 A 21 A 22 x t B 2
• Now we have z as predetermined variables and x
as jump variables
• Model has a saddlepath structure on its own
• Solved using Blanchard-Kahn etc.
Bellman’s dedication
• At the beginning of Bellman’s book
Dynamic Programming he dedicates it thus:
To Betty-Jo
Whose decision processes defy analysis
Control with RE
• How do rational expectations affect the optimal
policy?
– Somewhat unbelievably - no change
– Best policy characterised by the same algebra
• However, we need to be careful about the jump
variables, and Betty-Jo
• We now obtain pre-determined values for the costates λ
• Why?
Pre-determined co-states
• Look at the value function V t 12 s t Ss t
• Remember the reaction function is:
tz S 11
x
t S 21
S 12 z t
Ss t
S 22 x t
• So the cost can be written as V t 12 s t t
• We can minimise the cost by choosing some
co-states and letting x jump
Pre-determined co-states (cont.)
• At time 0 this is minimised by:
V0
1
2
z 0
0z 1
x 0 x z 0
0 2
0z
x 0
0
• We can rearrange the reaction function to:
tz N 11
x t N 21
• Where
N 12 z t
x
N 22 t
1
1
N 11 S 11 S 12 S 22 S 21 , N 22 S 22
etc
Pre-determined co-states (cont.)
• Alternatively the value function can be
written in terms of the x and the z’s as:
zt I
x t N 21
0 zt
zt
x T x
N 22 t
t
• The loss is:
V0
1
2
z 0
T ST
x
0
z0
x
0
Cost-to-go
•
•
•
•
At time 0, z0 is predetermined
x0 is not, and can be any value
In fact is a function of z0 (and implicitly u)
We can choose the value of λx at time 0 to
minimise cost
• We choose it to be 0
• This minimises the cost-to-go in period 0
Time inconsistency
•
•
•
•
•
•
This is true at time 0
Time passes, maybe just one period
Time 1 ‘becomes time 0’
Same optimality conditions apply
We should reset the co-states to 0
The optimal policy is time inconsistent
Different to non-RE
• We established before that the non-RE solution did
not depend on the initial conditions (or any z)
• Now it directly does
• Can we use the same solution methods?
– DP or LM?
– Yes, as long as we ‘re-assign’ the co-states
• However, we are implicitly using the LM solution
as it is ‘open-loop’ – the policy depends directly
on the initial conditions
Where does this fit in?
• Originally established in 1980s
– Clearest statement Currie and Levine (1993)
– Re-discovered in recent US literature
– Ljungqvist and Sargent Recursive
Macroeconomic Theory (2000, and new
edition)
• Compare with Stokey and Lucas
How do we deal with time
inconsistency?
• Why not use the ‘principle of optimality’
• Start at the end and work back
• How do we incorporate this into the RE
control problem?
– Assume expectations about the future are
‘fixed’ in some way
– Optimise subject to these expectations
A rule for future expectations
• Assume that:
x
e
t 1
N t 1 z t 1
• If we substitute this into the model we get:
1
x t ( A 22 N t 1 A12 ) ( N t 1 A11 A12 ) z t
1
( A 22 N t 1 A12 ) ( N t 1 B1 B 2 ) u t
J t zt K tut
A rule for future expectations
• The ‘pre-determined’ model is:
z t 1 A11 z t A12 x t B1u t
• Using the reaction function for x we get:
z t 1 ( A11 A12 J t ) z t ( B1 B 2 K t ) u t
Aˆ t z t Bˆ u t
Dynamic programming solution
• To calculate the best policy we need to
make assumptions about leadership
• What is the effect on x of changes in u?
• If we assume no leadership it is zero
• Otherwise it is K, need to use:
Vt
u t
Vt xt
xt u t
0
Vt
u t
Vt
xt
Kt
Dynamic programming (cont.)
• FOC for u for leadership:
1 ˆ
ˆ
ˆ
ˆ
u t ( R B t S t 1 B t ) ( B tS t 1 Aˆ t K t ( Q 22 J t Q 21 )) z t
Fˆt z t
where: Rˆ t R K tQ 22 K t
• This policy must be time consistent
• Only uses intra-period leadership
Dynamic programming (cont.)
• This is known in the dynamic game
literature as feedback Stackelberg
• Also need to solve for S
– Substitute in using relations above
• Can also assume that x unaffected by u
– Feedback Nash equilibrium
• Developed by Oudiz and Sachs (1985)
Dynamic programming (cont.)
• Key assumption that we condition on a rule
for expectations
• Could condition on a time path (LM)
• Time consistent by construction
– Principle of optimality
• Many other policies have similar properties
• Stochastic properties now matter
Time consistency
• Not the only time consistent solutions
• Could use Lagrange multipliers
• DP is not only time consistent it is subgame
perfect
• Much stronger requirement
– See Blake (2004) for discussion
What’s new with DSGE models?
• Woodford and others have derived welfare loss
functions that are quadratic and depend only on
the variances of inflation and output
• These are approximations to the true social utility
functions
• Can apply LQ control as above to these models
• Parameters of the model appear in the loss
function and vice versa (e.g. discount factor)
DGSE models in WinSolve
• Can set up micro-founded models
• Can set up micro-founded loss functions
• Can explore optimal monetary policy
– Time inconsistent
– Time consistent
– Taylor-type approximations
• Let’s do it!