Презентация «Forecasting with bayesian techniques MP»

Смотреть слайды в полном размере
Презентация «Forecasting with bayesian techniques MP»

Вы можете ознакомиться с презентацией онлайн, просмотреть текст и слайды к ней, а также, в случае, если она вам подходит - скачать файл для редактирования или печати. Документ содержит 72 слайда и доступен в формате ppt. Размер файла: 2.38 MB

Просмотреть и скачать

Pic.1
«Forecasting with bayesian techniques MP», слайд 1
Pic.2
«Forecasting with bayesian techniques MP», слайд 2
Pic.3
Introduction: Two Perspectives in Econometrics Let θ be a vector of parameters to be estimated using
Introduction: Two Perspectives in Econometrics Let θ be a vector of parameters to be estimated using data For example, if yt~ i. i. d. N(μ,σ2), then θ=[μ,σ2] are to be estimated from a sample {yt} …
Pic.4
Outline Why a Bayesian Approach to VARs? Brief Introduction to Bayesian Econometrics Analytical Exam
Outline Why a Bayesian Approach to VARs? Brief Introduction to Bayesian Econometrics Analytical Examples Estimating a distribution mean Linear Regression Analytical priors and posteriors for BVARs …
Pic.5
Why a Bayesian Approach to VAR? Dimensionality problem with VARs: y contains n variables, p lags in
Why a Bayesian Approach to VAR? Dimensionality problem with VARs: y contains n variables, p lags in the VAR The number of parameters in c and A is n(1+np), and the number of parameters in Σ is …
Pic.6
Why a Bayesian Approach to VAR? (2) Usually, only a fraction of estimated coefficients are statistic
Why a Bayesian Approach to VAR? (2) Usually, only a fraction of estimated coefficients are statistically significant parsimonious modeling should be favored What could we do? Estimate a VAR with …
Pic.7
Combining information: prior and posterior Bayesian coefficient estimates combine information in the
Combining information: prior and posterior Bayesian coefficient estimates combine information in the prior with evidence from the data Bayesian estimation captures changes in beliefs about model …
Pic.8
Shrinkage There are many approaches to reducing over-parameterization in VARs A common idea is shrin
Shrinkage There are many approaches to reducing over-parameterization in VARs A common idea is shrinkage Incorporating prior information is a way of introducing shrinkage The prior information can be …
Pic.9
Forecasting Performance of BVAR vs. alternatives Source: Litterman, 1986
Forecasting Performance of BVAR vs. alternatives Source: Litterman, 1986
Pic.10
Introduction to Bayesian Econometrics: Objects of Interest Objects of interest: Prior distribution:
Introduction to Bayesian Econometrics: Objects of Interest Objects of interest: Prior distribution: Likelihood function: - likelihood of data at a given value of θ Joint distribution (of unknown …
Pic.11
Bayesian Econometrics: Objects of Interest (2) The marginal likelihood… …is independent of the param
Bayesian Econometrics: Objects of Interest (2) The marginal likelihood… …is independent of the parameters of the model Therefore, we can write the posterior as proportional to prior and data:
Pic.12
Bayesian Econometrics: maximizing criterion For practical purposes, it is useful to focus on the cri
Bayesian Econometrics: maximizing criterion For practical purposes, it is useful to focus on the criterion: Traditionally, priors that let us obtain analytical expressions for the posterior would be …
Pic.13
Bayesian Econometrics : maximizing criterion (2) Maximizing C() gives the Bayes mode. In some cases
Bayesian Econometrics : maximizing criterion (2) Maximizing C() gives the Bayes mode. In some cases (i. e. Normal distributions) this is also the mean and the median The criterion can be generalized …
Pic.14
Analytical Examples Let’s work on some analytical examples: Sample mean Linear regression model
Analytical Examples Let’s work on some analytical examples: Sample mean Linear regression model
Pic.15
Estimating a Sample Mean Let yt~ i. i. d. N(μ,σ2), then the data density function is: where y={y1,…y
Estimating a Sample Mean Let yt~ i. i. d. N(μ,σ2), then the data density function is: where y={y1,…yT} For now: assume variance σ2 is known (certain) Assume the prior distribution of mean μ is …
Pic.16
Estimating a Sample Mean The posterior of μ: …has the following analytical form with So, we “mix” pr
Estimating a Sample Mean The posterior of μ: …has the following analytical form with So, we “mix” prior m and the sample average (data) Note: The posterior distribution of μ is also normal: μ~ …
Pic.17
Estimating a Sample Mean: Example Assume the true distribution is Normal yt~N(3,1) So, μ=3 is known
Estimating a Sample Mean: Example Assume the true distribution is Normal yt~N(3,1) So, μ=3 is known to… God A researcher (one of us) does not know μ for him/her it is a normally distributed random …
Pic.18
Posterior with prior N(1,1) Compute the posterior distribution as sample size increases
Posterior with prior N(1,1) Compute the posterior distribution as sample size increases
Pic.19
Posterior with Prior N(1,1/50) Then, we look at more informative (tight) prior and set ν =50 (higher
Posterior with Prior N(1,1/50) Then, we look at more informative (tight) prior and set ν =50 (higher precision)
Pic.20
Examples: Regression Model I Linear Regression model: where ut~ i. i. d. N(0,σ2) Assume: β is random
Examples: Regression Model I Linear Regression model: where ut~ i. i. d. N(0,σ2) Assume: β is random and unknown but σ2 is fixed and known Convenient matrix representation where The density function …
Pic.21
Examples: Regression Model I (2) Assume that the prior mean of β has multivariate Normal distributio
Examples: Regression Model I (2) Assume that the prior mean of β has multivariate Normal distribution N(m,σ2M): where the key parameters of the prior distribution are m and M Bayesian rule states: i. …
Pic.22
Examples: Regression Model I (3) We mix information – densities of data and prior – to get posterior
Examples: Regression Model I (3) We mix information – densities of data and prior – to get posterior distribution! Result: the density function of β is… … which means that the posterior distribution …
Pic.23
Since we do not like black boxes… there are 2 ways to get m* and M* (2 parameters to characterize po
Since we do not like black boxes… there are 2 ways to get m* and M* (2 parameters to characterize posterior) Since we do not like black boxes… there are 2 ways to get m* and M* (2 parameters to …
Pic.24
Define a “new” regression model Define a “new” regression model We simply stack our “ingredients” to
Define a “new” regression model Define a “new” regression model We simply stack our “ingredients” together to mix the information (prior and data) so that now β takes into account both! The GLS …
Pic.25
Examples: Regression Model II So far the life was easy(-ier), in the linear regression model β was r
Examples: Regression Model II So far the life was easy(-ier), in the linear regression model β was random and unknown, but σ2 was fixed and known What if σ2 is random and unknown?. . Bayesian rule …
Pic.26
Examples: Regression Model II () To manipulate the product …we assume the following distributions: N
Examples: Regression Model II () To manipulate the product …we assume the following distributions: Normal for data Normal for the prior for β (conditional on σ2): β|σ2 ̴ N(m, σ2M) and Inverse-Gamma …
Pic.27
Examples: Regression Model II (3) By manipulating the product (see more details in the appendix B) …
Examples: Regression Model II (3) By manipulating the product (see more details in the appendix B) …we get the following result with mean and variance of the posterior for β|σ2 ̴ N(m*, σ2M*) And …
Pic.28
Priors: summary In the above examples we dealt with 2 types of prior distributions of our parameters
Priors: summary In the above examples we dealt with 2 types of prior distributions of our parameters: Case 1 prior assumes β is unknown and normally distributed (Gaussian) σ2 is a known parameter the …
Pic.29
Bayesian VARs Linear Regression examples will help us to deal with our main object – Bayesian VARs A
Bayesian VARs Linear Regression examples will help us to deal with our main object – Bayesian VARs A VAR is typically written as where yt contains n variables, the VAR includes p lags, and the data …
Pic.30
VAR in a matrix form: example Consider, as an example, a VAR for n variables and p=2 Stack the varia
VAR in a matrix form: example Consider, as an example, a VAR for n variables and p=2 Stack the variables and coefficients Then, the VAR Let and rewrite where is a Kroneker product
Pic.31
How to Estimate a BVAR: Case 1 Prior Consider Case 1 prior for a VAR: coefficients in A are unknown
How to Estimate a BVAR: Case 1 Prior Consider Case 1 prior for a VAR: coefficients in A are unknown with multivariate Normal prior distribution: and known Σe “Old trick” to get the posterior: use GLS …
Pic.32
How to Estimate a BVAR: Case 2 (conjugate) Priors Before we see the case of an unknown Σe need to in
How to Estimate a BVAR: Case 2 (conjugate) Priors Before we see the case of an unknown Σe need to introduce a multivariate distribution to characterize the unknown random error covariance matrix Σe …
Pic.33
How to Estimate a BVAR: Conjugate Priors Assume Conjugate priors: The VAR parameters A and Σe are bo
How to Estimate a BVAR: Conjugate Priors Assume Conjugate priors: The VAR parameters A and Σe are both unknown prior for A is multivariate Normal: and for Σe is Inverse Wishart: Follow the analogy …
Pic.34
BVARs: Minnesota Prior Implementation The Minnesota prior – a particular case of the “Case 1 prior”
BVARs: Minnesota Prior Implementation The Minnesota prior – a particular case of the “Case 1 prior” (unknown model coefficients, but known error variance): Assume random walk is a reasonable model …
Pic.35
BVARs: Minnesota Prior Implementation The Minnesota prior The prior variance for the coefficient of
BVARs: Minnesota Prior Implementation The Minnesota prior The prior variance for the coefficient of lag k in equation i for variable j is: … and depends only on three hyperparameters: the tightness …
Pic.36
BVARs: Minnesota Prior Implementation The Minnesota prior Interpretation: the prior on the first own
BVARs: Minnesota Prior Implementation The Minnesota prior Interpretation: the prior on the first own lag is the prior on the own lag k is the prior std. dev. declines at a rate k, i. e. coefficients …
Pic.37
Remarks: Remarks: The overall tightness of the prior is governed by γ smaller γ  model for yit shri
Remarks: Remarks: The overall tightness of the prior is governed by γ smaller γ  model for yit shrinks towards random walk The effect of other lagged variables is controlled by w smaller  estimates …
Pic.38
BVARs: Prior Selection Minnesota and conjugate priors are useful (e. g. , to obtain closed-form solu
BVARs: Prior Selection Minnesota and conjugate priors are useful (e. g. , to obtain closed-form solutions), but can be too restrictive: Independence across equations Symmetry in the prior can …
Pic.39
Del Negro and Schorfheide (2004): DSGE-VAR Approach Del Negro and Schorfheide (2004) We want to esti
Del Negro and Schorfheide (2004): DSGE-VAR Approach Del Negro and Schorfheide (2004) We want to estimate a BVAR model We also have a DSGE model for the same variables It can be solved and linearized: …
Pic.40
Del Negro and Schorfheide (2004) We estimate the following BVAR: The solution for the DSGE model has
Del Negro and Schorfheide (2004) We estimate the following BVAR: The solution for the DSGE model has a reduced-form VAR representation where θ are deep structural parameters Idea: Combine artificial …
Pic.41
Del Negro and Schorfheide (2004) Parameter λ is a “weight” of “artificial” (prior) data from DSGE λ=
Del Negro and Schorfheide (2004) Parameter λ is a “weight” of “artificial” (prior) data from DSGE λ=0 delivers OLS-estimated VAR: i. e. DSGE not important Large λ shrinks coefficients towards the …
Pic.42
Likelihood of the VAR of a DSGE Model Recall the likelihood function for an unconstrained VAR Simila
Likelihood of the VAR of a DSGE Model Recall the likelihood function for an unconstrained VAR Similarly, the (Quasi-) likelihood for the “artificial” data: which is a prior density for the BVAR …
Pic.43
Likelihood of the VAR of a DSGE Model
Likelihood of the VAR of a DSGE Model
Pic.44
DSGE-VAR prior
DSGE-VAR prior
Pic.45
DSGE-VAR posterior
DSGE-VAR posterior
Pic.46
Results
Results
Pic.47
Results
Results
Pic.48
Kadiyala and Karlsson (1997) Small Model: a bivariate VAR with unemployment and industrial productio
Kadiyala and Karlsson (1997) Small Model: a bivariate VAR with unemployment and industrial production Sample period: 1964:1 to 1990:4. Estimate the model through 1978:4 Criterion to chose …
Pic.49
Kadiyala and Karlsson (1997) Compare different priors based on the VAR forecasting performance (RMSE
Kadiyala and Karlsson (1997) Compare different priors based on the VAR forecasting performance (RMSE) Standard VAR(p)… … can be rewritten (see slide 29): … and where
Pic.50
Prior distributions in K&K K&K use a number of competing prior distributions… Minnesota, Nor
Prior distributions in K&K K&K use a number of competing prior distributions… Minnesota, Normal-Wishart, Normal-Diffuse, Extended Natural Conjugate (see appendix E) … for and Parameters of …
Pic.51
Prior distributions in K&K
Prior distributions in K&K
Pic.52
Forecast Comparison in K&K: Small Model, unemployment
Forecast Comparison in K&K: Small Model, unemployment
Pic.53
Forecast Comparison in K&K: Large Model
Forecast Comparison in K&K: Large Model
Pic.54
Giannone, Lenza and Primiceri (2011) Use three VARs to compare forecasting performance Small VAR: GD
Giannone, Lenza and Primiceri (2011) Use three VARs to compare forecasting performance Small VAR: GDP, GDP deflator, Federal Funds rate for the U. S Medium VAR: includes small VAR plus consumption, …
Pic.55
Giannone, Lenza and Primiceri (2011) The marginal likelihood is obtained by integrating out the para
Giannone, Lenza and Primiceri (2011) The marginal likelihood is obtained by integrating out the parameters of the model: But the prior distribution of  is itself a function of the hyperparameters of …
Pic.56
Giannone, Lenza and Primiceri (2011) We interpret the model as a hierarchical model by replacing pγ(
Giannone, Lenza and Primiceri (2011) We interpret the model as a hierarchical model by replacing pγ(θ)=p(θ|γ) and evaluate the marginal likelihood: The hyperparameters γ are uncertain Informativeness …
Pic.57
Giannone, Lenza and Primiceri (2011)
Giannone, Lenza and Primiceri (2011)
Pic.58
In all cases BVARs demonstrate better forecasting performance vis-à-vis the unrestricted VARs In all
In all cases BVARs demonstrate better forecasting performance vis-à-vis the unrestricted VARs In all cases BVARs demonstrate better forecasting performance vis-à-vis the unrestricted VARs BVARs are …
Pic.59
Conclusions BVARs is a useful tool to improve forecasts This is not a “black box” posterior distribu
Conclusions BVARs is a useful tool to improve forecasts This is not a “black box” posterior distribution parameters are typically functions of prior parameters and data Choice of priors can go: from …
Pic.60
Thank You! Thank You!
Thank You! Thank You!
Pic.61
Appendix A: Remarks about the marginal likelihood Remarks about the marginal likelihood: If we have
Appendix A: Remarks about the marginal likelihood Remarks about the marginal likelihood: If we have M1,…. MN competing models, the marginal likelihood of model Mj, f({yt}|Mj) can be seen as: The …
Pic.62
Appendix A: Remarks about the marginal likelihood Remarks about the marginal likelihood: Predict the
Appendix A: Remarks about the marginal likelihood Remarks about the marginal likelihood: Predict the first observation using the prior: Record the first observable and its probability, f(y1o). Update …
Pic.63
Appendix B: Linear Regression with conjugate priors To calculate the posterior distribution for para
Appendix B: Linear Regression with conjugate priors To calculate the posterior distribution for parameters …we assume the following for distributions: Normal for data Normal for the prior for β …
Pic.64
«Forecasting with bayesian techniques MP», слайд 64
Pic.65
«Forecasting with bayesian techniques MP», слайд 65
Pic.66
Appendix C: How to Estimate a BVAR, Case 1 prior Use GLS estimator for the regression Continue (next
Appendix C: How to Estimate a BVAR, Case 1 prior Use GLS estimator for the regression Continue (next slide)
Pic.67
Appendix C: How to Estimate a BVAR, Case 1 Prior Continue So, the moments for the posterior distribu
Appendix C: How to Estimate a BVAR, Case 1 Prior Continue So, the moments for the posterior distribution are: The posterior distribution is then multivariate normal
Pic.68
Appendix D: How to Estimate a BVAR: Conjugate Priors Note that in the case of the Conjugate priors w
Appendix D: How to Estimate a BVAR: Conjugate Priors Note that in the case of the Conjugate priors we rely on the following VAR representation … while in the Minnesota priors case we employed Though, …
Pic.69
Appendix E: Prior and Posterior distributions in Kadiyala and Karlsson (1997)
Appendix E: Prior and Posterior distributions in Kadiyala and Karlsson (1997)
Pic.70
Appendix E: Posterior distributions of forecast for unemployment and industrial production in K&
Appendix E: Posterior distributions of forecast for unemployment and industrial production in K&K (1997), h=4, T0 =1985:4
Pic.71
Appendix E: Posterior distribution of the unemployment rate forecast in K&K (1997)
Appendix E: Posterior distribution of the unemployment rate forecast in K&K (1997)
Pic.72
Appendix E: Choosing λ
Appendix E: Choosing λ


Скачать презентацию

Если вам понравился сайт и размещенные на нем материалы, пожалуйста, не забывайте поделиться этой страничкой в социальных сетях и с друзьями! Спасибо!