Estimation of a Ramsay-Curve IRT Model using the Metropolis-Hastings Robbins-Monro Algorithm

Similar documents
On Implementation of the Markov Chain Monte Carlo Stochastic Approximation Algorithm

Estimation of the Markov-switching GARCH model by a Monte Carlo EM algorithm

Calibration of Interest Rates

WC-5 Just How Credible Is That Employer? Exploring GLMs and Multilevel Modeling for NCCI s Excess Loss Factor Methodology

COS 513: Gibbs Sampling

Consistent estimators for multilevel generalised linear models using an iterated bootstrap

Stochastic Approximation Algorithms and Applications

Maximum Likelihood Estimation

Technical Appendix: Policy Uncertainty and Aggregate Fluctuations.

A NEW POINT ESTIMATOR FOR THE MEDIAN OF GAMMA DISTRIBUTION

The histogram should resemble the uniform density, the mean should be close to 0.5, and the standard deviation should be close to 1/ 12 =

Relevant parameter changes in structural break models

Estimating Mixed Logit Models with Large Choice Sets. Roger H. von Haefen, NC State & NBER Adam Domanski, NOAA July 2013

Bayesian Multinomial Model for Ordinal Data

A Test of the Normality Assumption in the Ordered Probit Model *

Computational Statistics Handbook with MATLAB

A comment on Christoffersen, Jacobs and Ornthanalai (2012), Dynamic jump intensities and risk premiums: Evidence from S&P500 returns and options

Using MCMC and particle filters to forecast stochastic volatility and jumps in financial time series

PARAMETRIC AND NON-PARAMETRIC BOOTSTRAP: A SIMULATION STUDY FOR A LINEAR REGRESSION WITH RESIDUALS FROM A MIXTURE OF LAPLACE DISTRIBUTIONS

Probits. Catalina Stefanescu, Vance W. Berger Scott Hershberger. Abstract

Monitoring Processes with Highly Censored Data

An Improved Saddlepoint Approximation Based on the Negative Binomial Distribution for the General Birth Process

Market Risk Analysis Volume I

Week 1 Quantitative Analysis of Financial Markets Distributions B

Likelihood-based Optimization of Threat Operation Timeline Estimation

Application of MCMC Algorithm in Interest Rate Modeling

A Two-Step Estimator for Missing Values in Probit Model Covariates

Lecture 2: Forecasting stock returns

Lecture 2: Forecasting stock returns

Chapter 2 Uncertainty Analysis and Sampling Techniques

Inverse reinforcement learning from summary data

Iranian Journal of Economic Studies. Inflation Behavior in Top Sukuk Issuing Countries : Using a Bayesian Log-linear Model

Improving the accuracy of estimates for complex sampling in auditing 1.

A Practical Implementation of the Gibbs Sampler for Mixture of Distributions: Application to the Determination of Specifications in Food Industry

Empirical Distribution Testing of Economic Scenario Generators

A Markov Chain Monte Carlo Approach to Estimate the Risks of Extremely Large Insurance Claims

Extracting Information from the Markets: A Bayesian Approach

Asymmetric Type II Compound Laplace Distributions and its Properties

Gamma Distribution Fitting

Investing through Economic Cycles with Ensemble Machine Learning Algorithms

# generate data num.obs <- 100 y <- rnorm(num.obs,mean = theta.true, sd = sqrt(sigma.sq.true))

1 Explaining Labor Market Volatility

A Hidden Markov Model Approach to Information-Based Trading: Theory and Applications

On the Distributional Assumptions in the StoNED model

Modelling component reliability using warranty data

Unobserved Heterogeneity Revisited

SELECTION OF VARIABLES INFLUENCING IRAQI BANKS DEPOSITS BY USING NEW BAYESIAN LASSO QUANTILE REGRESSION

Market Correlations in the Euro Changeover Period With a View to Portfolio Management

Lecture 9: Markov and Regime

Outline. Review Continuation of exercises from last time

Getting started with WinBUGS

Option Pricing Using Bayesian Neural Networks

Fast Convergence of Regress-later Series Estimators

Adaptive Interest Rate Modelling

Fitting parametric distributions using R: the fitdistrplus package

Appendix. A.1 Independent Random Effects (Baseline)

Analysis of extreme values with random location Abstract Keywords: 1. Introduction and Model

Lecture 8: Markov and Regime

Adaptive Metropolis-Hastings samplers for the Bayesian analysis of large linear Gaussian systems

Empirical Analysis of the US Swap Curve Gough, O., Juneja, J.A., Nowman, K.B. and Van Dellen, S.

An Implementation of Markov Regime Switching GARCH Models in Matlab

International Journal of Computer Engineering and Applications, Volume XII, Issue II, Feb. 18, ISSN

GENERATION OF STANDARD NORMAL RANDOM NUMBERS. Naveen Kumar Boiroju and M. Krishna Reddy

LOSS SEVERITY DISTRIBUTION ESTIMATION OF OPERATIONAL RISK USING GAUSSIAN MIXTURE MODEL FOR LOSS DISTRIBUTION APPROACH

Using New SAS 9.4 Features for Cumulative Logit Models with Partial Proportional Odds Paul J. Hilliard, Educational Testing Service (ETS)

Monte Carlo Methods for Uncertainty Quantification

Journal of Economics and Financial Analysis, Vol:1, No:1 (2017) 1-13

Financial Econometrics Notes. Kevin Sheppard University of Oxford

Lecture 17: More on Markov Decision Processes. Reinforcement learning

International Journal of Computer Engineering and Applications, Volume XII, Issue II, Feb. 18, ISSN

MCMC Package Example

Adjusted Priors for Bayes Factors Involving Reparameterized Order Constraints

ADVANCED OPERATIONAL RISK MODELLING IN BANKS AND INSURANCE COMPANIES

The Time-Varying Effects of Monetary Aggregates on Inflation and Unemployment

ELEMENTS OF MONTE CARLO SIMULATION

Importance sampling and Monte Carlo-based calibration for time-changed Lévy processes

Efficiency Measurement with the Weibull Stochastic Frontier*

Is the Ex ante Premium Always Positive? Evidence and Analysis from Australia

Supplementary material for the paper Identifiability and bias reduction in the skew-probit model for a binary response

Bayesian Estimation of the Markov-Switching GARCH(1,1) Model with Student-t Innovations

M.S. in Quantitative Finance & Risk Analytics (QFRA) Fall 2017 & Spring 2018

The Leveled Chain Ladder Model. for Stochastic Loss Reserving

Discrete Choice Methods with Simulation

Estimating a Dynamic Oligopolistic Game with Serially Correlated Unobserved Production Costs. SS223B-Empirical IO

Monte Carlo and Empirical Methods for Stochastic Inference (MASM11/FMSN50)

Omitted Variables Bias in Regime-Switching Models with Slope-Constrained Estimators: Evidence from Monte Carlo Simulations

Volatility Models and Their Applications

STARRY GOLD ACADEMY , , Page 1

Algorithmic Trading using Reinforcement Learning augmented with Hidden Markov Model

Calibration of PD term structures: to be Markov or not to be

Fitting financial time series returns distributions: a mixture normality approach

[D7] PROBABILITY DISTRIBUTION OF OUTSTANDING LIABILITY FROM INDIVIDUAL PAYMENTS DATA Contributed by T S Wright

Modeling skewness and kurtosis in Stochastic Volatility Models

Accelerated Option Pricing Multiple Scenarios

Extend the ideas of Kan and Zhou paper on Optimal Portfolio Construction under parameter uncertainty

Risk Measurement in Credit Portfolio Models

1 Introduction. Term Paper: The Hall and Taylor Model in Duali 1. Yumin Li 5/8/2012

1 Residual life for gamma and Weibull distributions

Calibrating Financial Models Using Consistent Bayesian Estimators

No Predictable Components in G7 Stock Returns

Transcription:

1 / 34 Estimation of a Ramsay-Curve IRT Model using the Metropolis-Hastings Robbins-Monro Algorithm Scott Monroe & Li Cai IMPS 2012, Lincoln, Nebraska

Outline 2 / 34 1 Introduction and Motivation 2 Review of Ramsay Curves 3 Review of the MH-RM Algorithm 4 Simulation Study 5 Empirical Application 6 Conclusion

Outline 3 / 34 1 Introduction and Motivation 2 Review of Ramsay Curves 3 Review of the MH-RM Algorithm 4 Simulation Study 5 Empirical Application 6 Conclusion

Ramsay-Curve IRT 4 / 34 The need to estimate the latent trait distribution has received increasing attention in recent years. One approach is Ramsay Curve IRT, (RC-IRT, Woods & Thissen, 2006), which uses Bock and Aitkin s (1981) EM algorithm for estimation. However, the EM algorithm does not yield the observed information matrix upon convergence. Hence, no standard errors are available.

Ramsay-Curve IRT 5 / 34 The need to estimate the latent trait distribution has received increasing attention in recent years. One approach is Ramsay Curve IRT, (RC-IRT, Woods & Thissen, 2006), which uses Bock and Aitkin s (1981) EM algorithm for estimation. However, the EM algorithm does not yield the observed information matrix upon convergence. Hence, no standard errors are available. A solution: use the Metropolis-Hastings Robbins-Monro (MH-RM, Cai, 2010) algorithm.

Outline 6 / 34 1 Introduction and Motivation 2 Review of Ramsay Curves 3 Review of the MH-RM Algorithm 4 Simulation Study 5 Empirical Application 6 Conclusion

Review of Ramsay Curves For full details, see Woods and Thissen (2006). Given observed latent trait scores, RCs can estimate the distribution s shape. Essentially, the method uses B-spline regression, with a small number of parameters. 0.00 0.01 0.02 0.03 0.04 Densities Ramsay Curve Emp. Hist. -6-4 -2 0 2 4 6 7 / 34

Outline 8 / 34 1 Introduction and Motivation 2 Review of Ramsay Curves 3 Review of the MH-RM Algorithm 4 Simulation Study 5 Empirical Application 6 Conclusion

9 / 34 A Review of MH-RM Metropolis-Hastings Robbins-Monro (Cai, 2010) is motivated by: Fisher (1925) The gradient of the observed log likelihood = the expectation of the gradient of the complete log likelihood.

A Review of MH-RM 10 / 34 Metropolis-Hastings Robbins-Monro (Cai, 2010) is motivated by: Fisher (1925) The gradient of the observed log likelihood = the expectation of the gradient of the complete log likelihood. Iterative approach: Use MH sampler to find an MC approximation of the expectation. Use this approximation to update the parameter estimates.

A Review of MH-RM 11 / 34 Metropolis-Hastings Robbins-Monro (Cai, 2010) is motivated by: Fisher (1925) The gradient of the observed log likelihood = the expectation of the gradient of the complete log likelihood. Iterative approach: Use MH sampler to find an MC approximation of the expectation. Use this approximation to update the parameter estimates. Since the approximations are noisy, use the RM method to filter. The filtering operates through decreasing gain constants.

MH-RM Estimation in a Picture 12 / 34 Stage 1 (iterations 1-800): move to neighborhood of MLE. Parameter Estimate -0.7-0.5-0.3-0.1 0 800 1000 1550 Iteration Number

13 / 34 MH-RM Estimation in a Picture Stage 1 (iterations 1-800): move to neighborhood of MLE. Stage 2 (iterations 801-1000): find starting values for Stage 3 by averaging estimates across Stage 2 iterations. Parameter Estimate -0.7-0.5-0.3-0.1 0 800 1000 1550 Iteration Number

14 / 34 MH-RM Estimation in a Picture Stage 1 (iterations 1-800): move to neighborhood of MLE. Stage 2 (iterations 801-1000): find starting values for Stage 3 by averaging estimates across Stage 2 iterations. Stage 3 (iterations 1001-convergence): obtain MLE by using decreasing gain constants to average out the noise. Parameter Estimate -0.7-0.5-0.3-0.1 0 800 1000 1550 Iteration Number

15 / 34 Standard Errors in MH-RM Louis (1982) The observed information can be expressed as a function of the complete log likelihood, involving its gradient and hessian. With MH-RM, the pieces used in the approximation can be calculated by: recursive approximation, as in Cai (2010); or Monte Carlo approximation, as in Diebolt and Ip (1996).

Outline 16 / 34 1 Introduction and Motivation 2 Review of Ramsay Curves 3 Review of the MH-RM Algorithm 4 Simulation Study 5 Empirical Application 6 Conclusion

Simulation Study 17 / 34 Simulation Study Conditions Sample Size 1000 Number of Items 25 Slope Dist. N (1.7, 0.3) Location Dist. N (0, 1), truncated at ±2 True g(θ η) Normal, Skewed, and Bimodal Replications 100 Estimation Methods MH-RM and BA-EM

18 / 34 True g(θ η) used in Simulation Study 0.00 0.02 0.04 0.06 Simulation Densities Normal Skewed Bimodal -6-4 -2 0 2 4 6 The Skewed and Bimodal densities were found as mixtures of normals. Latent trait scores were drawn by rejection sampling. Item responses were simulated in accordance with the 2PL model.

Parameter Recovery Log likelihood values for MH-RM (x-axis) and BA-EM (y-axis). 11700 11500 11300 11700 11500 11300 Normal MH RM EM 11700 11500 11300 11700 11500 11300 Skewed MH RM EM 11500 11300 11500 11300 Bimodal MH RM EM Both MH-RM and EM perform maximum marginal likelihood estimation, so we shouldn t expect substantial differences in estimation performance. 19 / 34

Parameter Recovery Average RMSE for item parameters within a replication for MH-RM (x-axis) and BA-EM (y-axis). 0.10 0.14 0.18 0.10 0.14 0.18 Normal MH RM EM 0.10 0.15 0.20 0.25 0.10 0.15 0.20 0.25 Skewed MH RM EM 0.10 0.15 0.20 0.25 0.30 0.10 0.20 0.30 Bimodal MH RM EM Takeaway: MH-RM and BA-EM are producing comparable results. 20 / 34

21 / 34 Ramsay Curve Recovery: Normal g(θ η) Density 6 4 2 0 2 4 6 θ estimated RCs by replication shown in different colors true g(θ η) shown in black

22 / 34 Ramsay Curve Recovery: Skewed g(θ η) Density 6 4 2 0 2 4 6 θ estimated RCs by replication shown in different colors true g(θ η) shown in black

23 / 34 Ramsay Curve Recovery: Bimodal g(θ η) Density 6 4 2 0 2 4 6 θ estimated RCs by replication shown in different colors true g(θ η) shown in black

Standard Error Estimation 24 / 34 For item parameters, average standard errors (x-axis) and Monte Carlo standard deviations (y-axis). Normal Skewed Bimodal Monte Carlo SD 0.05 0.15 0.25 Monte Carlo SD 0.05 0.15 0.25 Monte Carlo SD 0.05 0.15 0.25 0.05 0.15 0.25 0.05 0.15 0.25 0.05 0.15 0.25 Standard Error Standard Error Standard Error Note: For all three g(θ η) shapes, the MC standard deviations are slightly larger.

25 / 34 Standard Error Estimation Examining coverage probabilities (based on 100 replications): 68% coverage 95% coverage g(θ η) a c all a c all Normal 0.66 0.60 0.63 0.94 0.92 0.93 Skewed 0.63 0.63 0.63 0.93 0.93 0.93 Bimodal 0.67 0.64 0.66 0.95 0.93 0.94 a = slope c = intercept all = slopes and intercepts

Outline 26 / 34 1 Introduction and Motivation 2 Review of Ramsay Curves 3 Review of the MH-RM Algorithm 4 Simulation Study 5 Empirical Application 6 Conclusion

Example: National Comorbidity Survey 27 / 34 Background: The National Comorbidity Survey (NCS, 1994) was a nationwide household survey of the United States. Ten Likert items (4-categories) measured current emotional distress. Example Items: During the past 30 days how often did you... worry too much about things? feel exhausted for no good reason? feel hopeless about the future? Response options: never (0), rarely (1), sometimes (2), often (3).

28 / 34 Distress Scale Results: Ramsay Curve 0.00 0.01 0.02 0.03 0.04 0.05 Densities Distress Normal -6-4 -2 0 2 4 6 Note: there appears to be a sizeable group of respondents at θ = 3.

Distress Scale Results: Slope Estimates and Standard Errors 29 / 34 Item 1 2 3 4 5 6 7 8 9 10 Slope Estimates 1.99 1.96 2.08 2.82 2.50 2.51 2.86 2.32 2.35 2.47 1.83 1.81 1.90 2.55 2.31 2.29 2.58 2.14 2.14 2.24 Standard Errors.06.07.05.07.06.07.08.06.06.07.08.09.07.08.07.08.10.07.07.08 Blue values correspond to RC estimation. Black values correspond to IRT estimation with a normal distribution.

Outline 30 / 34 1 Introduction and Motivation 2 Review of Ramsay Curves 3 Review of the MH-RM Algorithm 4 Simulation Study 5 Empirical Application 6 Conclusion

31 / 34 Discussion Overall, results support the validity and utility of the MH-RM implementation. Embedding RC-IRT in MH-RM (as opposed to EM) enables the computation of the observed information matrix. This, in turn, facilitates: test assembly limited-information goodness-of-fit testing differential item functioning

References I 32 / 34 Bock, R. D., & Aitkin, M. (1981). Marginal maximum likelihood estimation of item parameters: Application of an EM algorithm. Psychometrika, 46, 443 459. Cai, L. (2010). High-dimensional exploratory item factor analysis by a Metropolis-Hastings Robbins-Monro algorithm. Psychometrika, 75, 33 57. Cai, L. (2012). flexmirt: Flexible multilevel item factor analysis and test scoring [Computer software]. Seattle, WA: Vector Psychometric Group, LLC. Diebolt, J., & Ip, E. H. S. (1996). Stochastic EM: method and application. In W. Gilks, S. Richardson, & D. Spiegelhalter (Eds.), Markov chain Monte Carlo in practice (p. 259-273). London: Chapman and Hall. Fisher, R. A. (1925). Theory of statistical estimation. Proceedings of the Cambridge Philosophical Society, 22, 700-725.

References II 33 / 34 Louis, T. A. (1982). Finding the observed information matrix when using the EM algorithm. Journal of the Royal Statistical Society - Series B, 44(2), 226-233. Mislevy, R. J. (1984). Estimating latent distributions. Psychometrika, 49(3), 359-381. R Development Core Team. (2010). R: A language and environment for statistical computing [Computer software manual]. Vienna, Austria. Available from http://www.r-project.org/ (ISBN 3-900051-07-0) Woods, C. M., & Lin, N. (2008). Item response theory with estimation of the latent density using Davidian curves. Applied Psychological Measurement, 33(2), 102-117. Woods, C. M., & Thissen, D. (2006). Item response theory with estimation of the latent population distribution using spline-based densities. Psychometrika, 71, 281 301.

Acknowledgements 34 / 34 This research is supported by grants from the Institute of Education Sciences (R305B080016 and R305D100039) and the National Institute on Drug Abuse (R01DA026943 and R01DA030466). Thanks to Larry Thomas and Mark Hansen for providing datasets used in the empirical applications.