Unobserved Heterogeneity Revisited

Similar documents
Chapter 3. Dynamic discrete games and auctions: an introduction

Lecture 17: More on Markov Decision Processes. Reinforcement learning

UCLA Department of Economics Ph.D. Preliminary Exam Industrial Organization Field Exam (Spring 2010) Use SEPARATE booklets to answer each question

The Costs of Environmental Regulation in a Concentrated Industry

Final exam solutions

Making Complex Decisions

Obtaining Analytic Derivatives for a Class of Discrete-Choice Dynamic Programming Models

EE641 Digital Image Processing II: Purdue University VISE - October 29,

Estimating a Dynamic Oligopolistic Game with Serially Correlated Unobserved Production Costs. SS223B-Empirical IO

GMM for Discrete Choice Models: A Capital Accumulation Application

Lecture 7: Bayesian approach to MAB - Gittins index

Lec 1: Single Agent Dynamic Models: Nested Fixed Point Approach. K. Sudhir MGT 756: Empirical Methods in Marketing

Pakes (1986): Patents as Options: Some Estimates of the Value of Holding European Patent Stocks

Estimating Mixed Logit Models with Large Choice Sets. Roger H. von Haefen, NC State & NBER Adam Domanski, NOAA July 2013

A simple wealth model

A potentially useful approach to model nonlinearities in time series is to assume different behavior (structural break) in different subsamples

Handout 4: Deterministic Systems and the Shortest Path Problem

Stochastic Games and Bayesian Games

Financial Liberalization and Neighbor Coordination

EE266 Homework 5 Solutions

AM 121: Intro to Optimization Models and Methods

Identifying Dynamic Discrete Choice Models. off Short Panels

SUPPLEMENT TO EQUILIBRIA IN HEALTH EXCHANGES: ADVERSE SELECTION VERSUS RECLASSIFICATION RISK (Econometrica, Vol. 83, No. 4, July 2015, )

Notes on the EM Algorithm Michael Collins, September 24th 2005

Identification and Estimation of Dynamic Games when Players Beliefs are not in Equilibrium

Note on Oblivious equilibrium

Graduate Macro Theory II: The Basics of Financial Constraints

GRANULARITY ADJUSTMENT FOR DYNAMIC MULTIPLE FACTOR MODELS : SYSTEMATIC VS UNSYSTEMATIC RISKS

Monte Carlo and Empirical Methods for Stochastic Inference (MASM11/FMSN50)

Chapter 7: Estimation Sections

Labor Migration and Wage Growth in Malaysia

Online Appendix (Not intended for Publication): Federal Reserve Credibility and the Term Structure of Interest Rates

Identification and Counterfactuals in Dynamic Models of Market Entry and Exit

Economics 2010c: Lecture 4 Precautionary Savings and Liquidity Constraints

17 MAKING COMPLEX DECISIONS

Sunk Costs, Entry, and Exit in Dynamic Oligopoly

1 Dynamic programming

Estimating Market Power in Differentiated Product Markets

INCOME DISTRIBUTION, HOUSEHOLD HETEROGENEITY AND CONSUMPTION INSURANCE IN THE UK: A MIXTURE MODEL APPROACH

Econometric Methods for Valuation Analysis

Computer Vision Group Prof. Daniel Cremers. 7. Sequential Data

4 Reinforcement Learning Basic Algorithms

Taxing Firms Facing Financial Frictions

Information Aggregation in Dynamic Markets with Strategic Traders. Michael Ostrovsky

On modelling of electricity spot price

The method of Maximum Likelihood.

Point Estimators. STATISTICS Lecture no. 10. Department of Econometrics FEM UO Brno office 69a, tel

Risk Premia and the Conditional Tails of Stock Returns

Choice Models. Session 1. K. Sudhir Yale School of Management. Spring

Chapter 5 Univariate time-series analysis. () Chapter 5 Univariate time-series analysis 1 / 29

Markov Decision Processes: Making Decision in the Presence of Uncertainty. (some of) R&N R&N

When do Secondary Markets Harm Firms? Online Appendixes (Not for Publication)

Chapter 4: Asymptotic Properties of MLE (Part 3)

Long-Run Market Configurations in a Dynamic Quality-Ladder Model with Externalities. June 2, 2018

Steven Heston: Recovering the Variance Premium. Discussion by Jaroslav Borovička November 2017

Financial Econometrics Jeffrey R. Russell. Midterm 2014 Suggested Solutions. TA: B. B. Deng

Econ 8602, Fall 2017 Homework 2

ADVANCED MACROECONOMIC TECHNIQUES NOTE 7b

Directory. The Value of Life and the Rise in Health Spending p.1/34

16 MAKING SIMPLE DECISIONS

Decision Theory: Value Iteration

Frequency of Price Adjustment and Pass-through

Stochastic Differential Equations in Finance and Monte Carlo Simulations

Sequential Decision Making

Modelling Returns: the CER and the CAPM

Asset Pricing with Heterogeneous Consumers

A Markov Chain Monte Carlo Approach to Estimate the Risks of Extremely Large Insurance Claims

The Multinomial Logit Model Revisited: A Semiparametric Approach in Discrete Choice Analysis

Markov Decision Processes

INTERTEMPORAL ASSET ALLOCATION: THEORY

Handout 8: Introduction to Stochastic Dynamic Programming. 2 Examples of Stochastic Dynamic Programming Problems

Sentiments and Aggregate Fluctuations

Martingale Pricing Theory in Discrete-Time and Discrete-Space Models

Stochastic Games and Bayesian Games

Chapter 7: Estimation Sections

Slides III - Complete Markets

University of Toronto Department of Economics. Identification and estimation of dynamic games when players' beliefs are not in equilibrium

Consumption, Investment and the Fisher Separation Principle

Monetary Economics Final Exam

Dynamic Asset Pricing Model

Multi-armed bandit problems

Portfolio Investment

Economics 502. Nominal Rigidities. Geoffrey Dunbar. UBC, Fall November 22, 2012

Graduate Macro Theory II: Fiscal Policy in the RBC Model

Financial Risk Management

Equity correlations implied by index options: estimation and model uncertainty analysis

Regret Minimization and Security Strategies

Maturity, Indebtedness and Default Risk 1

Macroeconomics 2. Lecture 7 - Labor markets: Introduction & the search model March. Sciences Po

Small Sample Performance of Instrumental Variables Probit Estimators: A Monte Carlo Investigation

AnEstimableDynamicModelofEntry,Exit. and Growth in Oligopoly Retail Markets

Idiosyncratic risk and the dynamics of aggregate consumption: a likelihood-based perspective

Financial Constraints and the Franchising Decision

Ramsey s Growth Model (Solution Ex. 2.1 (f) and (g))

Economic stability through narrow measures of inflation

Online Appendix: Asymmetric Effects of Exogenous Tax Changes

Equity, Vacancy, and Time to Sale in Real Estate.

Resolving Failed Banks: Uncertainty, Multiple Bidding, & Auction Design

What s New in Econometrics. Lecture 11

Making Decisions. CS 3793 Artificial Intelligence Making Decisions 1

Transcription:

Unobserved Heterogeneity Revisited Robert A. Miller Dynamic Discrete Choice March 2018 Miller (Dynamic Discrete Choice) cemmap 7 March 2018 1 / 24

Distributional Assumptions about the Unobserved Variables A trade off We have already shown that the model is exactly identified up to a normalization if the distribution of unobserved variables is known. The model is underidentifed for counterfactuals on transitions. Assumptions on preferences and transitions can help: for example nonstationary transitions and stable preferences (an exclusion restriction). What if we want to relax assumption that the distribution of unobserved variables is known? Then we must place assumptions on the way systematic payoffs are parameterized: note these are identifying assumptions. Miller (Dynamic Discrete Choice) cemmap 7 March 2018 2 / 24

Motivating Example Rust s (1987) bus engine revisited Recall Mr. Zurcher decides whether to replace the existing engine (d 1t = 1), or keep it for at least one more period (d 2t = 1). Bus mileage advances 1 unit (x t+1 = x t + 1) if Zurcher keeps the engine (d 2t = 1) and is set to zero otherwise(x t+1 = 0 if d 1t = 1). Transitory iid choice-specific shocks, ɛ jt are Type 1 Extreme value. Zurcher sequentially maximizes expected discounted sum of payoffs: } E { β t 1 [d 2t (θ 1 x t + θ 2 s + ɛ 2t ) + d 1t ɛ 1t ] t=1 Miller (Dynamic Discrete Choice) cemmap 7 March 2018 3 / 24

Motivating Example ML Estimation when CCP s are known (infeasible) To show how the EM algorithm helps, consider the infeasible case where s {1,..., S} is unobserved but p(x, s) is known. Let π s denote population probability of being in unobserved state s. Supposing β is known the ML estimator for this "easier" problem is: {ˆθ, ˆπ} = arg max θ,π N n=1 ln [ S T π s l(d nt x nt, s, p, θ) s=1 t=1 where p p(x, s) is a string of probabilities assigned/estimated for each (x, s) and l(d nt x nt, s n, p, θ) is derived from our representation of the conditional valuation functions and takes the form: d 1nt + d 2nt exp(θ 1 x nt + θ 2 s + β ln [p(0, s)] β ln [p(x nt + 1, s)] 1 + exp(θ 1 x nt + θ 2 s + β ln [p(0, s)] β ln [p(x nt + 1, s)]) Maximizing over the sum of a log of summed products is computationally burdensome. Miller (Dynamic Discrete Choice) cemmap 7 March 2018 4 / 24 ]

Motivating Example Why EM is attractive (when CCP s are known) The EM algorithm is a computationally attractive alternative to directly maximizing the likelihood. Denote by d n (d n1,..., d nt ) and x n (x n1,..., x nt ) the full sequence of choices and mileages observed in the data for bus n. At the m th iteration: q (m+1) ns = Pr = θ (m+1) = arg max θ { s d n, x n, θ (m), π (m) s }, p π (m) s T t=1 l(d nt x nt, s, p, θ (m) ) S s =1 π(m) s T t=1 l(d nt x nt, s, p, θ (m) ) π s (m+1) = N 1 N q ns (m+1) n=1 N n=1 S s=1 T t=1 q ns (m+1) ln[l(d nt x nt, s, p, θ)] Miller (Dynamic Discrete Choice) cemmap 7 March 2018 5 / 24

Motivating Example Steps in our algorithm when s is unobserved and CCP s are unknown Our algorithm begins by setting initial values for θ (1), π (1), and p (1) ( ): Step 1 Compute q (m+1) ns q (m+1) ns = Step 2 Compute π (m+1) s as: π (m) s S s =1 π(m) s T t=1 l [d nt x nt, s, p (m), θ (m)] T t=1 l (d nt x nt, s, p (m), θ (m)) according to: π (m+1) s = N n=1 q ns (m+1) N Step 3 Update p (m+1) (x, s) using one of two rules below Step 4 Obtain θ (m+1) from: θ (m+1) = arg max θ N n=1 S s=1 T t=1 [ ( ) q ns (m+1) ln l d nt x nt, s n, p (m+1), θ Miller (Dynamic Discrete Choice) cemmap 7 March 2018 6 / 24

Motivating Example Updating the CCP s Take a weighted average of decisions to replace engine, conditional on x, where weights are the conditional probabilities of being in unobserved state s. Step 3A Update CCP s with: p (m+1) (x, s) = N n=1 T t=1 d 1nt q (m+1) ns I (x nt = x) N n=1 T t=1 q (m+1) ns I (x nt = x) Or in a stationary infinite horizon model use identity from model that likelihood returns CCP of replacing the engine: Step 3B Update CCP s with: p (m+1) (x nt, s n ) = l(d nt1 = 1 x nt, s n, p (m), θ (m) ) Miller (Dynamic Discrete Choice) cemmap 7 March 2018 7 / 24

First Monte Carlo Finite horizon renewal problem Suppose s {0, 1 } equally weighted. There are two observed state variables 1 total accumulated mileage: x 1t+1 = { t if d 1t = 1 x 1t + t if d 2t = 1 2 permanent route characteristic for the bus, x 2, that systematically affects miles added each period. We assume t {0, 0.125,..., 24.875, 25} is drawn from: f ( t x 2 ) = exp [ x 2 ( t 25)] exp [ x 2 ( t 24.875)] and x 2 is a multiple 0.01 drawn from a discrete equi-probability distribution between 0.25 and 1.25. Miller (Dynamic Discrete Choice) cemmap 7 March 2018 8 / 24

First Monte Carlo Finite horizon renewal problem Let θ 0t be an aggregate shock (denoting cost fluctuations say). The difference in current payoff from retaining versus replacing the engine is: u 2t (x 1t, s) u 1t (x 1t, s) θ 0t + θ 1 min {x 1t, 25} + θ 2 s Denoting the observed state variables by x t (x 1t, x 2 ), this translates to: v 2t (x t, s) v 1t (x t, s) = θ 0t + θ 1 min {x 1t, 25} + θ 2 s { [ ]} p 1t (0, s) +β ln f ( t x 2 ) t p Λ 1t (x 1t + t, s) Miller (Dynamic Discrete Choice) cemmap 7 March 2018 9 / 24

First Monte Carlo Table 1 of Arcidiacono and Miller (2011) Mean and standard deviations for 50 simulations. For columns 1 6, the observed data consist of 1000 buses for 20 periods. For columns 7 and 8, the intercept (θ0) is allowed. to Miller vary (Dynamic over Discrete time Choice) and the data consist cemmapof 7 2000 buses for 10 periods. March 2018 10 / 24

The Estimators Maximization We parameterize u jt (z t ) and G (ɛ t ) by θ, f jt (z t+1 z t ) with α, and following our motivating example, we define two estimators. Given any conditional choice probability mapping p, both maximize the joint log likelihood: ( θ, π, α) = arg max (θ,π,α,) N n=1 S s=1 π s log L (d n, x n x n1, s ; θ, p) where L (d n, x n x n1, s ; θ, p) is the likelihood of the (panel length) sequence (d n, x n ): = L (d n, x n x n1, s ; θ, π, p) T J t=1 j=1 d jnt l jt (x nt, s, θ, π, p)f jt (x n,t+1 x nt, s, θ) Miller (Dynamic Discrete Choice) cemmap 7 March 2018 11 / 24

The Estimators Using the Likelihood to obtain the CCP s The difference between the estimators arises from how p is defined. The first estimator is based on the fact that l jt (x nt, s n, θ, α, π, p) is the likelihood of observing individual n make choice j at time t given s n. Accordingly define p (x, s) to solve: p jt (x, s) = l jt (x, s; θ, α, π, p) The large sample properties are standard. Miller (Dynamic Discrete Choice) cemmap 7 March 2018 12 / 24

The Estimators An empirical approach to the CCP s Let L n (s n = s) denote the joint likelihood of the data for n and being in unobserved state s evaluated at ( θ, α, π, p). ( ) L n (s n = s) π s L d n, x n x n1, s ; θ, p Also denote ( by ) L n the likelihood of observing (d n, x n ) given parameter values π, θ, p : L n S ( ) s=1 π sl d n, x n x n1, s ; θ, p = S s=1 L n (s n = s) As an estimated sample approximation, N 1 N n=1 the fraction of the population in s. ] [ L n (s n = s)/ L n is Miller (Dynamic Discrete Choice) cemmap 7 March 2018 13 / 24

The Estimators Another CCP "fixed point" Similarly: ] 1 N 1 N n=1 [I (x nt = x) L n (s n = s)/ L n is the estimated fraction of the population in s with x at t. ] 2 N 1 N n=1 [d jnt I (x nt = x) L n (s n = s)/ L n is the estimated fraction choosing j at t as well. We define: p jt (x, s) = [ N ] /[ n (s n = s) N ] n (s n = s) d jnt I (x nt = x) L I (x nt = x) L n=1 L n n=1 L n Compared to the first one this estimator has similar properties but imposes less structure. Miller (Dynamic Discrete Choice) cemmap 7 March 2018 14 / 24

Unobserved Markov Chain Extending the estimation framework Up until now we have been assuming that the unobserved component is time invariant. Now suppose s t is a Markov chain where π (s t+1 s t ) is an exogenous probability transition where: f jt (x t+1, s t+1 x t, s t ) = π (s t+1 s t ) f jt (x t+1 x t, s t ) and π 1 (s 1 x 1 ) is the probability of being in (unobserved) state s 1 conditional on (observed) state x 1 in the first period. The intuition for the simpler case follows through in this generalization. Miller (Dynamic Discrete Choice) cemmap 7 March 2018 15 / 24

Unobserved Markov Chain Extending the algorithm We obtain the CCP s, p (m) (x, s), at the m th step using one of the two ways we described above. The EM algorithm is used in obtaining θ (m) in the same way as before. So we only have to determine:, the probability of n being in unobserved state s at time t, 2 π (m) 1 (s 1 x 1 ), the probability distribution over the initial unobserved states conditional on the initial observed states, 3 π (m) (s s), the transition probabilities of the unobserved states. 1 q (m) nst Miller (Dynamic Discrete Choice) cemmap 7 March 2018 16 / 24

Unobserved Markov Chain The remaining pieces of the algorithm 1 q (m+1) nst follows from Bayes rule: 2 Averaging over q (m+1) ns1 : π (m+1) q (m+1) nst = L(m) n (s nt = s) L n (m) 1 (s x ) = N n=1 q (m+1) ns1 I (x n1 = x) n=1 N I (x n1 = x) 3 Let q ns t s denote the probability of n being in unobserved state s at time t conditional on the data and also on being in unobserved state s at time t 1. We base π (m+1) (s s) on sample analogs of the identity: π(s s) = E [ ] n qns t sq nst 1 E n [q nst 1 ] where E n is the expectation taken over the population. Miller (Dynamic Discrete Choice) cemmap 7 March 2018 17 / 24

A Third Two Stage Estimator An unrestricted first stage estimator Alternatively, subject to identification, we could estimate unrestricted CCPs in a first stage, and then plug them into the structural part of the econometric model. 1 Estimate f jt (x n,t+1 x nt, s n, α) that is α and p jt (x, s) from an unrestricted likelihood formed from: J j=1 [ pjt (x nt, s n ) f jt (x n,t+1 x nt, s n, α) ] d jnt 2 Estimate θ using the conditional choice probabilities and the unobserved transitions on the unobserved variables obtained in the first step. Apart from not estimating θ, the essential difference in the first stage of this alternative estimator and the previous one, is that the likelihood components ) of this one come from p jt (x, s) not x, L j (d t s; θ, α, π, p. Miller (Dynamic Discrete Choice) cemmap 7 March 2018 18 / 24

Second Monte Carlo Structure Entrants pay startup cost to compete in the market, but not incumbents. Paying startup cost now transforms entrant into incumbent next period. Declining to compete in any given period is tantamount to exit. When a firm exits another firm potentially enters next period. Miller (Dynamic Discrete Choice) cemmap 7 March 2018 19 / 24

Second Monte Carlo Dynamics There are two sources of dynamics in this model. An entrant depreciates startup cost over its anticipated lifetime. Since it is more costly for an entrant to start operations, than for an incumbent to continue, the number of incumbents signals how much competition the firm faces in the current period, and consequently affects its own decision whether to exit the industry or not. Miller (Dynamic Discrete Choice) cemmap 7 March 2018 20 / 24

Second Monte Carlo Two observed state variables Each market has a permanent market characteristic, denoted by x 1, common to each player within the market and constant over time, but differing independently across markets, with equal probabilities on support {1,..., 10}. The number of firm exits in the previous period is also common knowledge to the market, and this variable is indicated by: x 2t I d (h) 1,t 1 h=1 This variable is a useful predictor for the number of firms that will compete in the current period. Intuitively, the more players paying entry costs, the lower the expected number of competitors. Miller (Dynamic Discrete Choice) cemmap 7 March 2018 21 / 24

Second Monte Carlo Unobserved (Markov chain state) variables, and price equation The unobserved state variable s t {1,..., 5} follows a first order Markov chain. We assume that the probability of the unobserved variable remaining unchanged in successive periods is fixed at some π (0, 1), and that if the state does change, any other state is equally likely to occur with probability (1 π) /4. We generated also price data on each market, denoted by w t, with the equation: w t = α 0 + α 1 x + α 2 s t + α 3 I d (h) 1t + η t h=1 where η t is distributed as a standard normal disturbance independently across markets and periods, revealed to each market after the entry and exit decisions are made. Miller (Dynamic Discrete Choice) cemmap 7 March 2018 22 / 24

Second Monte Carlo Utility and number of firms and markets The flow payoff of an active firm i in period t, net of private information ɛ (i) 2t is modeled as: U 2 ( x (i) t, s (i) t ), d ( i) I t = θ 0 + θ 1 x + θ 2 s t + θ 3 d (h) 1t + θ 4 d (i) 1,t 1 h=1 ) We normalize exit utility as U 1 ( x (i) t, s (i) t, d ( i) t = 0 We assume ɛ (i) jt is distributed as Type 1 Extreme Value. The number of firms in each market in our experiment is 6. We simulated data for 3, 000 markets, and set β = 0.9. Starting at an initial date with 6 entrants in the market, we ran the simulations forward for twenty periods. Miller (Dynamic Discrete Choice) cemmap 7 March 2018 23 / 24

Second Monte Carlo Table 2 of Arcidiacono and Miller (2011) Mean and standard deviations for 100 simulations. Observed data consist of 3000 markets for 10 periods with 6 firms in each market. In column 7, the CCP s are updated with the model. Miller (Dynamic Discrete Choice) cemmap 7 March 2018 24 / 24