Lecture 3 Chiong and Shum Random Projection Estimation

  • Slides: 23
Download presentation
Lecture 3: Chiong and Shum, “Random Projection Estimation of Discrete-Choice Models with Large Choice

Lecture 3: Chiong and Shum, “Random Projection Estimation of Discrete-Choice Models with Large Choice Sets” Stephen P. Ryan Olin Business School Washington University in St. Louis

Overview • Discrete choice models are ubiquitous workhorse in many settings • Most important

Overview • Discrete choice models are ubiquitous workhorse in many settings • Most important decisions of your life are discrete: • • • Job Housing Car Education Spouse (perhaps dynamic!) • Enormous amount of research on these models, starting with Mc. Fadden (1974) and moving forward • Chiong and Shum (2017): What to do when you have a huge number of choices?

Review of discrete choice models •

Review of discrete choice models •

Choice Probabilities •

Choice Probabilities •

So What’s the Problem? • Consider the problem of buying a place in STL

So What’s the Problem? • Consider the problem of buying a place in STL • Let’s go to Redfin and see what it says about our choice set…

Hmm… • That didn’t seem so bad? • I think we could plausibly estimate

Hmm… • That didn’t seem so bad? • I think we could plausibly estimate a model with 2, 142 choices • But instead of thinking about your decision at the margin, we thought about modeling utility functions of everyone who is currently located somewhere? • Apparently there are 141, 312 occupied housing units in STL • What about some other important situations? • What about your choice of a mate?

Maybe this starts to get a little harder • The set of available mates

Maybe this starts to get a little harder • The set of available mates in a given period of time is fairly large • Population of Germany is about 83 million • Assuming uniform distribution of ages, people live to 83, one million people your age • How many possible combinations of one million people? • Answer: A LOT • Many, many settings look like this either with lots of choices outright or in combination with timing, location, bundling, etc. • So what is the problem? Two-fold: 1. Computational time of running model with millions of choices 2. The probabilities are getting really, really small -> this can cause problems computationally

Chiong and Shum’s Solution • Random projection • A technique from the ML literature

Chiong and Shum’s Solution • Random projection • A technique from the ML literature for dimension reduction • Essential idea: take a high dimension vector and encode it in a lower dimension approximation • Combine with convex-analytic property of discrete choice models (cyclic monotonicity) to generate inequality-based estimator for underlying parameters • Data: aggregate market shares (probabilities scaled up) • Their model relaxes assumptions on error term: unspecified distribution and correlation across errors

Convex Analysis and Discrete Choice •

Convex Analysis and Discrete Choice •

Cyclic Monotonicity •

Cyclic Monotonicity •

Alternative Approaches • Estimation approaches with large number of choices is not a new

Alternative Approaches • Estimation approaches with large number of choices is not a new problem • Fox (2007) uses maximum-score approach to estimate semiparametric discrete choice model using pairwise inequalities • Relies on rank-order property, generalized version of IIA satisfied by exchangeable errors • Train, Mc. Fadden, Ben-Akiva (1987) sample the choices made by consumers • Depends on multinomial logit structure • Present paper allows for non-exchangeable joint error, like nested logit

Random Projection •

Random Projection •

Sparse Random Projection Matrix

Sparse Random Projection Matrix

Compressed Estimator • Where the X is replaced with the compressed version. Equivalently, one

Compressed Estimator • Where the X is replaced with the compressed version. Equivalently, one can write:

Johnson-Lindenstrauss Lemma • There exists a linear map such that a low-distortion embedding: •

Johnson-Lindenstrauss Lemma • There exists a linear map such that a low-distortion embedding: • Reduced dimensional k trades off tractability and error • Does not depend on d, original dimension of data

Convergence • Authors show that for any data, D, estimator using compressed data converges

Convergence • Authors show that for any data, D, estimator using compressed data converges in probability to corresponding estimator using all data as k grows large • From property of random projection matrix, argue pointwise convergence, then invoke Convexity Lemma of Pollard (1991) that all pointwise and uniform convergence same for convex functions

The Payoff: Identified Sets

The Payoff: Identified Sets

Monte Carlo Evidence

Monte Carlo Evidence

Monte Carlo Evidence

Monte Carlo Evidence

Application: Supermarket Scanner Data •

Application: Supermarket Scanner Data •

Results

Results

Conclusion • Injection of machine learning methods inside nonlinear model • Random projection methods

Conclusion • Injection of machine learning methods inside nonlinear model • Random projection methods reduce dimensionality of choice set • Semi-parametric approach closest to Fox (2007) • Allows for zero share products • Randomness -> positive privacy implications