ECON 452 (A & B) Winter 2001 C. Ferrall / A. Gregory

OUTLINE FOR PART II

  1. Introduction

  2.  
    1. The goal of Part II is for you to learn three skills:
      1. How to carry out econometric studies using data from cross-sectional surveys (in particular, DLI surveys stored in the QED Data Archive).
      2. How to report results from your study so people will understand them
      3. How to model a kind of variable that often appears in survey data: limited-dependent variables.

      4.  
    2. Some Questions About These Goals
      1. Are these good goals?
      2. Does the focus in Part II complement the focus in part I?
      3. Are the projects designed to achieve these goals?
      4. Why am I asked to find and read an article using similar data?
      5. Should I expect a certain amount of frustration in carrying out these projects?
      6. Are the deadlines firm?

      7.  
  3. Getting and Reporting Econometric Results Based on Survey Data

  4.  
    1. Example: The Economics of Abuse

    2.  
    3. Survey Data isn't always easy to work with
      1. Sometimes you have to worry about sampling weights
      2. DLI files contain many masked or censored variables
      3. Even when not masked or censored, pay attention to missing observations
      4. Most survey questions have qualitative and categorical answers

      5.  
    4. "Section II: The Data"
    5. See the annotations in Bowlus and Seitz paper for more details.
      1. Tell your reader about your data sources.
      2. Tell your reader how you selected observations and manipulated the data
      3. Show your reader tables of summary statistics and in the text briefly discuss the patterns in the tables
      4. Show your reader some informative graphs
      5. Place supporting material or information that is difficult to display in prose into a data appendix
         
  5. A Quick Review of Multiple Linear Regression (MLRM)

  6. Based on: Multiple Regression Notes
    Supplementary material:  Mike Abbott's Econ 351 notes (and the corresponding sections of Gujarati)
     
    1. Ordinary Least Squares (OLS) is perfect for the MLRM
      1. The MLRM is defined by about seven assumptions, A0-A7
      2. The OLS estimator is easy to derive and it's formula only depends on A0
      3. The statistical properties of OLS depend crucially on A0-A7
      4. Using OLS and all the assumptions of MLRM you can carry out hypothesis tests, construct confidence intervals, and make predictions
      5. Doing all this is a breeze in Stata

      6.  
    2. "Section III: Empirical Results"
      1. Show your reader well-constructed tables of regressions
      2. Briefly discuss or intepret most of the regression coefficients
      3. Bury unimportant coefficients or models in appendices, footnotes, or even deeper!

      4.  
    1. When the assumptions of the MLRM don't hold neither do the nice properties of OLS
      1. It's not hard to correct for violations of constant variance (A6) using standard errors robust to heteroscedasticity
      2. And you already spent six weeks worrying about serial correlation (~A5), so we'll skip that
      3. It's tougher to deal with endogenous or error-correlated regressors (~A3), and we must skip this problem due to time constraints.
      4. So we will focus on the linear assumption in the MLRM.

      5.  
  7. An Even Quicker Introduction to Maximum Likelihood Estimation

  8.  
    1. Maximum Likelihood starts with a model of the whole population distribution that depends on some unknown population (or model) parameters
    2. The sample probability function treats data as variable and model parameters as fixed.  The likelihood function is really just the probability function, but it treats model parameters as variable and data as fixed

    3.  
    4. Finding parameter values that maximize the likelihood function generally leads to statistically consistent and asymptotically normal estimates of the population parameters.

    5.  
    6. Under A0-A7 the ML estimates of the NMLRM are the same as the OLS estimates!

    7.  
    8. Let's see that using Stata's maxlik command ...

    9.  
  9.  Limited Dependent Variables (LDVs)

  10.  
    1. Introduction
      1. There are lots of Kinds and Examples of LDVs (Binary, Censored, Multinomial, and Truncated Outcomes)
      2. Economic Models of Individual Choices Often Lead to LDV StatisticalModels

      3.  
    2. Binary Outcomes are the simplest type of LDV
      1. The Linear Probability Model (LPM) just means running OLS on a binary variable
      2. Logit and Probit are non-linear models for binary outcomes estimated using ML.
        1. Interpretiting coefficients in these models are little trickier than in the MLRM
        2. MLE of  logit/probit is easy in Stata
        3. You can do hypothesis testing and prediction
      3. Present results of logit/probit models pretty much like OLS, with a few differences

      4.  
    3. Multinomial Outcomes require more structure than binary outcomes
      1. If you want to model letter marks (A,B,C,etc), use the Ordered Logit/Probit Model
      2. If you want to model something like which way people get to work use the Multinomial Logit/Probit Model
      3. MLE works more or less the same as before.