Week 6 - Multiple Linear Regression
Overview
Let’s model!
Now, we can build powerful models with heaps of dependent variables. Want to predict wages? Let’s control for education, for experience, for gender, for age, for age squared (yes!). YES. Only our degrees of freedom can hold us back.
Reading Guide
Chapter 6: Linear Regression with Multiple Regressors
SW 6.1 Omitted Variable Bias
A discussion that connects nicely with our previous discussion of the zero conditional mean discussion and causal inference.
SW 6.2 The Multiple Regression Model
Hooray!
SW 6.3 The OLS Estimator in Multiple Regression
This section doesn’t get into derivation, and neither do we!
SW 6.4 Measures of Fit in Multiple Regression
The only new thing here is a revised \(SER\) forumla and the introduction of the Adjusted \(R^2\). Note that the lecture video also discusses the root mean standard error, \(RMSE\), which is a lot like the \(SER\) except that it uses \(n\) rather than degrees of freedom as a denominator.
SW 6.5 The Least Squares Assumptions in Multiple Regression
Take the three from univariate regression and add … no multicollinearity. Sorted.
SW 6.6 Distribution of the OLS Estimators in Multiple Regression
Just the intuition, don’t worry about the appendix.
SW 6.7 Multicollinearity
Make sure you understand the examples, but remember that in practice, any statistical package will fix perfect multicollinearity on its own. Imperfect multicollinearity, on the other hand, is something to think about when crafting your models.
SW 6.8 Conclusion
Treat yourself.
Slides
In-class exercises
Download PDF here, which contains regression output for questions (2) - (4)
Exercises
Consider a dataset on earnings in the United States. We are interested in the returns to education - how much an extra year of schooling “buys” you in terms of weekly wages (...as of 1980). You’re also worried about whether one’s education suffers from omitted variable bias.
You estimate two equations: \[\begin{aligned} \widehat{wage} &= 146.95 + 60.21educ\\ \widehat{educ} & = 5.84 + 0.075IQ\end{aligned}\]
Based on these results, is 60.21 an overestimate or underestimate of the returns to education? How do you know?
You estimate another equation: \(\widehat{wage} = -128.89 +42.06 educ + 5.14 IQ\)
What is the interpretation of the coefficient on \(educ\)? What is the interpretation of the constant?
Now, you control for experience and age and estimate the following population regression model:
\[wage_i = \beta_0 + \beta_1 educ_i + \beta_2 IQ_i + \beta_3 exper_i + \beta_4 age_i + \beta_5 age_i^2 + u_i\]
A one-year increase in age is associated with what change in wages? (mind the squared term)
Finally, because you are worried about omitted variable bias, you include father’s and mother’s education.
Why might parent’s education might directly affect wages?
Which other independent variables do you think parent’s education might affect? Explain.
How did controlling for parent’s education affect the returns to education? The returns to IQ?
Other resources
As requested, slower graphs! Also added a graph on collider bias, the webpage explanation helps there.
— Nick HK (@nickchk) November 29, 2018
These graphs are intended to show what standard causal inference methods actually do to data, and how they work.
This is what controlling for a binary variable looks like: pic.twitter.com/dTZxqY5JxA