Information about Exam
Reference: Wooldridge
The goal: We want to test our ideas about how the world works using data.
Econometrics is a toolbox that helps us to do this.
We usually want to know how one particular factor (X) influences one particular outcome (Y).
What is the causal relationship between X and Y? For example, we could ask about the causal relationship between:
We believe that there is some process – the data generating process (DGP) - that explains how the outcome we observe is generated.
We describe this DGP mathematically with an econometric model.
When writing down the econometric model we therefore have to think about how the functional form (exponential, quadratic, etc.) of the relationship between Y and our Xs.
Simple Regression model: $$ y=\beta_0 + \beta_1 x + u $$ (mention only one factor, x, explicitly)
Multiple Regression model: $$ y=\beta_0 + \beta_1 x_1 + \beta_2 x_2 + u $$ (mention more than one factor, here $x_1$ and $x_2$, explicitly)
Why? Because $\beta_j$ describes the causal relationship between $x_j$ and $y$. This is important for good decision making.
We use data to do this. In a way Economists can also be considered Data Detectives.
Collecting and analysing data.
The most popular method for analysing data is OLS (others are FD, IV).
In some very special circumstances (see OLS assumptions) OLS gives us good estimates of $\beta_j$.
Greek letter = parameter (unobserved and unknown)
Greek letter with hat = estimator of parameter (what you see in Stata)
$\hat{\beta}_j$ is not the same thing as $\beta$.
Formal definition: $\hat{\beta}_j$ is unbiased if $E[\hat{\beta}_j] = \beta_j$.
Intuition:
An estimator is consistent if as n (the sample size) tends to infinity, the distribution of $\hat{\beta}_j$ collapses to the single point $\beta_j$.
In other words: consistency means that the estimator is correct if we have an infinetley large sample.
Under assumptions MLR.1-4, $\hat{\beta}_j$ is a consistent estimator of $\beta_j$.
Now we know when $\hat{\beta}_j$ is correct on average in many samples (i.e., unbiased), but we typically only have one sample.
Can we ever know the exact value of $\beta_j$? No.
But, under some circumstances we can draw conclusions about the value of $\beta_j$. For example:
We specify the null and alternative hypothesis. Typically:
Null hypothesis
Note: Sometime the null hypothesis is that $\beta_j$ is equal to a particular value, and sometimes the alternative hypothesis is one sided.
Assumption MLR.5: Homoskedasticity: The error u has the same variance given any value of the explanatory variables. $$Var(u│x_1, x_2,...,x_k) = \sigma^2$$
Under assumptions MLR.1-5, OLS estimator is the Best Linear Unbiased Estimator (BLUE).
(MLR.1-5 are collectively known as the Gauss-Markov assumptions)
If MLR.5 is violated we have Heteroskedasticity.
Note: The Normality assumption is only necessary in small samples.
Answer: It implies that the OLS estimator $\hat{\beta}_j$ follows a normal distribution too.
And when $\hat{\beta}_j$ is normally distributed, our t-statistic is t-distributed.
If our t-statistic is t-distributed we can know the probability of getting the t-statistic that we have if there is no effect (p-values).
The ratio of $\hat{\beta}_j-\beta_j$ to the standard deviation follows a standard normal distribution.
The ratio of $\hat{\beta}_j-\beta_j$ to the standard error (called t-statistic) follows a t-distribution.
Our starting point ($H_0$) is usually to assume that the true effect is zero but our estimator is different from zero by chance.
However, if the true effect is zero, large and small t-statistics are unlikely.
When the t-statistic is large/small enough, we conclude that the true effect is probably not zero and thus reject $H_0$.
When is a t-statistic large/small enough? This depends:
In the past we would have looked up a critical value, c, in a statistical table and compared it to our t-statistic.
Today, Stata does this for us automatically.
Look at p-value:
Compare p-value to significance level.
If p-value $\leq$ significance level $\Rightarrow$ reject $H_0$.
If p-value $>$ significance level $\Rightarrow$ don't reject $H_0$ (Note: we never accept $H_0$)
Or, use the rule of thumb for two sided alternatives:
Given a 5% significance level and $df>60$, a t-value lower than -2 or higher than 2 implies statistical significance.
Or check if 95 Confidence Interval includes 0.
We can use OLS for estimation and hypothesis testing, and all is fine and dandy: OLS is then a very “good” estimator (unbiased, low variance) and it’s easy to test hypotheses.
We can try to fix it.
Or we have to interpret our results differently.
Let’s have a closer look at the MLR assumptions.
Include more relevant variables
Use a different method
Diff-in-Diff estimator of effect of incinerator: $\Delta 1981 - \Delta 1978 = -\$11,863$
Parallel trends assumption: The outcome of the treatment and control group would have followed a parallel trend in the absence of the treatment.
Possible if we have panel data.
We relate changes in X to changes in Y.
Allows us to difference away time constant part of the error term (remember the dog and the scale).
Possible to get a causal effect if $x$ is endogenous.
All we need is an instrumental variable ($z$) that fulfills two assumptions:
instrument relevance: $Cov(z,x) \neq 0$
IV estimates are Local Average Treatment Effects (LATEs)
Monotonicity assumption:
If the monotonicity assumption holds, IV estimates the average treatment effect on the treated (ATE)
Unless you have an experiment, for most research questions it is very difficult to get an unbiased estimator.
But, OLS estimates are still often very interesting even if you can’t interpret them causally.
Final Test: 60 minutes. See precise details in Blackboard. (Specific time, focus on last parts of course.)
Final MCQ test: 50 minutes. See precise details in Blackboard. ('Any time', covers whole course.)
Topic 10 arguing with data (last lecture) will not be assessed.
These questions will be very similar or even identical to questions from tutorial exercises.
They are supposed to be easy points for students who study!
Be precise and to the point.
Avoid irrelevant and wrong explanations.
Read the instructions and questions carefully.
You might not have time to answer all questions $\Rightarrow$ Answer easy questions first.
Try to be to the point and answer only what the questions ask. For most questions this will be a one word or one sentence answer.
Example: How many coefficients in this model (excluding the constant) are significant at the 20% level? [1 mark]
Question: How many coefficients in this model (excluding the constant) are significant at the 20% level? [2 mark]
Student answer: 4, because the R-squared is 0.4
Comment: 4 is correct, but the explanation is wrong. This wrong and irrelevant addition would lead to reduction of marks.
I will try to mark important words, but still read the questions carefully.
Example:
I have a document with exam answers. [More accurately, they are built into blackboard.]
The answers on this document are very short: often one number or a few words, never longer than 2 sentences.
No marks will be given for workings.
You will be able to see what you answered, as well as correct answer, for most of the questions. But not until a day or two after all students have finished.
Use the discussion board and help each other out.
I have little time next weeks, but can tune in to the discussion from time to time.
I enjoyed this class. I hope you have learned something that you remember after the exam.
Special thanks to those who turned up in person. Made teaching much nicer for me.
Any (positive or negative) feedback on the course? Great! Talk to me or send me an email.
Good luck with the exam!