No Slide Title

Download Report

Transcript No Slide Title

Econ 140
Classical Regression
Lecture 8
Lecture 9
1
Today’s Plan
Econ 140
• For the next few lectures we’ll be talking about the
classical regression model
– Looking at both estimators for a and b
– Inferences on what a and b actually tell us
• Today: how to operationalize the model
– Looking at BLUE for bi-variate model
– Inference and hypothesis tests using the t, F, and c2
– Examples of linear regressions using Excel
Lecture 9
2
Estimating coefficients
Econ 140
• Our model: Y = a + bX + e
• Two things to keep in mind about this model:
1) It is linear in both variables and parameters
• Examples of non-linearity in variables:
Y = a + bX2 or Y = a + bex
• Example of non-linearity in parameters:
Y = a + b2X
• OLS can cope with non-linearity in variables but not
in parameters
Lecture 9
3
Estimating coefficients (3)
Econ 140
2) Notation: we’re not estimating a and b anymore
• We are estimating coefficients which are estimates
of the parameters of a and b
• We will denote the coefficients as  or aˆ and  or bˆ
• We are dealing with a sample size of n
– For each sample we will get a different  and  pair
Lecture 9
4
Estimating coefficients (4)
Econ 140
• In the same way that you can take a sample to get an
estimate of µy you can take a sample to get an estimate of
the regression line, of  and 
Y
Sample 1
Regression
(Population)
Line
Sample 2
Lecture 9
X
5
The independent variable
Econ 140
• We also have a given variable X, its values are known
– This is called the independent variable
• Again, the expectation of Y given X is
E(Y | X) = a + bX
• With constant variance
V(Y) = 2
Lecture 9
6
A graph of the model
Econ 140
Y
(Y1, X1)
Yˆ
e  Y  Yˆ
Y
Yˆ    X
X
Lecture 9
7
What does the error term do?
Econ 140
• The error term e  Y  Yˆ gives us the test statistics and
tells us how well the model Y = a+bX+e fits the data
• The error term represents:
1) Given that Y is a random variable, e is also random,
since e is a function of Y
2) Variables not included in the model
3) Random behavior of people
4) Measurement error
5) Enables a model to remain parsimonious - you don’t
want all possible variables in the model if some have
little or no influence
Lecture 9
8
Rewriting beta
Econ 140
• Our complete model is Y = a + bX + e
• We will never know the true value of the error e so we will
estimate the following equation:
Y    X  
• For our known values of X we have estimates of , , and 
• So how do we know that our OLS estimators give us the
BLUE estimate?
– To determine this we want to know the expected value
of  as an estimator of b, which is the population
parameter
Lecture 9
9
Rewriting beta(2)
Econ 140
• To operationalize, we want to think of what we know
• We know from lecture 2 that there should be no correlation
between the errors and the independent variables
  i  0, E  i   0
  i X i  0, E  i X i   0

2
• We also know V ( X )   2 V (Y )  

• Now we have that E(Y|X) = a + bX + E(|X)
• The variance of Y given X is V(Y) = 2 so V(|X)= 2
Lecture 9
10
Rewriting beta(3)
Econ 140
• Rewriting 
– In lecture 2 we found the following estimator for 
XY  nXY


2
2
X

n
X

• Using some definitions we can show:
E() = b
2
V ( ) 
2
x

Lecture 9
11
Rewriting beta (4)
Econ 140
• We have definitions that we can use:
yi  (Yi  Y )
2
2
xi  ( X i  X ) So that  xi   ( X i  X )
• Using the definitions for yi and xi we can rewrite  as
xy

 x2
• We can also write  xy 
Lecture 9

xY
12
Rewriting beta (5)
   ciYi
• We can rewrite  as
Econ 140
where ci 
xi
2
x
 i
• The properties of ci :
 ci  0,  ( X i  X )  0,
2.  ci xi  1
1.
Lecture 9
3.
2 1
c
 i
4.
 ci X i  1 
2
x
 i
 xi X i
 xi 2
13
Showing unbiasedness
Econ 140
• What do we know about the expected value of beta?
   ciYi
E (  )   ci E (Yi )
• We can rewrite this as E (  )   ci (a  bX i )
• Multiplying the brackets out we get:
E (  )   ci (a)   ci (bX i )
• Since b is constant,
E (  )   ci (a)  b ci X i
Lecture 9
14
Showing unbiasedness (2)
Econ 140
• Looking back at the properties for ci we know that
 ci  0,  ci X i  1
• Now we can write this as
E (  )  0  b(1)  b
• We can conclude that the expected value of  is b and that
 is an unbiased estimator of b
Lecture 9
15
Gauss Markov Theorem
Econ 140
• We can now ask: is  an efficient estimator?
• The variance of  is
2
2
V (  )   ci V (Y ) 
 xi2
Where    ciYi
• How do we know that OLS is the most efficient estimator?
– The Gauss-Markov Theorem
Lecture 9
16
Gauss Markov Theorem (2)
Econ 140
• Similar to our proof on the estimator for my.
• Suppose we use a new weight ci  ci  di
• We can take the expected value of E()
   ci E (Yi )
  ci (a  bX i )
 a( ci   di )  b( ci X i   di X i )
Lecture 9
17
Gauss Markov Theorem (3)
Econ 140
• We know that
E (  )  a (0   di )  b(1   di X i )
 a  di  b  b di X i
– For  to be unbiased, the following must be true:
 di  0
 di X i  0
Lecture 9
18
Gauss Markov Theorem (4)
Econ 140
• Efficiency (best)?
• We have    ciYi where ci  ci  di
• Therefore the variance of this new  is
•
2
2
V (  )   ci V (Y ) 
2+
 xi
Sdi2 +2Scidi
• If each di  0 such that ci c’i then
2
d
 i 0
• So when we use weights c’I we have an inefficient
estimator
Lecture 9
19
Gauss Markov Theorem (5)
Econ 140
• We can conclude that
•
Lecture 9
XY  nXY


 X 2  nX 2
is BLUE
20
Wrap up
Econ 140
• What did we cover today?
• Introduced the classical linear regression model (CLRM)
• Assumptions under the CLRM
1) Xi is nonrandom (it’s given)
2) E(ei) = E(ei|Xi) = 0
3) V(ei)= V(ei|Xi) = 2
4) Covariance (eiej) = 0
• Talked about estimating coefficients
• Defined the properties of the error term
• Proof by contradiction for the Gauss Markov Theorem
Lecture 9
21