Econometrics I Ch2
Econometrics I Ch2
Econometrics I Ch2
1
Lecture Plan
2
#. What are the concept regression Analysis ?
14
2.1.1 Definition
4
Definition
9
When do we apply regression analysis?
10
When do we apply regression analysis?...
11
What are some of the considerations we should
make when choosing statistical analysis?
12
What are some of the considerations or
assumptions that lead to a choice of analysis?
14
What are the different types of
regression models?
15
I. Simple regression model
16
I. Simple regression model...
EXAMPLES:
Qd= f(P) Demand function
QS = f(P) Supply function
17
II. Multiple regression model
If k=1, that is, there is only one X-variable, we have the simple
regression
18
II. Multiple regression model...
19
III. What is multivariate regression model?
20
III. What is multivariate regression model?..
21
2.1.4.Variables in regression models
22
Explain the variables involved in a
regression model?
Yi = the ith value of the dependent variable. Y i is also called the dependent variable or the
regressand or the explained variable
Xi = the ith value of the independent variable. Xi is also called the independent
variable or the regressor or the explanatory variable
23
Explain the variables involved in a regression model?...
2. Unobservable variables
The εi is the random error term for the ith member of the
population
24
Explain the variables involved in a
regression model?...
25
What is the significance of the stochastic term?
26
What are the justifications for the inclusion of
the disturbance term in a regression model?
27
What are the justifications for the inclusion of
the disturbance term in a regression model?...
28
What are the justifications for the inclusion of
the disturbance term in a regression model?...
o Deviations of the points from the line may be due to errors of measurement
of the variables, due to the methods of collection, processing statistical
information, etc.
o The variables included in the model may be measured inaccurately & the
stochastic term is expected to account for these errors.
o Problems that arise due to the methods of data collection, processing statistical
information, etc. can be captured by the error term.
29
What are the justifications for the inclusion of
the disturbance term in a regression model?...
31
Assumptions of the Classical Linear Stochastic Regression
Model
5
Assumptions
5
Assumption
Proof:
Mean: (Y ) xi u i
X i Since (u i ) 0
X i u i ( X i )
2
(ui ) 2
2
(since (ui ) 2 2 )
var( Yi ) 2
……………………………………….(
5
Assumptions
5
Assumptions
5
Assumption
Cov(Ui,Uj ) E (ui u j ) 0
5
Assumption…
5
39
Assumption…
5
Assumption…
then it is assumed that they are not perfectly correlated with each
other.
5
Method of Simple linear regression Estimation
5
Method of Simple linear regression Estimation
o Specifying the model and stating its underlying assumptions are
the first stage of any econometric application.
o The next step is the estimation of the numerical values of the
parameters of economic relationships.
o The parameters of the simple linear regression model can be
estimated by various methods. Three of the most commonly used
methods are:
1. Ordinary least square method (OLS)
2. Maximum likelihood method (MLM)
3. Method of moments (MM)
5
1. Ordinary least square method (OLS)
o The model is called the true relationship between Y and X
because Y and X represent their respective population value are
called the true parameters since they are estimated from the
population value of Y and X.
o But, it is difficult to obtain the population value of Y and X
because of technical or economic reasons. So we are forced to
take the sample value of Y and X.
o The parameters estimated from the sample value of Y and X are
called the estimators of the true parameter and are symbolized as
α and 𝛽 .
5
1. Ordinary least square method (OLS)
ˆ and ˆ are estimated from the sample of Y and X and ei represents the sample
counterpart of the population random disturbance U i .
Estimation of and by least square method (OLS) or classical least square
(CLS) involves finding values for the estimates ˆ and ˆ which will minimize the
o
sum of square of the squared residuals ( ei2 ).
5
Conti..
From the estimated relationship Yi ˆ ˆX i ei , we obtain:
e 2
i (Yi ˆ ˆX i ) 2 ………………………. (2.7)
To find the values of ˆ and ˆ that minimize this sum, we have to partially
differentiate e 2
i with respect to ˆ and ˆ and set the partial derivatives equal to
zero.
ei2
1. 2 (Yi ˆ ˆX i ) 0.......... .......... .......... .......... .......... .....( 2.8)
ˆ
Note: at this point that the term in the parenthesis in equation 2.8and 2.11 is the
residual, e Yi ˆ ˆX i . Hence it is possible to rewrite (2.8) and (2.11) as
Y X
i i ˆX i ˆX i2 ……………………………………….(2.13)
47
Conti..
Equation (2.9) and (2.13) are called the Normal Equations. Substituting the values
of ̂ from (2.10) to (2.13), we get:
Y Xi i X i (Y ˆX ) ˆX i2
Y Xi i Y X i ˆ (X i2 XX i )
XY nXY = ˆ ( X i2 nX 2)
XY nXY
ˆ ………………….(2.14)
X i2 nX 2
48
Conti…
Equation (2.14) can be rewritten in somewhat different way as follows;
( X X )(Y Y ) ( XY XY XY XY )
( X X ) 2 X 2 nX 2 (2.16)
49
Conti…
Now, denoting ( X i X ) as xi , and (Yi Y ) as yi we get;
x y
ˆ i 2 i ……………………………………… (2.17)
xi
50
Estimation of a function with zero intercept
Suppose it is desired to fit the line Yi X i U i , subject to the restriction 0.
Subject to: ˆ 0
51
Conti…
The composite function then becomes
Z (Yi ˆ ˆX i ) 2 ˆ , where is a Lagrange multiplier.
z
2 0 (iii)
52
Conti…
Substituting (iii) in (ii) and rearranging we obtain:
X i (Yi ˆX i ) 0
Yi X i ˆX i 0
2
X Y
ˆ i 2 i ……………………………………..(2.18)
X i
This formula involves the actual values (observations) of the variables and not their
deviation forms, as in the case of unrestricted value of ˆ .
53
Statistical Properties of Least Square Estimators
54
Conti….
How are we to choose among the different econometric
methods, the one that gives „good‟ estimates?
55
Conti…
Closeness‟ of the estimate to the population parameter is
measured by the mean and variance or standard deviation of
the sampling distribution of the estimates of the different
econometric methods.
56
Cont..
We assume the usual process of repeated sampling i.e. we
assume that we get a very large number of samples each of
size „n‟.
57
Conti….
We next compare the mean (expected value) and the
variances of these distributions and we choose among the
alternative estimates the one whose distribution is
concentrated as close as possible around the population
parameter.
58
PROPERTIES OF OLS ESTIMATORS
59
Conti…
According to the this theorem, under the basic assumptions of the
Some times the theorem referred as the BLUE theorem i.e. Best,
Linear, Unbiased Estimator. An estimator is called BLUE if:
60
A)Linear:
a linear function of the a random variable, such as, the dependent
variable Y
̂ is linear in Y
where
61
B. Unbiasedness
Proposition: ˆ & ˆ are the unbiased estimators of the true parameters &
From your statistics course, you may recall that if ˆ is an estimator of then
E(ˆ) the amount of bias and if ˆ is the unbiased estimator of then bias =0 i.e.
E(ˆ) 0 E(ˆ)
In our case, ˆ & ˆ are estimators of the true parameters & .To show that they
are the unbiased estimators of their respective parameters means to prove that:
(ˆ ) and (ˆ )
62
C) Minimum variance:
63
Statistical test of Significance of the OLS Estimators (First Order tests)
After the estimation of the parameters and the determination of the least
square regression line, we need to know how „good‟ is the fit of this line
This knowledge is essential because the closer the observation to the line,
the better the goodness of fit, i.e. the better is the explanation of the
64
Cont.….
We divide the available criteria into three groups: the theoretical a priori
criteria, the statistical criteria, and the econometric criteria.
Under this section, our focus is on statistical criteria (first order tests). the two
most commonly used first order tests in econometric analysis are
i.e. R2). This test is used for judging the explanatory power of the
independent variable(s).
B. B) The standard error tests of the estimators. This test is used for judging
65
Tests of the ‘ Goodness of fit’ with R2
A. R2 shows the percentage of total variation of the dependent variable that
can be explained by the changes in the explanatory variable(s) included in
the model.
By fitting the line Yˆ ˆ 0 ˆ1 X we try to obtain the explanation of the variation of
the dependent variable Y produced by the changes of the explanatory variable X.
66
Conti…
.Y
Y = e Y Yˆ
Y Y = Yˆ Yˆ ˆ 0 ˆ1 X
= Yˆ Y
Y.
X
Figure „d‟. Actual and estimated values of the dependent variable Y 67
Conti……
As can be seen from fig.(d) above, 𝑌 − 𝑌 represents measures the
68
Conti…
In summary:
ei Yi Yˆ = deviation of the observation Yi from the regression line.
69
Conti…
Now, we may write the observed Y as the sum of the predicted value ( Yˆ ) and the
residual term (ei.).
Yi Yˆ ei
predicted Yi
Observed Yi Re sidual
From equation (2.34) we can have the above equation but in deviation form
y yˆ e . By squaring and summing both sides, we obtain the following expression:
70
Conti…
y 2 ( yˆ 2 e) 2
y 2 ( yˆ 2 ei2 2 yei)
yˆe 0 ………………………………………………(2.46)
71
Conti….
Therefore;
yi2
yˆ 2 ei2 ………………………………...(2.47)
Total Explained Un exp lained
var iation var iation var ation
OR,
Total sum of Explained sum Re sidual sum
square of square of square
TSS ESS RSS
i.e
TSS ESS RSS ……………………………………….(2.48)
72
Conti…
Mathematically; the explained variation as a percentage of the total variation is
explained as:
ESS yˆ 2
……………………………………….(2.49)
TSS y 2
From equation (2.37) we have yˆ ̂x . Squaring and summing both sides give us
yˆ 2 ˆ 2 x 2 (2.50)
We can substitute (2.50) in (2.49) and obtain:
ˆ 2 x 2
ESS / TSS …………………………………(2.51)
y 2
73
CONTI…
xy x x y
2 2
2 i2 , Since ˆ i 2 i
x y xi
xy xy
………………………………………(2.52)
x 2 y 2
74
Conti….
Comparing (2.52) and (2.54), we see exactly the expressions. Therefore:
xy xy
ESS/TSS = r2
x 2 y 2
The limit of R2: The value of R2 falls between zero and one. i.e. 0 R 2 1 .
75
Interpretation of R2
76
B) TESTING THE SIGNIFICANCE OF OLS PARAMETERS
To test the significance of the OLS parameter estimators we need the following:
Variance of the parameter estimators
Unbiased estimator of 2
The assumption of normality of the distribution of error term.
We have already derived that:
77
Conti….
ˆ 2
var( ˆ )
x 2
ˆ 2 X 2
var(ˆ )
nx 2
e 2 RSS
ˆ 2
n2 n2
For the purpose of estimation of the parameters the assumption of normality is not
used, but we use this assumption to test the significance of the parameter estimators;
because the testing methods or procedures are based on the assumption of the
normality assumption of the disturbance term.
78
Cont.…
Hence before we discuss on the various testing methods it is important to see
whether the parameters are normally distributed or not.
We have already assumed that the error term is normally distributed with mean zero
and variance 2 , i.e. U i ~ N ( 20) . , Similarly, we also proved that
2
1. ˆ ~ N ,
2
x
2 X 2
2. ˆ ~ N , 2
nx
79
Cont.…
To show whether ˆ and ˆ are normally distributed or not, we need to make use of
one property of normal distribution. “........ any linear function of a normally
distributed variable is itself normally distributed.”
ˆ ki Yi k1Y1 k 2 Y2i .... k n Yn
2 2 X 2
ˆ ~ N , 2 ; ˆ ~ N , 2
x nx 80
Conti…
The OLS estimates ˆ and ˆ are obtained from a sample of observations on Y and
X. Since sampling errors are inevitable in all estimates, it is necessary to apply test
of significance in order to measure the size of the error and determine the degree of
confidence in order to measure the validity of these estimates. This can be done by
using various tests. The most common ones are:
i) Standard error test ii) Student’s t-test iii) Confidence interval
81
Conti…
All of these testing procedures reach on the same conclusion. Let us now see these
testing methods one by one.
i) Standard error test
This test helps us decide whether the estimates ˆ and ˆ are significantly different
from zero, i.e. whether the sample from which they have been estimated might have
come from a population whose true parameters are zero. 0 and / or 0 .
Formally we test the null hypothesis
H 0 : i 0 against the alternative hypothesis H 1 : i 0
SE(ˆ ) var(ˆ )
Second: compare the standard errors with the numerical values of ˆ and ˆ .
Decision rule:
If SE(ˆi ) 12 ˆi , accept the null hypothesis and reject the alternative
If SE(ˆi ) 12 ˆi , reject the null hypothesis and accept the alternative
Numerical example: Suppose that from a sample of size n=30, we estimate the
following supply function.
Q 120 0.6 p ei
SE : (1.7) (0.025)
84
Conti…
Test the significance of the slope parameter at 5% level of significance using the
standard error test.
SE( ˆ ) 0.025
( ˆ ) 0.6
1
2 ˆ 0.3
SE( ˆ ) var( ˆ )
SE(ˆ ) var(ˆ )
Second: compare the standard errors with the numerical values of ˆ and ˆ .
86
Conti…
Decision rule:
If SE(ˆi ) 12 ˆi , accept the null hypothesis and reject the alternative
If SE(ˆi ) 12 ˆi , reject the null hypothesis and accept the alternative
87
Conti…
The acceptance or rejection of the null hypothesis has definite economic meaning.
Namely, the acceptance of the null hypothesis 0 (the slope parameter is zero)
implies that the explanatory variable to which this estimate relates does not in fact
influence the dependent variable Y and should not be included in the function, since
the conducted test provided evidence that changes in X leave Y unaffected. In other
words acceptance of H0 implies that the relation ship between Y and X is in fact
Y (0) x , i.e. there is no relationship between X and Y.
Numerical example: Suppose that from a sample of size n=30, we estimate the
following supply function.
Q 120 0.6 p ei
SE : (1.7) (0.025)
Test the significance of the slope parameter at 5% level of significance using the
standard error test. 88
Conti…
SE ( ˆ ) 0.025
( ˆ ) 0.6
1
2 ˆ 0.3
( X X ) 2
sx
n 1
n sample size
We can derive the t-value of the OLS estimates
ˆi
t ˆ
SE( ˆ )
with n-k degree of freedom.
ˆ
tˆ
SE(ˆ )
90
Conti…
Where:
SE = is standard error
k = number of parameters in the model.
Since we have two parameters in simple linear regression with intercept different
from zero, our degree of freedom is n-2. Like the standard error test we formally
test the hypothesis: H 0 : i 0 against the alternative H 1 : i 0 for the slope
parameter; and H0 : 0 against the alternative H1 : 0 for the intercept.
ˆ 0 ˆ
t*
SE( ˆ ) SE( ˆ )
91
cont..
Step 2: Choose level of significance. Level of significance is the probability of
making „wrong‟ decision, i.e. the probability of rejecting the hypothesis when it is
actually true or the probability of committing a type I error. It is customary in
econometric research to choose the 5% or the 1% level of significance. This means
that in making our decision we allow (tolerate) five times out of a hundred to be
„wrong‟ i.e. reject the hypothesis when it is actually true.
92
Conti…
Step 3: Check whether there is one tail test or two tail test. If the inequality sign in
the alternative hypothesis is , then it implies a two tail test and divide the chosen
level of significance by two; decide the critical rejoin or critical value of t called tc.
But if the inequality sign is either > or < then it indicates one tail test and there is no
need to divide the chosen level of significance by two to obtain the critical value of
to from the t-table.
Example:
If we have H 0 : i 0
against: H1 : i 0
93
Cont..
Then this is a two tail test. If the level of significance is 5%, divide it by two to
obtain critical value of t from the t-table.
Step 4: Obtain critical value of t, called tc at 2 and n-2 degree of freedom for two
tail test.
Step 5: Compare t* (the computed value of t) and tc (critical value of t)
If t*> tc , reject H0 and accept H1. The conclusion is ˆ is statistically
significant.
If t*< tc , accept H0 and reject H1. The conclusion is ˆ is statistically
insignificant.
94
Conti..
Numerical Example:
Suppose that from a sample size n=20 we estimate the following consumption
function:
C 100 0.70 e
(75.5) (0.21)
The values in the brackets are standard errors. We want to test the null hypothesis:
H 0 : i 0 against the alternative H 1 : i 0 using the t-test at 5% level of
significance.
the t-value for the test statistic is
95
Conti…
ˆ 0 ˆ 0.70
t* = 3 .3
SE(ˆ ) SE(ˆ ) 0.21
96
III) Confidence interval
Rejection of the null hypothesis doesn‟t mean that our estimate ˆ and ˆ is the
correct estimate of the true population parameter and . It simply means that
our estimate comes from a sample drawn from a population whose parameter is
different from zero.
In order to define how close the estimate to the true parameter, we must construct
confidence interval for the true parameter, in other words we must establish limiting
values around the estimate with in which the true parameter is expected to lie within
a certain “degree of confidence”. In this respect we say that with a given probability
the population parameter will be with in the defined confidence interval (confidence
limits). 97
Conti…
We choose a probability in advance and refer to it as confidence level (interval
coefficient). It is customarily in econometrics to choose the 95% confidence level.
This means that in repeated sampling the confidence limits, computed from the
sample, would include the true population parameter in 95% of the cases. In the
other 5% of the cases the population parameter will fall outside the confidence
interval.
In a two-tail test at level of significance, the probability of obtaining the specific
t-value either –tc or tc is 2 at n-2 degree of freedom. The probability of obtaining
98
Cont..
ˆ
any value of t which is equal to at n-2 degree of freedom is
SE ( ˆ )
1
2
2 i.e. 1 .
i.e. Pr t c t* t c 1 …………………………………………(2.57)
ˆ
but t* …………………………………………………….(2.58)
SE( ˆ )
Pr SE( ˆ )t c ˆ SE( ˆ )t c 1 by multiplying SE( ˆ )
Pr ˆ SE( ˆ )t c ˆ SE( ˆ )t c 1 by subtracting ˆ
H1 : 0
Decision rule: If the hypothesized value of in the null hypothesis is within the
is outside the limit, reject H0 and accept H1. This indicates ˆ is statistically
significant. 100
Cont.…
Numerical Example:
Suppose we have estimated the following regression line from a sample of 20
observations.
Y 128.5 2.88X e
(38.2) (0.85)
101
Cont.…
Solution:
a. The limit within which the true lies at 95% confidence interval is:
ˆ SE( ˆ )t c
ˆ 2.88
SE(ˆ ) 0.85
t c at 0.025 level of significance and 18 degree of freedom is 2.10.
The results of the regression analysis derived are reported in conventional formats.
It is not sufficient merely to report the estimates of ‟s. In practice we report
regression coefficients together with their standard errors and the value of R2. It
has become customary to present the estimated equations with standard errors
placed in parenthesis below the estimated parameter values. Sometimes, the
103
Conti…
placed in parenthesis below the estimated parameter values. Sometimes, the
estimated coefficients, the corresponding standard errors, the p-values, and some
other indicators are presented in tabular form.
These results are supplemented by R2 on ( to the right side of the regression
equation).
104
# END!!!!!!!
14