0% found this document useful (0 votes)
5 views18 pages

Unit-4

Copyright
© © All Rights Reserved
Available Formats
Download as PDF, TXT or read online on Scribd
Download as pdf or txt
0% found this document useful (0 votes)
5 views18 pages

Unit-4

Copyright
© © All Rights Reserved
Available Formats
Download as PDF, TXT or read online on Scribd
Download as pdf or txt
Download as pdf or txt
You are on page 1/ 18

UNIT 4 SIMPLE LINEAR REGRESSION

MODEL: ESTIMATION
Structure
4.0 Objectives
4.1 Linear Regression Model
4.2 Population Regression Function (PRF)
4.2.1 Deterministic Component

4.2.2 Stochastic Component

4.3 Sample Regression Function (SRF)


4.4 Assumptions of Classical Regression Model
4.5 Ordinary Least Squares Method of Estimation
4.6 Algebraic Properties of OLS Estimators
4.7 Coefficient of Determination
4.7.1 Formula of Computing R2

4.7.2 F-Statistic for Goodness of Fit

4.7.3 Relationship between F and R2

4.7.4 Relationship between F and t2

4.8 Let Us Sum Up


4.9 Answers/ Hints to Check Your Progress Exercises

4.0 OBJECTIVES
After going through this unit, you should be able to
 describe the classical linear regression model;
 differentiate between Population Regression Function (PRF) and Sample
Regression Function (SRF);
 find out the Ordinary Least Squares (OLS) estimators;
 describe the properties of OLS estimators;
 explain the concept of goodness of fit of regression equation; and
 describe the coefficient of determination and its properties.


Dr. Pooja Sharma, Assistant Professor, Daulat Ram College, University of Delhi
Regression Model: Two
Variables Case 4.1 INTRODUCTION
In Unit 5 of the course BECC 107: Statistical methods for Economics we
discussed the topics correlation and regression. In that Unit we gave a brief idea
about the concept of regression. You already know that there are two types of
variables in regression analysis: i) dependent (or explained) variable, and ii)
independent (or explanatory) variable. As the name (explained and explanatory)
suggests the dependent variable is explained by the independent variable.
Usually we denote the dependent variable as Y and the independent variable as
X. Suppose we took up a household survey and collected n pairs of observations
in X and Y. The relationship between X and Y can take many forms. The general
practice is to express the relationship in terms of some mathematical equation.
The simplest of these equations is the linear equation. It means that the
relationship between X and Y is in the form of a straight line, and therefore, it is
called linear regression. When the equation represents curves (not a straight line)
the regression is called non-linear or curvilinear.
Thus in general terms we can express the relationship between X and Y as
follows in equation (4.1).
𝑌 = 𝑓(𝑋) … (4.1)
In this block (Units 4, 5 and 6) we will consider simple linear regression models
with two variables only. The multiple regression model comprising more than
one explanatory variable will be discussed in the next block.
Regression analysis may have the following objectives:
 To estimate the mean or average value of the dependent variable, given the
values of the independent variables.
 To test the hypotheses regarding the underlying economic theory. For
example, one may test the hypotheses that the price elasticity of demand is
(–)1 that is, the demand is perfectly elastic, assuming other factors affecting
the demand are held constant.
 To predict the mean value of the dependent variable given the values of the
independent variable.

4.2 POPULATION REGRESSION FUNCTION


A population regression function hypothesizes a theoretical relationship between
a dependent variable and a set of independent or explanatory variables. It is a
linear function. The function defines how the conditional expectation of a
variable Y responds to the changes in independent variable X.
𝑌 = 𝐸 (𝑌 |𝑋 ) + 𝑢 … (4.2)
The function consists of a deterministic component 𝐸 (𝑌|𝑋) and a non-
deterministic or ‘stochastic’ component 𝑢, as depicted in equation (4.2).
44
We are concerned about examining the determinants of dependent variable (Y) Simple Linear Regression
Model: Estimation
conditional upon the given values of impendent variables (X).
4.2.1 Deterministic Component
The conditional expectation of Y constitutes the deterministic component of the
regression model. It is obtained in the form of a deterministic line. It is also
known as the Population Regression Line (PRL). The non-deterministic or
stochastic component is represented by a random error term, denoted by 𝑢 .
Let us take an example. Suppose we want to examine the impact of weekly
personal disposable income (PDI) on the weekly expenditure for a set of
population, then we consider weekly PDI as the independent variable (Y) and
weekly expenditure as the dependent variable (X). For each given value of
weekly PDI, the average value of weekly expenditure is plotted on the vertical
axis. People with higher income are likely to spend more, therefore intuitively,
the relationship between weekly PDI and weekly expenditure is positive. Thus
the following Population Regression Line is obtained and plotted on a graph as
explained below.
𝐸(𝑌 |𝑋 ) = 𝛽 + 𝛽 𝑋 … (4.3)
Note that in equation (4.3), 𝛽 and 𝛽 are the parameters. Here 𝛽 is the intercept
of the population regression function. It indicates the expected value of the
dependent variable when the explanatory variable is zero. Further, 𝛽 is the slope
of the population regression function. It indicates the magnitude by which the
dependent variable will change if there is a one unit change in the independent
variable. The population parameters describe the relationship between the
dependent variable and the independent variable in the population.

Weekly Exp.

0 Weekly Personal Disposable Income

Fig. 4.1: Weekly Personal Disposable Income


45
Regression Model: Two
Variables Case
Look into the circled points in Fig. 4.1. These points represent the mean or the
average value of Y corresponding to various 𝑋 . They are called the conditional
means or conditional expectation values. If we connect the various expected
values of Y, the resulting lines is called the Population Regression Line (PRL).
4.2.2 Stochastic Component
When we collect data from a sample, we do not a deterministic relationship
between X and Y. For example, for the same level of income the expenditure of
two persons could be different. Suppose there are two persons with monthly
income of Rs. 20000 per month. While the monthly expenditure of one person is
Rs. 15000, that of the other person could be Rs. 19000. The differences in
monthly expenditure for the second person could be higher due to his health
condition or living style. Such differences in the dependent variable are captured
by the stochastic error term. In Fig. 4.1, for a particular value of X, the value of
the Y variable is depicted by a vertical dotted line. The expected value of Y for a
particular value of X is circled (see Fig. 4.1).
Thus, there is a need to specify the stochastic relationship between X and Y. The
specification of the sample regression function (SRF) is
𝑌 =𝛽 +𝛽 𝑋 +𝑢 … (4.5)
In equation (4.5) the term 𝑢 is called stochastic error or random error.
The first component of equation (4.5) is the deterministic component (𝛽 +
𝛽2𝑋𝑖, which we have already discussed. The deterministic component is the
mean or average expenditure in the example under consideration. The
deterministic component is also called the systematic or deterministic
component.
The second component 𝑢 is called the random component (determined non-
systematically by factors other than income). The error term 𝑢 is also known as
the ‘noise component’. The error term 𝑢 is a random variable. The value of 𝑢
cannot be controlled or known.
There are three reasons for including the error term 𝑢 in a regression model: (i)
The error term represents the influence of those variables that are not explicitly
introduced in the regression model. For example, there are several variables that
influence consumption expenditure of a household (such as number of family
members, health status, neighbourhood, etc.). These variables affect the
dependent variable, and there exists intrinsic randomness between X and Y. (ii)
Human behaviour is not predictable. This sort of randomness is reflected and
captured by the random error term. (iii) The errors in measuring data such as
rounding off of annual family income, absence of many students from the school,
etc.
Because of the randomness the actual value of the data would either remain
above or below the expected value of the dependent variable. In other words, the
46 actual value will deviate from the average, that is, the systematic component.
Having understood the elementary concept of Population Regression Function Simple Linear Regression
Model: Estimation
and Population Regression Line (PRL), the following section describes the
estimation of PRL using the sample.
Check Your Progress 1
1) What are the objectives of estimating regression models?

.......................................................................................................................
.......................................................................................................................
.......................................................................................................................
.......................................................................................................................
.......................................................................................................................
.......................................................................................................................

2) Why does the average value of the dependent variable differ from the
actual value?
.......................................................................................................................
.......................................................................................................................
.......................................................................................................................
.......................................................................................................................
.......................................................................................................................

3) Why do we include an error term (𝒖𝒊 ) to the regression model?


.......................................................................................................................
.......................................................................................................................
.......................................................................................................................
.......................................................................................................................
.......................................................................................................................

4.3 SAMPLE REGRESSION FUNCTION


We rarely have the data related to the entire population at our disposal. We only
have a sample from the population. Thus, we need to use the sample to estimate
the population parameters. We may not be able to find out the population
regression line (PRL) because of sampling fluctuations or sampling error.
Suppose we have two samples from the given population. Using the samples
separately, we obtain Sample Regression Lines (SRLs). A sample represents the
population. In Fig. 4.2 we have shown two sample regression lines, SRL1 and
SRL2.

47
Regression Model: Two
Variables Case

SRL1

Expenditure
SRL2

0
PDI

Fig: 4.2: Two Sample Regression Lines


Both the sample regression lines represent the population regression line.
However, due to sampling fluctuation, the slope and intercept of both the SRLs
are different. Analogous to population regression function (PRF) that underlies
the PRL, we develop the concept of Sample Regression Function (SRF)
comprising Sample Regression Line (SRL) and the error term 𝑢 .

Exp.

SRL: 𝑌 = 𝑏 + 𝑏 𝑋
𝑌
𝑌 PRL: 𝐸(𝑌|𝑋 ) = 𝛽 + 𝛽 𝑋

𝑒
𝑢
𝑢
𝑒 𝑌

0
𝑋 𝑋 PDI

Fig 4.3: Population Regression Line and Sample Regression Line


48
In Fig. 4.3 we depict the population regression line (PRL) and the sample Simple Linear Regression
Model: Estimation
regression line (SRL). We observe that the slopes of both the lines are different.
Thus, 𝑏 ≠ 𝛽 and 𝑏 ≠ 𝛽 . Let us consider a particular value of the explanatory
variable, 𝑋 . The corresponding value of the explained variable is 𝑌 . On the
basis of the sample regression line we obtain estimated value of the explained
variable, 𝑌 . Now let us find out the distinction between the error term (u) and the
residual (e). The distance between the actual value 𝑌 and the corresponding
point on the population regression line is 𝑢 . This error 𝑢 is not known to us,
because we do not know the values of 𝛽 and 𝛽 . What we know is 𝑌 , which is
estimated on the basis of 𝑏 and 𝑏 . The distance between 𝑌 and 𝑌 is the
residual, 𝑒 .
The population regression line as given in equation (4.2) is
𝑌 = 𝐸(𝑌 |𝑋 ) + 𝑢
The sample regression line that we estimate is given by
𝑌 =𝑏 +𝑏 𝑋 … (4.6)
In equation (4.6) the symbol (^) is read as ‘hat’ or ‘cap’. Thus, 𝑌 is read as ‘𝑌 -
hat’.
You should remember that what we observe are proxies b1 , b 2 and e in place of
𝛽 , 𝛽 and 𝑢 .
𝑌 =𝑌 +𝑒 =𝑏 +𝑏 𝑋 +𝑒 ... (4.7)
where Yˆi = estimator of E(Y|Xi), the estimator of the population conditional
mean 𝑌 is an estimator (or a sample statistic) in equation (4.7). A particular value
obtained by the estimator is considered an estimate.
The actual value of Y is obtained by adding the residual term to the estimated
value of Y, also referred as the residual. The residual is the estimated value of
random error term of the population regression function. The sample regression
function in equation (4.7) is combination of sample regression line given by Yˆi
and the estimated residual term ei. The dark straight line in Fig. 4.3 is the
Population Regression Line (PRL) and it is given by the following equation:
𝐸(𝑌|𝑋) = 𝛽 + 𝛽 𝑋 . …(4.8)
Therefore, the Population Regression function (PRF) can be expressed as
𝑌 = 𝐸(𝑌 |𝑋 ) + 𝑢
Or,
𝑌 =𝛽 +𝛽 𝑋 +𝑢 … (4.9)

Thus, the Population Regression Function in equation (4.9) is a combination of


population regression line (PRL) 𝐸 (𝑌 |𝑋 ) and random error term 𝑢 . The SRF is
only an approximation of PRF. We attempt to find the most appropriate sample
that yields estimators 𝑏 and 𝑏 which are as close as possible to population
49
Regression Model: Two
Variables Case
parameters 𝛽 and 2 . In other words, 𝑏 is as close as possible to 1 , and 𝑏 is as
close as possible to 2 .

4.4 ASSUMPTIONS OF CLASSICAL REGRESSION


MODEL
A linear regression model is based on certain assumptions as specified below. If a
regression model fulfils the following assumptions, it is called the classical linear
regression model (CLRM). The assumptions of CLRM are as follows:
(i) The regression model is linear in parameters. It may or may not be
linear in variables. For example, the equation given below is linear in
parameters as well as variables as shown in equation (4.10)
Yi = β1 + 𝛽 𝑋 + ui …(4.10)
(ii) The explanatory variable is not correlated with the disturbance term u.
This assumption requires that ∑ 𝑢 𝑋 = 0 . In other words, the
covariance between error term and explanatory variable is zero. This
assumption is automatically fulfilled if X is non-stochastic. It requires
that the 𝑋 values are kept fixed in repeated samples.
(iii) The expected value or mean value of the error term u is zero. In
symbols, 𝐸(𝑢 |𝑋 ) = 0. It does not mean that all error terms are zero.
It implies that the error terms cancel out each other.
(iv) The variance of each 𝑢 is constant. In symbols, 𝑣𝑎𝑟(𝑢 ) = 𝜎 . The
conditional distribution of the error term has been displayed in Fig.
4.4(a). The corresponding error variance for a specific value of the
error term has been depicted in Fig. 4.4(b). From the figure you can
make out that the error variance is constant at all levels of the X
variable. It describes the case of ‘homoscedasticity’.

Y PRL:𝐸(𝑌|𝑋 )=𝛽 + 𝛽 𝑋

+𝑢

-𝑢

0 X

Fig 4.4 (a) Conditional Distribution of Error Term 𝒖𝒊

50
Simple Linear Regression
Model: Estimation
Y

PRF: 𝑌 = 𝛽 + 𝛽 𝑋

0 X

Fig 4.4 (b) Homoscedasticity (equal variance)


Fig. 4.5 depicts the case of unequal error variance, i.e., heteroscedasticity. Here
the variance of the error terms varies across the values of Xi.
y

PRF:𝑌 = 𝛽 + 𝛽 𝑋

0 X

Fig. 4.5: Case of Heteroscedasticity (Unequal Variance)


(v) There is no correlation between the two error terms. This is the
assumption of no autocorrelation.
𝑐𝑜𝑣 𝑢 , 𝑢 =0 𝑖≠𝑗
It implies that there is no systematic relationship between two error
terms. This assumption implies that the error terms 𝑢 are random. 51
Regression Model: Two
Variables Case
Since two error terms are assumed to be uncorrelated, any two Y
values will also be uncorrelated, i.e., 𝑐𝑜𝑣 𝑌 , 𝑌 = 0.
Fig 4.6(i) depicts the case of no autocorrelation. Fig 4.6(ii) depicts
positive autocorrelation, and Fig 4.7(iii) shows the case of negative
autocorrelation.

(i) No Autocorrelation (ii) Positive Autocorrelation (iii) Negative Autocorrelation

Fig 4.6: Various Cases of Autocorrelation


(vi) The regression model is correctly specified, that is, there is no
specification error in the model. If certain relevant variable is not
included or certain irrelevant variable is included in the regression
model then we commit model specification error. For instance,
suppose we study the demand for automobiles. If we take the price of
automobiles only and do not include the income of the consumer
income then there is some specification error. Similarly, if we do not
take into account costs of adverting, financing, gasoline prices, etc.,
we will be committing model specification error (we will discuss the
issue of specification error in Unit 13).

4.5 ORDINARY LEAST SQUARES METHOD OF


ESTIMATION
As mentioned in Unit 1 of this course, we need to estimate the parameters of the
regression model. There are quite a few methods of estimation of the parameters.
In this course will discuss about two such methods: (i) Least Squares, and (ii)
Maximum Likelihood. We discuss about the Ordinary Least Squares (OLS)
method below.
The Ordinary Least Squares (OLS) method estimates the parameters of a linear
regression model by minimising the error sum of squares (ESS). In other words,
it minimizes the sum of the squares of the differences between the observed
dependent variable (𝑌 ) and the predicted or expected value of the dependent
variable (𝑌 ).
52
In symbols, Simple Linear Regression
Model: Estimation
𝑒 = Yi − 𝑌

𝑒 = Yi − 𝑌
∑ 𝑒 =∑ Yi − 𝑌 ... (4.11)

In OLS method we minimise ∑ 𝑒 .


We know that
𝑌 = 𝑏1 + 𝑏2 𝑋𝑖
If we substitute the value of 𝑌 in equation (4.11) we obtain
∑ 𝑒 = 𝛴(𝑌𝑖 − 𝑏1 − 𝑏2 𝑋𝑖 )2

The first order condition of minimization requires that the partial derivatives are
equal to zero. Note that we have to decide on the values of 𝑏 and 𝑏 such that
ESS is the minimum. Thus, we have take partial derivates with respect to 𝑏 and
𝑏 . This implies that

=0 … (4.13)

and

=0 … (4.14)

From equation (4.13) we have

2𝛴 (𝑌 − 𝑏 − 𝑏 𝑋 ) (−1) = 0

By re-arranging terms in the above equation we obtain

𝛴𝑌 = 𝑛𝑏 + 𝑏 𝛴𝑋 … (4.15)

In equation (4.15), note that n is the sample size.

From equation (4.14) we have

2𝛴 (𝑌 − 𝑏 − 𝑏 𝑋 ) (−𝑋 ) = 0

By re-arranging terms in the above equation we obtain

𝛴𝑋 𝑌 = 𝑏 𝛴𝑋 + 𝑏 𝛴𝑋 … (4.16)

Equations (4.15) and (4.16) are called normal equations. We have two equations
with two unknowns (𝑏 and 𝑏 ) .
Thus, by solving these two normal equations we can find out unique values of 𝑏
and 𝑏 .
53
Regression Model: Two
Variables Case
By solving the normal equations (4.15) and (4.16) we find that
𝑏 = 𝑌 − 𝑏 𝑋¯ … (4.17)
and
( ¯ )( )
𝑏 = ( )
= ( )

Let us take the variables X and Y in deviation forms such that


𝑥 =𝑋 −𝑋 𝑦 =𝑌 − 𝑌
Thus,

𝑏 = … (4.18)

As you can see from the formula for b2, it is simpler to write the estimator of the
slope coefficient in deviation form. Expressing the values of a variable from its
mean value does not change the ranking of the values, since we are subtracting
the same constant from each value. It is crucial to note that b1 and b2 are
expressed in terms of quantities computed from the sample, given by the formula
in expressions in (4.17) and (4.18).
We mention below the formulae for variance and standard deviation of the
estimators b1 and b2

𝑉𝑎𝑟(𝑏 ) = 𝜎 = 𝜎 … (4.19)

𝑆𝐸 (𝑏 ) = 𝑉𝑎𝑟(𝑏 ) … (4.20)

𝑉𝑎𝑟(𝑏 ) = 𝜎 =

𝑆𝐸 (𝑏 ) = √var(𝑏 ) … (4.21)

σ = = = … (4.22)
. .

S.E. of the residual (𝑒 ) = √σ … (4.23)


The formulae mentioned in equations (4.19), (4.20), (4.21), (4.22) and (4.23) are
the variance and standard errors of estimated parameters b1 and b2.
Smaller the value of σ , closer is the actual Y value to its estimated value. Recall
that any linear function of a normally distributed variable to itself normally
distributed. If 𝑏 and 𝑏 are linear functions of normally distributed variable 𝑢
they themselves are normally distributed. Thus,

𝑏 ~𝑁 𝛽 , 𝜎 … (4.24)

𝑏 ~𝑁 𝛽 , 𝜎 … (4.25)
54
Check Your Progress 2 Simple Linear Regression
Model: Estimation
1) Distinguish between the error term and the residual by using appropriate
diagram.
.........................................................................................................................
.........................................................................................................................
.........................................................................................................................
.........................................................................................................................
.........................................................................................................................
2) Prove that the sample regression line passes through the mean values of X
and Y.
.........................................................................................................................
.........................................................................................................................
.........................................................................................................................
.........................................................................................................................
.........................................................................................................................

4.6 ALGEBRAIC PROPERTIES OF OLS


ESTIMATORS
The OLS estimators b1 and b2 fulfil certain important properties.

a) SRF obtained by OLS method passes through sample mean values of X


and Y. This mainly implies that the point ( 𝑋¯, 𝑌¯) passes through the
Sample Regression Line.

𝑌¯ = 𝑏 + 𝑏 𝑋¯ …(4.26)
Mean value of residuals 𝑒¯ is always zero 𝑒¯ = = 0. This implies that on
an average, the positive and negative residual terms cancel each other.

b) 𝛴𝑒 𝑋 = 0 …(4.27)

The sum of product of residuals 𝑒 and the values of explanatory variable


X is zero, i.e., the two variables are uncorrelated.

c) 𝛴𝑒 𝑌𝑖 = 0 …(4.28)

The sum of product of residuals 𝑒 and estimated 𝑌 is zero, i.e., 𝑒 𝑌𝑖 = 0.

4.7 COEFFICIENT OF DETERMINATION


Let us consider the regression model:
𝑌 =𝛽 +𝛽 𝑋 +𝑢
Recall from equation (4.7) that
𝑌 = 𝑌𝑖 + 𝑒
55
Regression Model: Two
Variables Case
If we subtract 𝑌 from both sides of the above equation, we obtain
(𝑌 − 𝑌) = (𝑌𝑖 − 𝑌) + (𝑌 − 𝑌𝑖 ) … (4.29)
[Since 𝑒 = 𝑌 − 𝑌𝑖 ]
In equation (4.20) there are three terms: (i) (𝑌 − 𝑌) which is the variation in 𝑌 ,
(ii) (𝑌𝑖 − 𝑌¯ ) which is the explained variation, and (iii) (𝑌 − 𝑌𝑖 ) which is the
unexplained or residual variation.
Now, let us use the lower case letters to indicate deviation from mean of a
variable. Equation (4.30) can be written as
𝑦 =𝑦 +𝑒 … (4.30)
Since ∑ 𝑒 = 0, we have 𝑒¯ = 0.
Therefore, we have 𝑌 = 𝑌 , that is, the mean values of the actual Y and the
estimated Y are the same.
Recall that
𝑌 =𝑏 +𝑏 𝑋 +𝑒 ... (4.7)
and
𝑌¯ = 𝑏 + 𝑏 𝑋¯ …(4.26)
If we subtract equation (4.26) from equation (4.7), we get
𝑦 =𝑏 𝑥 +𝑒 …(4.31)
If find OLS estimator of (4.31), we obtain
yˆi  b2 xi .
Therefore,
𝑦 =𝑦 +𝑒 ... (4.32)
Now let us takes squares of equation (4.32) on both sides and sum it over the
sample. After re-arranging terms, we obtain
yi2  yˆ i2   ei2 … (4.33)

Or, equivalently,
yi2  b22 xi2  ei2 … (4.34)

Equation (4.34) can be expressed in the following manner;


TSS = ESS + RSS … (4.35)
where TSS = Total Sum of Squares
ESS = Explained Sum of Squares
RSS = Residual Sum of Squares
Let us divide equation (4.35) by TSS. This gives us
56
1= + … (4.36) Simple Linear Regression
Model: Estimation

Now, let us define

𝑅 = … (4.37)

The 𝑅 is called the coefficient of determination. It is considered as measure of


goodness of fit of a regression model. It is an overall ‘goodness of fit’ that tells us
how well the estimated regression line fits the actual Y values.
4.7.1 Formula of Computing R2
Using the definition of 𝑅 given at equation (4.37), we can write equation (4.36)
as:

1=𝑅 + =𝑅 +

Therefore,

𝑅 =1− … (4.38)

You should note that 𝑅 gives the percentage of TSS explained by ESS. Thus, if
𝑅 = 0.75, we can say that 75 per cent variation in the dependent variable is
explained by explanatory variable in the regression model. The value of R2 or
coefficient of determination lies between 0 and 1. This is mainly because it
represents the ratio of explained sum of squares to total sum of squares.
Now let us look into the algebraic properties of 𝑅 and interpret it. When 𝑅 = 0
we have ESS = 1. It indicates that no proportion of the variation in the dependent
variable is explained by ESS. If R2 = 1, the sample regression is a perfect fit. If
R2 = 1, all the observations lie on the estimated regression line. A higher value of
the R2 implies a better fit of a regression model.
4.7.2 F-Statistic for Goodness of Fit
The statistical significance of a regression model is tested by the F-statistic. By
using the t-test we can test the statistical significance of a particular parameter of
the regression model. For example, the null hypothesis 𝐻 : 𝛽 = 0 implies that
there is no relationship between Y and X in the population. By using F-statistic,
we can test the null hypothesis that all the parameters in the model are zero.
Therefore, we use F-statistics for goodness of fit.

F-statistics for goodness of fit is given by the following:


/( )
𝐹= … (4.39)
/( )

where k is the number of parameters in regression equation and n is the sample


size.
57
Regression Model: Two
Variables Case
4.7.3 Relationship between F and R 2
/( )
From equation (4.39) we know that 𝐹 = . If we divide the numerator
/( )
and the denominator by TSS, we have
/ /( ) /( )
𝐹= = … (4.40)
/ /( ) ( )/( )

Note that the F-statistic is an increasing function of 𝑅 . An increase in the value


of 𝑅 means an increase in the numerator and a decrease in the denominator.
Now let us explain the interpretation of F-static obtained in equation (4.41). The
value obtained by applying equation (4.41) to a dataset is the calculated value of
F or F-calculated. We compare this value with the tabulated value or critical
value of F given at the end of the book. For comparison purpose the degrees of
freedom are ((𝑘 − 1), (𝑛 − 𝑘)).
If F-calculated is greater than F-critical we reject the null hypothesis 𝐻 : 𝛽 = 0.
An implication of the above is that the independent variables explain the
dependent variable. In other words, there exists a statistically significant
relationship between Y and X.
If F-calculated is less than F-critical we do not reject the null hypothesis 𝐻 : 𝛽 =
0. Thus there is no significant relationship between Y and X.
4.7.4 Relationship between F and t2
There is relationship between the F-statistic and the t-statistic in a regression
model. Suppose, the number of explanatory variables k = 2.
/( )
𝐹=
/( )

For the two-variable model,


/( )
𝐹= = …(4.41)
/( ) /( )

We know that ESS 𝑌 − 𝑌¯ and 𝑅𝑆𝑆 = 𝑒

Therefore,
(𝑌𝑖 ¯) ([ ] [ ¯ ])
F= = ... (4.42)
⁄( )

Estimation of error variance = σ = = …(4.43)

𝐹= . 𝑏 (𝑋 − 𝑋¯) …(4.44)

We know that

𝑣𝑎𝑟(𝑏 ) =

Substituting equation (4.43) in equation (4.44) we get,

𝐹= = ( )
=[ ( )]
=𝑡 ... (4.45)
58
Therefore, the F-statistic is equal to square of the t-statistic (𝐹 = 𝑡 ). The above Simple Linear Regression
Model: Estimation
result, however, is true for the two-variable model only. If the number of
explanatory variable increases in a regression model, the above result may not
hold.
Check Your Progress 3
1) Is it possible to carry out F-test on the basis of the coefficient of
determination? Explain how.
.........................................................................................................................
.........................................................................................................................
.........................................................................................................................
.........................................................................................................................
.........................................................................................................................
2) Can the coefficient of determination be greater than 1? Explain why.
.........................................................................................................................
.........................................................................................................................
.........................................................................................................................
.........................................................................................................................
.........................................................................................................................

4.8 LET US SUM UP


In this unit we discussed about the classical linear regression model, which is
based on certain assumptions. We distinguished between the population
regression function and the sample regression function. We explained why a
stochastic error term is added in a regression equation. We explained the meaning
of each of the assumptions of the classical regression model. The procedure of
obtaining OLS estimators of a regression model is given in the Unit. The unit
further elaborated on the notion of goodness of fit and concept of R-squared.

4.9 ANSWERS/ HINTS TO CHECK YOUR


PORGRESS EXERCISES
Check Your Progress 1
1) The objectives of carrying out a regression model could be as follows:
 To estimate the mean or the average value of the dependent variable,
given the values of independent variables.
 To test the hypotheses regarding the underlying economic theory. For
example, one may test the hypotheses that price elasticity of demand is
(–)1.
 To predict or forecast the mean value of the dependent variable given
the value of the independent variable.

59
Regression Model: Two
Variables Case
2) The relationship between Y and X is stochastic in nature. There is an error
term added to the regression equation. The inclusion of the random error
term leads to a difference between the expected value and the actual value
of the dependent variable.
3) There are three reasons for inclusion of the error term in the regression
model. See Sub-Section 4.2.2 for details.
Check Your Progress 2
1) Go through Section 4.3. You should explain the difference between the
error term and the residual by using Fig. 4.3.
2) In the OLS method we minimise 𝛴𝑒 by equating its partial derivates to
zero. The condition = 0 gives us the first normal equation:
𝑌 = 𝑛𝑏 + 𝑏 𝛴𝑋 . If we divide this equation by the sample size, n, we
obtain 𝑌¯ = 𝑏 + 𝑏 𝑋¯ . Thus, the estimated regression passes through the
point 𝑋¯, 𝑌¯ .
Check Your Progress 3
1) Yes, we can carry out F-test on the basis of the 𝑅 value. Go through
equation (4.40).
2) The value of R2 or the coefficient of determination lies between 0 and 1.
This is mainly because it represents the ratio of ESS to TSS. It indicates
the proportion of variation in Y that has been explained by the
explanatory variables. The numerator ESS cannot be more than the TSS.
Therefore, R2 cannot be greater than 1.

60

You might also like