馃攳
Understand Autocorrelation in 15 minutes - Dr. Tehseen Jawaid - YouTube
Channel: unknown
[4]
Bismillah-Hir-Rehman-Nir-Raheem
[7]
Today we are discuss the second assumption of OLS
[12]
So, What is the second assumption?
[16]
Error term observation are independent of each other
[34]
Or
[37]
They are not co-related with each other
[49]
Well, first we look at co-relation.
[53]
Like this we have the movement of a variable
[57]
This is the movement of the other variable
[59]
So see the movement of both is on the same time
[62]
Its increase is seeing its increase
Its decrease is seeing its decrease
[65]
So if the moment is closer to each other and associate
Co-relation will be so strong
[71]
And if the movement does not happen with
each other, then the co-relation will be so weak
[76]
But this is the co-relation when two series
are co-relating with each other
[81]
If I have series x and it has any values 10, 20, 30, 40, 50
[88]
And I want to know if x is correlating with its own values
[92]
Now I'm going to create another value
here Xt this is current value
[96]
And I made a variable here t-1
[99]
t-1 means, each value with its previous value
[103]
40, 30, 20, 10.
[106]
If these two series are correlating with each other
[110]
So I can conclude this here
[112]
that x is correlating with its own value
[115]
So to correlate is called as correlation
[119]
And when any series starts correlating with its own value
[123]
So when it comes to its own values, so such
correlation is called Auto corelation
[131]
Auto correlation can exist in every variable like we said GDP
[135]
Today's GDP will correlate with last
year's GDP Absolutely logical
[139]
There is an investment, It is absolutely logical
There is an saving, it is absolutely logical
There is an consumption, it is absolutely logical
[146]
So every variable, most probably
[150]
Correlates with its old values
[155]
Now we look ahead. Let's assume
that is a regression model.
[169]
Y we have dependent variable, x
and z we have independent variable
[174]
And we have an error term.
[177]
We have discussed this
[179]
These error terms represent all variables
that are not part of the model
[188]
And those that are part of the model appear
as an independent variable in the model
[194]
Actually error, understand one thing about error
[198]
An error is called a mistake
[204]
And there is a basic characterstic of mistake
[207]
The main characteristic of a mistake is that
the mistake or error is not intentionally do.
[214]
So when not done intentionally, there is no consistency
[218]
Mean's no sequence or trend exists in the error
[222]
So that's why it's an error term
[224]
Because this is also a mistake
[226]
So what is the most important characteristic
of this error, the error should be raindom
[232]
The error should be random.
The error should not be consistence.
[235]
There should not be exist any trend in this
[241]
Now what happened, A researcher developed this model
[245]
But due to lack of information
Due to lack of literature review
[251]
It could not do the modeling correctly
and the model stopped here.
[258]
Mean, Z was a important variable.
[260]
Which was part of this model but not considered in it.
[263]
Now you know that variables that
are present in the model are present
[266]
and those that do not exist, that all exist in the error term.
[268]
In exactly the same way that Z was miss here
[271]
The Z that was missed went into the error term
[277]
It has many variables other then these variable are
all in the error term so Z also goes there.
[282]
Z was an important variable that was very important for Y
[285]
And we have already seen this
[287]
Any variable and if it correlates with its own value
[290]
Develop a relationship or make a friend. It is not logical
[294]
Because, All variables in logical are auto-correlated
[297]
So it is not a problematic thing that the variables are correlated.
[302]
The same will happen here
[304]
Now we see that for example we have annual data, time series data.
[307]
So the value of z for each year is correlating
[312]
Correlating, showing consistency, creating
a trend it is doing a Z
[318]
Now consider one thing here. There is a box
in which I keep the remote control car
[325]
There is a big carton and the car is not visible to anyone
[329]
I moved the car and it came here
[333]
Actually, Who will be doing the move, this car
but we will be seeing this box
[338]
We must have seen this box. In exactly the same way
[342]
Here Z is the consistence
[345]
Z forming a trend Z is correlating with its previous value
[349]
But Z's are not separate variables
[352]
It's added to the error term so we'll see it like this
[358]
This error term is consistency, the error
term is correlating with its previous value
[364]
When the error term is correlated with its previous value
[367]
So we said that any series that correlates
with its previous value is called autocorrelation
[380]
But then we also saw that there are
no problems with auto-correlation
[383]
Each variable may be autocorrelated
[385]
The important point here is that the problem
in autocorrelation will occur at this point
[390]
When autocorrelation is present in the error term
[394]
If the variable is autocorrelated. It is
logical that it is acceptable and it should be
[399]
But when you talk about error, the
error should not be autocorrelated
[402]
Because, as we have already seen
above, errors should be random
[406]
If there is any consistency in the error term
[409]
A correlation exists by its own value
[412]
So this assumption will be valid
[414]
It is saying error term observation are independent
of each other or they are not correlated with each other.
[419]
If this is not , it means that
autocorrelation exists in error
[434]
So what was the problem
that we saw? Autocorrelation
[446]
Autocorrelation
[448]
Now let's understand what is meant
by consistency in autocorrelation
[452]
We know that the actual difference of Y-Y(cap) is called error
[459]
Assume this is a possibility
[461]
The firm invites any analyst every year
Get your profitability predicted
[466]
So if the error is positive in one year,
it means that the actual profit has increased
[469]
Estimated was given less
[471]
Next year, come again positive
Next year, come again positive
[476]
When the error is coming with the
same sign, it is coming with the consist sign
[480]
So this means that the analyst must be doing some
kind of problem in forecasting that the error is repeating.
[486]
Consistency is happening
[488]
Or if it comes negative then negative
[493]
This means error consistency
[495]
A trend exists in error
[497]
that each incoming error carries the sign of the previous error
[501]
plus plus plus, this
minu, minus, minus
[504]
So we will call it error consistency
Error is autocorrelated
[507]
In exactly the same way, only the same
sign does not show consistency
[511]
plus minus, plus minus,
plus minus, plus minus
[515]
This is also autocorrelation
[517]
Because the opposite is being predicted in this too
[512]
Because consistency is also in the
opposite sign plus minus plus minus
[522]
You can easily forecast
[525]
If the autocorrelation is such that the error is
predicting the same sign. For your incoming error
[530]
So this is called positive correlation
[533]
And if opposite sign is predicting
then it is called negative correlation
[539]
Now the question is if any autocorrelation
occurs in the regression model
[543]
So what is illness, what problem comes in the model?
[550]
Well see this is a distribution
[556]
This is the distribution of a beta coefficient
[560]
So what happens now if positive correlation exists
[565]
So see, consistency is and how consistency is all on one side
[569]
mean are all plus or all minus
[571]
Most are on the same site
[573]
For that reason
It is in the center
[578]
Because of that, if the autocorrelation
with the same sign is mean positive.
[583]
So in them, because all the things are towards
the same sign,
[585]
they are not going towards the opposite
So the deviation is reduced
[589]
And in positive autocorrelation the
distribution will look something like this
[592]
This means that the variable was previously insignificant
[595]
If there is a positive autocorrelation,
it may be looking significant here
[600]
If the variable was previously insignificant, it may
appear significant due to positive correlation
[606]
In exactly the same way as in negative correlation
This is also consistency
[609]
But plus minus plus minus. Deviation increased
[612]
The distribution will look like this
[615]
Because of this, the variable that was previously
significant will start to look insignificant.
[623]
This means that when autocorrelation exists in the model
[627]
The significance of the variable, the probe
value, and its standard error are not reliable.
[635]
You cannot say this with confirmation
[638]
That variable is really significant or insignificant
[642]
So the issue that comes up in regression due
to autocorrelation is that of significant values
[647]
So you can say here
[650]
Issues in significant which is illness.
[655]
What is illness, which issues come to the significant,
the significant is not remain reliable
[666]
So the problem we discussed is Autocorrelation.
[669]
We talked about illness. Significant Issues.
Means, the significant is not reliable
[673]
Now let's talk about Measure
[676]
The measure we use for this autocorrelation
is Darwin Watson statics
[681]
Darwin Watson's range is zero to four
[684]
If its value is 2, it means no Autocorrelation
[689]
If the value is between zero and two,
you would say a positive Autocorrelation
[695]
And between one and two four, then
you will say negative Autocorrelation
[699]
So theoretically we will say that if the
value is exist 2 then there is no autocorrelation
[705]
Even if the value is 1.9 or 2.1,
the values will exist in autocorrelation
[709]
But now we also need to know
[711]
Is the autocorrelation coming in 1.9 so severe
that we should move towards its removal
[717]
So let's do a test to check the severity
[721]
and this is Serial Correlation LM test
[731]
The hypothesis of serial correlation LM test is no autocorrelation
[737]
If its P value is greater than 0.5
[741]
If our level of significance is five percent
[743]
If it will be 10 percent, then we will see from 0.1
[745]
If it is more, then accept the hypothesis
[748]
If it is less, then reject the hypothesis
[752]
Accept means autocorrelation does not exist
[756]
If this hypothesis is rejected, it means that autocorrelation exists
[760]
So if value other than 2 is more or less than that
[763]
The serial correlation LM test should confirm
that the hypothesis is being rejected
[768]
And the serial correlation severe is present in this model
[772]
So then which one will we move towards,
we will move towards Removal
[779]
If this value is different from two, but the hypothesis
is accepted, then no removal needs to be applied.
[785]
So now let's talk about removals,
removals are very easy
[789]
And it is logical that as our regression model was
[796]
And in this the researcher forgot to
add an important variable. which is Z
[800]
And Z had gone into error term.
And the error consistence was visible
[805]
Actually the error was just showing consistence
[807]
But the consistence was z
[809]
So due to which variable the
consistence error was visible.
[813]
Extract it and make it part of the model
[822]
Now the error term which was the consistency due to z
[827]
Which was trending because of Z
[830]
The autocorrelation due to Z was visible in this
[833]
Now that Z has come out of it
[835]
So now the error will go back to its old characterstic
[838]
Means error would be random now
[843]
Now the error has become random and as soon as the error
has become random, the correlation between them has disappeared
[847]
Autocorrelation was eliminated from the model
[849]
So what is the most appropriate removal of it?
Additional of relevant variable
[864]
Well it is not necessary to auto remove from one variable
[867]
Maybe you have to add one, maybe you have to
add second,maybe you have to add a third variable
[871]
Well, the question arises how to know which variable to add
[875]
So there are two things to see in this, number one theory
[877]
The theoretical relationship of the
dependent variable to which variable
[880]
Have you taken all those variables?
[883]
Second, what variables have been used in the literature
[886]
From there you will have the data to
determine which is the important variable
[889]
Which could have been in the model
and you didn't take it into the model
[894]
Number two, cochran orwatt and
gernalise least square (gls)
[904]
Number three removal is AR1
[908]
GLS and AR1 is also a method to remove autocorrelation
[913]
These technically randomize the error term
[915]
If there is a trend and consistency in it, they eliminate it
[918]
But the problem in interpretation is that
Y does not remain Y, Y becomes delta Y
[923]
With some technical explanation it becomes DeltaX
[927]
So, in effect, the autocorrelation is eliminated,
the error term becomes random
[930]
But those that are variables change their form
[933]
So they don't have the same interpretation
that we can do directly above
[938]
If a researcher is using a large sample
in their data.The sample size is large
[949]
And he doesn't want to add any more
variables, he thinks I have enough variables
[954]
So here is one more test, HAC test
[962]
Hetroscedasticity auto correlation consistence test
[965]
If a researcher performs the regression by selecting the HAC test
[970]
So the HAC test, due to autocorrelation,
any changes in probe value and standard error
[977]
Adjusts them and after selecting the HAC test
[981]
Now when the regression comes to you, it is reliable
[984]
It can be significant and can be interpreted
[988]
This is the end of today's lecture.
[991]
If you like this video, please like, share, subscribe
to our channel, and press the bell icon. Thank you.
Most Recent Videos:
You can go back to the homepage right here: Homepage





