馃攳
Heteroskedasticity summary - YouTube
Channel: unknown
[0]
Hi there, in this video I am going to be talking about
[4]
homoskedasticity
[6]
as one of the Gauss-Markov assumptions.
[9]
So,
[9]
first of all, what do we mean by homoskedasticity? Well,
[13]
in fact we mean homoskedasticity of our errors.
[16]
Which means that the variance of our errors, given our independent variables 'x',
[23]
is constant.
[25]
So, if I was to think about there being some relationship between y and x,
[30]
and I had some sort of sample of data,
[33]
which looks something
[35]
like this,
[37]
and perhaps I then fit a straight line to this data so that I am using a
[41]
linear model - linear in my independent variable 'x'. Then
[47]
we can sort of think about the errors which our model is making
[51]
are basically constant across our
independent variable 'x'.
[56]
They are basically the same
[57]
as I increase my 'x' variable - all the errors lie within
[62]
straight error bars.
[65]
Well, we can contrast this with the situation where we have
[70]
heteroskedastic errors.
[72]
So here it would be the case if I had my y and x, and I had some points,
[77]
some data points which as
[80]
x increases there is a larger variance in y,
[84]
if I then go ahead and fit a straight line to that data -
[89]
so perhaps my straight line would do something like that.
[91]
We can see that the errors
[93]
which our model is making, are increasing in magnitude,
[99]
as 'x' increases.
[101]
So, if I fit an error line
[105]
indicating the direction of increase of my errors,
[108]
then you can see that my errors are increasing along my 'x' variable.
[113]
So, this is what we call
[117]
heteroskedasticity,
[119]
so
[120]
homo- in this context means that the errors are the same,
[125]
so that's this sort of case, and hetero here means that the errors are different.
[130]
Well, mathematically how do we write that? We write that the variance of our
[136]
errors 'u'
[137]
given our 'x', is some sort of function
[140]
of x - it depends on 'x'.
[142]
Here it is some sort of positive
[144]
function because as my x increases, the magnitude of my errors increase as well.
[150]
So,
[151]
why do we care about our errors being
[155]
homoskedastic? Well,
[157]
as I said it is one of the Gauss-Markov assumptions, and
[160]
if it is violated this means that our
[164]
least-squared estimators are no longer BLUE.
[168]
In particular, they are no longer best.
[171]
So, there are other linear, unbiased estimators which have a lower sampling
[176]
variance.
[177]
Intuitively this means that there are
other
[180]
estimators which are linear and unbiased, which more often, or more
[184]
frequently than least-squares will get
closer to the true population parameters.
[190]
And the intuition
[191]
from this is essentially that - if I have heteroskedastic errors,
[195]
there is some sort of information which
[199]
is inherent in my system which I'm not including in my model.
[202]
And perhaps if I include that information into my model, so I include
[207]
the fact that I expect my errors to increase as 'x' increases,
[211]
then perhaps I can come up with an
estimator which actually
[215]
gets closer to my y values
[220]
more of the time. So that is the underlying intuition for why heteroskedasticity
[226]
means that I can construct another
estimator,
[229]
which has
[230]
a lower variance than least-squares.
[233]
In the next few videos I'm going to give some
[235]
actual examples of where heteroskedasticity arises,
[239]
and that's going to conclude our discussion of the Gauss-Markov assumptions.
Most Recent Videos:
You can go back to the homepage right here: Homepage





