A Gentle Introduction to Non-Parametric Statistics (15-1) - YouTube

Channel: Research By Design

[8]
We are now going to look at a special class of tests that give us the ability to do statistical analyses in
[15]
circumstances when parametric tests just won't do. They are called
[20]
non-parametric statistics, and I'm going to tell you why we need them.
[26]
From all of our studies of parametric statistics like t-tests and ANOVA,
[31]
we have learned how to compare groups using scale level data.
[35]
But what if you only have nominal data?
[40]
Well, you have options.
[43]
Let's imagine that we are having an end-of-semester party and we need to order soft drinks.
[49]
Somebody suggests that we just buy a case of whatever is on sale because people will drink
[55]
whatever is available, but you think that maybe people prefer one brand over others.
[62]
So you offer the class a choice of six brands of soft drinks and ask everyone to choose their favorite.
[70]
But how are you going to analyze these data?
[73]
How can you tell if people truly have a preference for one type of soft drink
[78]
or if they are simply choosing one at random?
[83]
Most of the statistics that we have used so far
[87]
estimate population parameters from a distribution of scores in a sample.
[92]
So, for example, the t-test uses sample variance as an estimator of population variance.
[100]
Tests like this are called "parametric" tests,
[104]
and they work because the central limit theorem shows us that the distribution of sample means
[110]
will form an approximately normal distribution.
[115]
But there is a class of tests that do not rely upon parameter estimations or
[120]
distribution assumptions, and these are called "non-parametric" statistics.
[127]
Non-parametrics are especially useful with nominal and ordinal level data, although,
[132]
they can be used with interval and ratio level data, as well.
[137]
So, for example, you want to make a comparison about whether males or females
[143]
are more likely to help a person in either emotional distress or physical peril.
[150]
Or perhaps you want to know if patterns of soft drink consumption are the same at a state university as at a private school.
[158]
Maybe you want to know if certain dormitories are
[161]
over-represented or underrepresented in their number of student government representatives.
[168]
Nonparametric statistics have several advantages over parametric statistics.
[174]
Number one:
[176]
Non-parametric tests are not susceptible to outliers the way that parametric tests are.
[182]
One outlier score can affect a parametric test by inflating the variance and hence the error term.
[191]
This can invalidate a conclusion drawn by the parametric test. So you make a type I error or perhaps even a type II error.
[201]
Outliers in correlation can change the direction or the strength of a relationship between variables,
[208]
but non-parametric tests do not have this limitation because they measure central tendency using ranked data
[217]
Ranking uses the median rather than a mean.
[221]
So, for example, if the tenth number in your distribution is an outlier, that
[226]
outlier can really shift the mean when you add it to other numbers.
[231]
But when you are considering ranks, it is still just the tenth number no matter how big the actual value is.
[239]
So the main advantage of nonparametric statistics is that they do not require
[245]
restrictive assumptions about the distribution.
[248]
For instance, they do not assume that soft drink consumption is normally distributed.
[254]
Non-parametric tests can also be used when you have non-normal, interval, or ratio data, such as data that are highly skewed.
[263]
Typically, therefore, non-parametric statistics are only used when the data are too skewed to use a parametric test.
[272]
Another advantage is that non-parametric statistics are typically easier to calculate by hand, and we will practice
[280]
doing some of them by hand in order to learn them.
[284]
However, the main disadvantage of non-parametric statistics
[288]
is that they are not as powerful as parametric tests.
[292]
That means that they are more likely to miss an effect that truly exists or
[298]
that you make a Type II error.
[302]
So as you get better with statistics you will be using mostly parametric statistics,
[308]
But you should keep some tools in your back pocket that you can use at times that the parametric statistics fail.
[315]
Each parametric test has a corresponding
[319]
non-parametric test.
[321]
Here you can see each parametric test and it's corresponding non-parametric
[327]
alternative. So, for example, if you plan to do an independent samples t-test,
[332]
but the data were highly skewed you could use a Mann-Whitney U Test,
[338]
instead. If you wanted to do a one-way ANOVA,
[342]
but instead of groups you were looking for differences over five years of a program,
[348]
you could do the Kruskal-Wallis test, instead.
[352]
Each of these non-parametric statistics can be conducted in SPSS,
[358]
but there is another test called the
[362]
"chi-square" test that is an ever-present help in times of statistical trouble, and
[367]
It would be the perfect tool for analyzing our soft-drink data.
[373]
Anytime you want to know "does my sample look like the population,"
[379]
you can use the chi-square.
[382]
This adaptability makes chi-square one of the most popular and versatile of the non-parametric statistics.
[390]
The Bear Handout on the page about choosing the right statistical test
[394]
also has a comparison of the parametric statistics alongside their non-parametric alternatives.