Transcript ppt

Introduction to Inferential Statistics
Statistical analyses are initially divided into:
Descriptive Statistics or Inferential Statistics.
Descriptive Statistics are used to organize and/or summarize the
parameters associated with data collection (e.g., mean, median, mode,
variance, standard deviation)
Inferential Statistics are used to infer information about the relationship
between multiple samples or between a sample and a population (e.g., t-test,
ANOVA, Chi Square).
The Scientific Method uses Inferential Statistics to determine if the
independent variable has caused a significant change in the dependent
variable and if that change can be generalizable to the larger population.
If a dependent variable within the population is normally distributed and we
can calculate or estimate the mean and standard deviation of the population,
then we can use probabilities to determine significance of these
relationships.
To begin we must collect a representative sample of our much larger
population.
Sampling
A Representative Sample means that all significant subgroups of the
population are represented in the sample.
Random Sampling is used to increase the chances of obtaining a
representative sample. It assures that everyone in the population of
interest is equally likely to be chosen as a subject.
Random Number Generators or Tables are used to select random
samples. Each number must have the same probability of occurring as
any other number. The larger the random sample, the more likely it will be
representative of the population.
Sampling Distribution
Generalization refers to the degree to which the mean of a representative sample
is similar to or deviates from the mean of the larger population. This generalization
is based on a distribution of sample means for all possible random samples.
This distribution is called a Sampling Distribution.
It describes the variability of sample means that could occur just by chance and
thereby serves as a frame of reference for generalizing from a single sample mean
to a population mean.
It allows us to determine whether, given the variability among all possible sample
means, the one observed sample mean can be viewed as a common outcome (not
statistically significant) or as a rare outcome (statistically significant).
All possible random samples of even a modest size population would consist of a
very large number of possibilities and would be virtually impossible to calculate by
hand. Therefore we use statistical theory to estimate the parameters.
Sampling Distribution
The Central Limit Theorem states that the distribution of sample means
approaches a normal distribution when n is large.
In such a distribution of unlimited number of sample means, the mean of the
sample means will equal the population mean.
Standard Error of the Mean
The standard deviation of the distribution of sample means is called the Standard
Error of the Mean or Standard Error for short. It is represented by the following
formula:
Since the standard deviation of the population is often unavailable, a good estimate
of the standard error uses the standard deviation of the sample (s). This newly
modified formula of the Standard Error is shown below:
Statistical Significance
Now that we have the parameters of the Sampling Distribution we can see how
to use this distribution to determine if the mean difference between two samples or
between a sample and a population are significantly different from each other.
The Research Hypothesis (H1) states that the sample means of the groups are
significantly different from one another.
The Null Hypothesis (H0) states that there is no real difference between the
sample means.
Anytime you observe a difference in behavior between groups, it may exist for two
reasons:
1.) there is a real difference between the groups, or
2.) there is no real difference between the groups, the results are due to error involved in
sampling.
This error can be described in two ways:
Type I error is when you reject the null hypothesis when shouldn't have
Type II error is when you fail to reject the null hypothesis when you
should have
Statistical Significance
The probability of committing a Type I
error is designated by alpha.
An alpha level of 0.05 is reasonable and
widely accepted by all scientists.
The null hypothesis can be rejected if
there is less than 0.05 probability of
committing a Type I error ( p < .05 ).
Two-Tailed Hypothesis
If the research hypothesis does not predict a direction of the results, we say it is a
Two-Tailed Hypothesis because it is predicting that alpha will be split between
both tails of the distribution. If the sample mean falls in either of these areas we can
reject the null hypothesis. This is shown below:
One-Tailed Hypothesis
If the scientific hypothesis predicts a direction of the results, we say it is a OneTailed Hypothesis because it is predicting that alpha will fall only in one specific
directional tail. If the sample mean falls in this area we can reject the null
hypothesis. This is shown below: