| |
INTRODUCTION TO STATISTICAL ANALYSIS
Hypotheses Testing
a. The common approach to testing a hypothesis is to establish a set
of two mutually exclusive and exhaustive hypotheses about the value in
question. For example:
Ho: Mean = 10 pounds
Ha: Mean does not equal 10 pounds
b. Next we perform our statistical analysis (perhaps in this
particular example we are testing that the mean weight loss from a new
diet drug is 10 pounds). If our statistical analysis is found to be
significant then we reject the null hypothesis (Ho) and find that the
mean does not equal 10 pounds.
c. Occasionally, the hypotheses will be directional (also known as a
one-tail test). For example:
Ho: mean = 10 pounds
Ha: mean > 10 pounds
d. When we write hypotheses we normally hope to reject the null
hypothesis and find that there is a difference. The exception to this
being in some forms of multivariate data analysis it is required to
accept the null hypothesis in order to test the assumptions of the
statistical technique. For what we are doing in this course the proper
course of action will be to reject the null hypothesis.
How do we determine when we can reject or accept a null hypothesis? The
answer lies in examining the significance level, labeled as alpha ( ).
e. The significance level is normally labeled as "p value" and is
often reported in statistical reports. Researchers must report their
significance level a-priori (prior to conducting the statistical test).
f. If we wait until after we have conducted the statistical tests,
the results do not change but it would be comparable to deciding what is
a good and passing grade on an exam after the exam has been graded. It
can be done, but it is not recommended.
g. There are three common significance levels that we use in
statistical analysis (.10, .05, and .01). If we split this significance
into the two-tails of the curve then we have what is known as a
two-tailed (non-directional) test. We will discuss this more in the
coming weeks.
h. We test whether we have met the required level of significance by
comparing the statistical value that we compute with a known table of
critical values.
i. If the calculated value is greater than the value within the
table of critical values, then we reject the null hypothesis.
ii. If the calculated value is less than the value within the
table of critical values, then we fail to reject (some statisticians
indicate that they accept) the null hypothesis.
i. What does the significance level tell us? The significance level
tells us how likely we are to make an error when we reject the null
hypothesis. If we are testing at the .10 level then we are saving that
10% of the time (10 times out of 100) we will incorrectly reject the
null hypothesis when we should have failed to reject (accept)
i. This concept becomes more important if we consider the various
possibilities for statistical research, and the researcher must take
very seriously the level he or she will set alpha.
ii. If we are conducting research on finances, then it is
probably acceptable to use the .10 level. After all, if we are
testing whether a given investing technique will increase our mean
earnings, then there is little physical harm if we reject the
technique when we should have accepted it.
iii. However, if we are conducting life-saving research on a new
drug, then are we willing to reject the use of a drug that could
save someone’s life? Probably not, we may want to test at a more
restrictive level.
Types of Error
j. Type I Error – this is where the researcher will reject the null
hypothesis when he or she should have accepted the null. Symbolized as
Alpha ( ).
k. Type II Error – this is where the researcher will accept the null
hypothesis when he or she should have accepted the null. Symbolized as
Beta ( )
l. The two types of error are inversely related. As Type I error
increases, Type II error decreases; and likewise, as Type I error
decreases, Type II error increases.
m. Prior to conducting research, the issue of error must be
considered and the researcher will have to determine which type of error
is considered more acceptable. For simplification of this concept see
the table below:
Conclusion of Statistical Test |
Ho is true |
Ho is false |
Ho is not rejected
|
Correct Conclusion |
Type II Error |
Ho is rejected
|
Type I Error |
Correct Conclusion |
The Issue of Power
n. The term power refers to the ability of a statistical test to
reject a false null hypothesis. Thus, it is obvious that we want to
create powerful research designs. We have already discussed that the
probability of a type I error is called alpha, and is usually set at
.05, .01, or .001.
o. Another convention is to set the probability of making a type II
error at .20. This convention demonstrates that most researchers are
less concerned with making a type II error than they are with making a
type I error. Power is defined as 1 - b . So
if b = .20 then power is equal to .80. If
this level of power cannot be achieved in an experiment, then a failure
to reject the null could mean that no relationship exists within the
population, or that the design was simply too weak to detect a real
difference.
p. Power is affected by several considerations:
i. Sample size. As sample size increases, so do does the power of
the study.
ii. Alpha level. As alpha is increased, Beta is decreased,
therefore increasing power.
iii. Direction of the test. A one tail test is more powerful than
a two tail test.
iv. Effect size. As effect size increases, power increases.
Assumptions of Statistical Tests
q. Each test that we will discuss will have its own assumptions.
While a violation of an assumption does not mean that you cannot use the
statistical technique, multiple violations could present a problem.
r. Many of the assumptions are the same across most of the
statistical techniques that we will study.
i. Parametric statistics – statistics that require normal
distribution
ii. Nonparametric statistics – statistics that do not require
normal distribution
|