Validity (cont.)/Control

Download Report

Transcript Validity (cont.)/Control

Validity (cont.)/Control
RMS – October 7
Validity
• Experimental validity – the soundness of the
experimental design
– Not the same as measurement validity (the
goodness of the operational definition)
• Internal validity
• Construct validity
• External validity
Internal Validity
• Concerns the logic of the relationship
between the independent and dependent
variables
• It is the extent to which a study provides
evidence of a cause-effect relationship
between the IV and the DV
confounding
• Error that occurs when the effects of two
variables in an experiment cannot be
separated – results in a confused
interpretation of the results
• Example
– Group A (IV = 0)
– Group B (IV = 1)
– Behaviour observed
How do you know what could be
confounding?
• Need to make judgments as you design the
experiment
• Be particularly careful with subject variables
Construct Validity
• Extent to which the results support the theory
behind the research
– Would another theory predict the same
experimental results?
• Hypotheses cannot be tested in a vacuum
– The conditions of a study constitute auxilliary
hypotheses that must also be true so that you can
test the main hypothesis
Example
• H1: Anxiety is conducive to learning
• Participants selected on basis of whether they bit their
fingernails (sign of anxiety)
• Observe how fast they can learn to write by holding a
pencil in their toes (a learning task)
• Did not just test impact of anxiety of learning
– Tested that fingernail biting is a measure of anxiety (H2)
– Tested that writing with toes is a good learning task (H3)
• If either is false, you could have negative results even if
the main hypothesis is true
Similarities
Construct Validity
• Try to rule out other
possible theoretical
explanations of the results
• Must design a new study to
help you choose between
competing theoretical
explanations
Internal Validity
• Try to rule out alternative
variables as potential causes
of the behaviour of interest
• May be able to redesign the
study to control for the
source of confounding
External Validity
• How well the findings of an experiment
generalize to other situations or populations
– Strictly speaking results are only valid for other
identical situations
– Can be hard to know which situational variables
are important
• Ecological validity is related
– Extent to which an experimental situation mimics
a real-world situation
Statistical Validity
• Extent to which data are shown to be the result of a causeeffect relationship rather than an accident
– Does the relationship exist or was it caused by pure chance
•
•
•
•
Similar to internal validity
Notion of power (is n big enough?)
Is measure accurate?
Is the statistical test appropriate?
– 3 kinds of lies: lies, damn lies, and statistics
• Statistical test establishes that an outcome has a certain
low probability of happening by chance alone
– No guarantee that it’s not a random error in sampling or
measurement
So we did the green study again and got no
link. It was probably a – RESEARCH
CONFLICTED ON GREEN JELLY BEAN/ACME
LINK; MORE STUDY RECOMMENDED
- xkcd.com/882/
Threats to Internal Validity - 1
• Events outside the laboratory (history):
– Can occur when different experimental conditions are
presented to subjects at different times
• Example: Feelings of depression (failure condition on
Monday, success condition on Wednesday)
– What if Monday was rainy and Wednesday was sunny?
• Maturation:
– A source of error in an experiment related to the
amount of time between measures
– Subjects may change between conditions because of
naturally occurring processes
• Aging doesn’t just occur with people!
Threats to Internal Validity - 2
• Effects of Testing:
– Changes caused by testing procedure (not processes
unrelated to the experiment)
• Learn how to do the task, learn the style of a test
• Regression Effect:
– tendency of subjects with extreme scores on a 1st measure
to score closer to the mean on a second testing
– Not a perfect correlation between 2 variables
• SAT and GPA or repeated SAT
• Arises when there is an error associated with the measurement of
the variables (e.g., students know some answers and make
lucky/unlucky guesses at the rest)
• Random error – that part of the value of a variable that can be
attributed to chance
Threats to Internal Validity - 3
• Mortality:
– The dropping out of some subjects before an
experiment is completed
– Selective subject loss
– Can introduce bias depending on the reason they
are dropping out
• Selection:
– Any bias in selecting groups can undermine
internal validity
Threats to Construct Validity – 1
• Difficult as there are an indefinite number of
theories that may account for a given
relationship
• Strategy: ask whether alternative theoretical
explanations of the data are less plausible
than the theory believed to be supported by
the research
Threats to Construct Validity - 2
• Loose connection to theory and method
– The anxiety/learning example
– Often research suffers from poor operational
definition of theoretical constructs
• Ambiguous effect of IV
– Experimenter can control all reasonable confound
variables
– Participant may compromise result by seeing the
situation differently than the experimenter
Human-subject research
• Good subject tendency:
– Participants want to help the research achieve
their goals (and may not understand the goals)
• Evaluation apprehension:
– Tendency of participants to alter their behaviour
to appear as socially desirable as possible
Threats To External Validity
• Other subjects
– Are the participants truly representative of the sample
to which you are trying to generalize?
• Other times
– Would the same experiment conducted at another
time produce the same results?
– Caution: the web is rapidly changing
• Other settings
– Lab vs field
– Structured vs unstructured environments
Reading: DRAM errors in the wild
• What’s the research problem?
• What was their approach?
• Hypotheses?
– IV, DV
– Measurement validity
• Internal validity
• Construct validity
• External validity
Control
• Basic idea: isolate the effect of the treatment
on the dependent variable
– Ensure there are no reasonable alternative causes
• Basic design: control vs. experimental group
• Setting:
– Field vs. lab
– Experiment vs. non-experiment
Allocation (sampling)
• Random, stratified random, matching
– Video
http://www.youtube.com/watch?v=rASK8PpqakM
&feature=related
Experimental Design
• Within subjects
– All subjects receive all conditions
• Between subjects
– All subjects receive only one of the conditions
• Mixed design
– Some factors between, some within