Eeks! Is My Survey Valid?

Download Report

Transcript Eeks! Is My Survey Valid?

SURVEY CONSTRUCTION, VALIDITY,
RELIABILITY
What to think about when
creating a survey instrument.
TOPICS
What is a survey? What are the steps in a survey study?
How do I construct questions?
Validity
Reliability
WHAT IS A SURVEY? WHAT ARE THE STEPS IN A
SURVEY STUDY?
“A survey is a system for collecting information from or about people to describe, compare, or
explain their knowledge, attitudes, and behavior.” (Fink, 2003)
Steps:
 1. Define the objectives/goals*
 2. Design the study: population*, sample*, sample size*, timeline*, instrument construction
 3. Integrate validity and reliability into survey instrument development (including pilot testing)
 A. Review, test, revise
 B. Repeat as necessary
 4. After IRB approval: Administer the survey (internet, mail, mixed mode)*
 5. Data cleaning & management*
 6. Data analysis*
 7. Reporting results*
*Not today 
ADVANTAGES AND DISADVANTAGES OF SURVEYS
Advantages
Disadvantages
Anonymous
Wording can bias response
Inexpensive
Impersonal
Easy to compare and Analyze
Doesn’t get full story
Lots of people, lots of data
Low response rates (Consider the
Tailored Design Method by Don Dillman)
Use pre-existing instruments
Self-selection bias
Not generalizable
SOCIAL EXCHANGE
Establish Trust




Token of appreciation
Sponsorship of legitimate authority
Make task appear important
Invoke other exchange relationships
Increase Rewards
Reduce Social Costs
 Avoid: subordinate language, embarrassment, inconvenience
 Minimize requests for personal informaiton
WRITING QUESTIONS
Match survey questions to research objectives, goals, research questions, or
hypotheses!!!
Straight forward questions yield straight forward responses.
Avoid questions about which you are “just curious”
Include necessary demographics whenever possible (describe sample)
If you can use a previously developed survey instrument, do it!




Validity and Reliability already done
Get author permission, some cost money
Make sure it is accepted in the literature
Do NOT change it! (without the original author’s permission, but then you lose the previous validity and
reliability)
QUESTION DESIGN
Physical format/appearance
 Qualtrics is nice
 Visual layout is clean
Order of questions




Questions about announced subject first!
Order should be logical
Group items that are similar
Personal and demographic questions at the end
CAUTIONS
You get what you measure: Choose your questions carefully
Keep it simple: Be aware of the literacy level of your respondents
Shorter is better: Concise, easier to understand, easier to answer
Ranking is hard: And often misunderstood leading to invalid data
People don’t often read the instructions
Each concept gets its own question! (Beware of AND and OR)
Questions should be concrete and behavioral, not conceptual
Binary questions (yes/no) contain less information than ordinal questions (Hierarchy of information
content)
Define terms before asking questions
Avoid jargon and “loaded” words
QUESTION TYPES
Open vs. Closed
Levels of Measurement (Information)




Nominal (Sex, race/ethnicity)
Ordinal (Likert, Likert-type, unipolar ordinal)
Interval
Ratio
ORDINAL QUESTIONS
Ordinal
 Unipolar
 e.g. never, rarely, sometimes, often, always
 4 to 5 points
 Bipolar
 Likert and Likert-type
 strongly disagree, disagree, somewhat disagree, neither agree nor disagree, somewhat agree, agree, strongly agree
 Even vs. odd number of responses
 Left-side bias
Continuous as ordinal
 Down-coding age and income
VALIDITY
Face Validity
Content Validity
Criterion Related Validity
 Concurrent (correlated with a “gold standard”)
 Predictive (ability to forecast)
Construct Validity
 Convergent
 Divergent
FACE VALIDITY
Generally untrained judges
Does the instrument appear to measure the construct?
Very weak evidence
CONTENT VALIDITY
Definition: “The extent to which an instrument adequately samples the research
domain of interest when attempting to measure phenomena (Carmines and Zeller,
1979)




Begin with literature review to identify the entire domain of content
Develop items associated with the identified domain of content
Content Validity Index (Proportion of experts who agree item is relevant), Cohen’s Kappa
In mixed mode surveys, triangulate quantitative responses with opnen-ended responses
CRITERION RELATED VALIDITY
Compares new instrument against another instrument or predictor
Concurrent validity
 “Gold Standard”
 Known/published psychometric properties
 Published and used by researchers in the field
 Could choose test that measures “opposite”
 Correlation coefficient
Predictive validity
 Useful in predicting future behavior/events/attitudes/outcomes
 Correlation coefficient
CONSTRUCT VALIDITY
Theoretical relationship of new instrument with other constructs or behaviors
 New instrument correlates with other similar (not exact) measures (convergent validity) and does not
correlate with others (divergent validity
 New instrument discriminates one group from another
 Not always quantifiable
 Requires theory related to constructs
RELIABILITY
Type of Reliability*
Characteristics
Test-Retest
Consistency across time
Same respondents, short time lapse,
group of respondents
Alternate Form
Differently worded items to obtain
same information
Like different forms of SAT, requires
two entire forms
Internal Consistency
Groups of items measuring same
construct
Cronbach’s Alpha
Intraobserver
Same observer measures twice
Interobserver
More than one observer measures
same experimental unit
CRONBACH’S ALPHA
Internal Consistency
 Only appropriate when scale scores are developed
 Only use items that are thought to measure a single construct
Prefer scales to have Cronbach’s alphas greater than .7
If Cronbach’s alpha is greater than .9, are some items redundant?
Not generally appropriate for knowledge tests (unless they measure a single
construct)
Not appropriate when scale scores are not calculated
SCALING AND SCORING
Hypothesize which items form a scale
 Perform a factor analysis
 Eliminate items that crossload, fail to discriminate among respondents, strongly correlate with other
items, are frequently misunderstood or left blank
 Calculate Cronbach’s alpha, and alpha if item deleted
Interpretability is important!
 What does a big number mean, what does a small number mean?
 When you read the items do they make sense for a single construct
OTHER CONSIDERATIONS
Sample size requirements increase with:
 Decreasing levels of information
 Smaller effect size detection
Software for sample size and power analysis
Pilot testing is critical
 Helps identify errors in the survey
 Identifies potential design/redesign issues
 Predicts potential problems you might encounter
END NOTES
Consider applying for CTR-IN pilot grants!
 http://www.isu.edu/healthsciences/ichr/
Teri Peterson
 Best way to contact is through email: [email protected]
 Phone is less reliable (X5333 or X4861)