Taking the Guesswork out of Writing Better Multiple Choice Questions Presented to AMATYC February 2, 2015 by Patricia L.

Download Report

Transcript Taking the Guesswork out of Writing Better Multiple Choice Questions Presented to AMATYC February 2, 2015 by Patricia L.

Taking the Guesswork out of
Writing Better Multiple Choice Questions
Presented to AMATYC
February 2, 2015
by
Patricia L. Gregg, PhD (Patti)
Georgia Perimeter College
Multiple choice tests have their place …
… but it isn’t necessarily “everywhere”
Question “stem”
Should give a clear indication of what
the question is about
Don’t be too broad or generic
(e.g. Which of the following is true?)
Phrase in form of a question
Don’t use incomplete statements, fill in
the blank, etc.
Be consistent with grammar and
syntax
Don’t give clues about right answer
based on the wording of the stem
Use simple and positive constructions Avoid convoluted and negative
(e.g. Name the most common cause of …) constructions
(e.g. Name the least uncommon cause of …)
Use scenarios when possible but
Don’t include a lot of distracting or
irrelevant information (red herrings)
Answer options
Categories should be mutually
exclusive
Avoid overlapping categories
Have one clear right (or best) answer
Avoid “all of the above” “none of the
above” “both a and c” etc.
Be consistent with grammar and
syntax
Don’t give clues about right answer
based on the wording of the options
Should be same length and
construction
Don’t make the right answer longer or
shorter than other options, or phrased
differently
Should be equally plausible
No obvious wrong answers
Aren‘t plausible distractors “trick questions”?
Plausible distractors help to
• minimize lucky guesses
• ensure that you are truly measuring students’
learning, not their ability to recognize blatantly
irrelevant information
• diagnose key concepts that students may be
misunderstanding or confusing with one another
Poll #1
ITEM ANALYSIS can help you understand
• which of your incorrect answer options are
plausible “distractors”
• whether your test question effectively
“discriminates” between high-performing and
low-performing students
Can we test higher order thinking
with multiple choice questions?
Yes, but we need to
be vigilant!
Too often we claim to be
assessing application,
analysis, interpretation, etc.
but we are really assessing
recognition, computation or
calculation.
• Not all word problems test higher order skills!
• Examples of question items that could measure
higher order skills would require students to:
 Interpret data or graphs
 Recognize spurious
conclusions
 Distinguish between relevant
and irrelevant information
_____triangles have three sides of equal length.
a.
b.
c.
d.
Isosceles
Equilateral
Right
Scalene
Leading with the blank is an awkward
construction, especially for non-native English
speakers and others with reading challenges.
Which type of triangle has
all three sides of equal length?
a. Isosceles
b. Equilateral
c. Right
d. Scalene
Adapted from Cheung & Bucat, How Can We Construct Good Multiple Choice Items?
This question tests the same knowledge as the previous
example, but the construction is more familiar.
The mean of a distribution of test scores is the:
a. Most frequently occurring score
b. 50th percentile
c. Arithmetic average
d. Measure of score range
A university developed an aptitude test to use for
admission into its Honors Program. The test was
administered to a group of seven applicants who
obtained the following scores: 70, 72, 72, 80, 89, 94, 98.
The mean score on the aptitude test is:
a. 72
c. 82
b. 80
d. 90
from Zimmaro, Writing Good Multiple-Choice Exams
In order to correctly answer this question, the student must
still know the definition of “mean” but also must be able to
calculate the mean. However, the question does not assess
higher order skills such as application or interpretation.
We want students to see the relevance
of course content, but too many details in
a word problem end up testing reading
speed and comprehension, instead of
math skills!
Suppose you are a mathematics professor who wants to
determine whether or not your teaching of a unit on
probability has had a significant effect on your students.
You decide to analyze their scores from a test they took
before the instruction and their scores from another exam
taken after the instruction. Which of the following t-tests is
appropriate to use in this situation?
a. Dependent samples.
b. Heterogenous samples.
c. Homogenous samples.
d. Independent samples.
When analyzing your students’ pretest and
posttest scores to determine if your teaching has
had a significant effect, an appropriate statistic
to use is the t-test for:
a.
b.
c.
d.
Dependent samples.
Heterogenous samples.
Homogenous samples.
Independent samples.
Burton et. al., How to Prepare Better Multiple Choice Items: Guidelines for
University Faculty
cited in Duquesne University, Good, Better, Best: Multiple Choice Exam Construction
Test items with carefully constructed
answer options can help you diagnose
what concepts your students may be
confusing and/or misunderstanding.
The number of bacteria N(t) present in
a culture at time t hours is given by
N (t) = 2200(2)t. Find the number of
bacteria present when t = 4 hours.
(a) 35200
(b) 17600
(c) 70400
(d) 19360000
Test item courtesy of a GPC faculty member
The correct answer is (a).
A student would select (b) if they multiplied 2 x 4
instead of raising 2 to the 4th power.
A student would get (c) if they multiplied 2200 by
32, thinking that (2)4 is 32 rather than 16.
The student who selects (d) demonstrates no
understanding of the order of operation. In this
case, the student multiplies 2200 by 2 and then
raises it to 4. Since that is not one of the choices,
19360000 (4400 raised to 2) is the closest.
http://www.act.org/compass/sample/pdf/numerical.pdf
How could this item be stronger?
Do we really need 5 answer options?
The COMPASS publishers probably do this to
minimize the probability of guessing the correct
answer, BUT it is really hard to come up with 5
plausible alternatives, so if one is blatantly wrong,
we may as well just have 4 options.
Do all the answer options contribute to
diagnosis?
It seems unlikely that a student would
select 0% or 45% unless they had almost
no grasp of simple arithmetic, even if they
didn’t understand how to calculate
percentages.
Does the number of math courses taken
contribute to the strength of the question?
It may be an unnecessary complication for weaker
readers.
The skill being testing by this item is calculation of
percentages.
We want the student to demonstrate their
understanding that the second group is 60% of
25% not 60% of the total class; therefore the last
remaining group is 40% of 25%.
One possible re-write:
This year, 75% of the graduating class of
Harriet Tubman High School had been
accepted to college. Of the remaining class
members, 60% had applied to college but had
not been accepted. What percent of the
graduating class did not apply to college?
The calculation is exactly the same, but the red
herrings are removed.
http://www.act.org/compass/sample/pdf/numerical.pdf
Poll #2
How could this item be stronger?
Ensure that all answer options contribute to
diagnosis.
Eliminate extraneous information
e.g. exercise 3 times per week for 20 minutes.
IF this test item were intended to measure
higher order skills, we would keep the exercise
information and possibly add some additional
“what if” scenarios with different variables.
PAPERS
Burton, Sudweeks, Merrill & Wood (1991, 33 pp.) https://testing.byu.edu/handbooks/betteritems.pdf
Cheung & Bucat (2002, 10 pp.) http://www3.fed.cuhk.edu.hk/chemistry/files/constructMC.pdf
Zimmaro (2010, 41 pp.)
https://ctl.utexas.edu/sites/default/files/documents/Writing-Good-Multiple-Choice-Exams-04-28-10.pdf
INSTITUTIONAL WEBSITES
Brigham Young University:
https://testing.byu.edu/handbooks/14%20Rules%20for%20Writing%20Multiple-Choice%20Questions.pdf
Duquesne University: http://www.duq.edu/about/centers-and-institutes/center-for-teaching-excellence/teaching-andlearning/multiple-choice-exam-construction
Indiana University: http://www.iub.edu/~best/pdf_docs/better_tests.pdf
University of Texas: http://ctl.utexas.edu/teaching/assessment/planning/multiple-choice
Vanderbilt University: http://cft.vanderbilt.edu/guides-sub-pages/writing-good-multiple-choice-test-questions/
COMMERCIAL PUBLISHERS
Censeo Corporation, Guidelines for Writing Effective Tests: A Practical “Short Course” for Test Authors (2007, 9 pp).
Document not available online; appears company may no longer be in business; I have a PDF copy to share.
Magna Publications:
http://www.facultyfocus.com/articles/educational-assessment/seven-mistakes-avoid-writing-multiple-choice-questions/
Patricia (Patti) Gregg
Associate Director, Assessment and Reporting
Georgia Perimeter College
[email protected]
678-891-2571
@DrPattiG
www.linkedin.com/pub/patricia-gregg