Smarter Balanced Assessments

Download Report

Transcript Smarter Balanced Assessments

Smarter Balanced Assessments
A COMPLETE CHANGE IN THE PARADIGM OF
STUDENT ASSESSMENT
Smarter Balanced stats
 29 states
 15 million students
 40,000 test items
 12 week testing window for the mandatory
summative test
 Not a timed test, likely shorter than current practice
 Fast turn around time on results
Why is technology needed?
 Computer adaptive questions (more on this later)
 Some CAT are interactive in nature (ex: draw something)
 Performance tasks require
 Stimulus





Information Processing tools and tasks






Video clips
Photos
2-D and 3-D modeling
databases
Data analysis software
Simulations
Spreadsheets
Equation editor tools
Chat rooms
Products or performances
Important terms to understand
 Computer adaptive testing
 Performance assessment
 Formative/interim assessment
 Summative assessment
Computer adaptive testing (CAT)
 Based on student responses, the computer program
adjusts the difficulty of questions throughout the
assessment.
 For example, a student who answers a question
correctly will receive a more challenging item, while
an incorrect answer generates an easier question.
 By adapting to the student as the assessment is
taking place, these assessments present an
individually tailored set of questions to each student
and can quickly identify which skills students have
mastered.
Weight-lifting analogy
 A “fixed” test is like lifting a 50 lb dumbbell
 Either you can or you can’t—doesn’t tell if you could
lift one 5 lbs more or less
 If everyone can lift a 5 lb dumbbell, doesn’t tell you
how much more you are able to lift
 CAT brings up the difficulty in small increments so
that it is clear precisely “how much weight” you can
handle
Validity
 Program is being designed using Artificial
Intelligence tools currently used for assessments
such as GRE, GMAT, SAT
 About 10% of the constructed and extended response
(includes written response of any length from short
answer to sentences to essays) will be checked by a
human scorer
Testing versus assessment
 T: Results in a quantifiable score that demonstrates
what students know or can do at a point in time
 A: Multiple measures over time that reflect mastery,
growth, depth of knowledge, gaps in learning
 T: Right or wrong answers
 A: Multiple correct answers are acceptable; Answers
illustrate student thinking and
understandings/misunderstandings of content
Testing versus assessment
 T: Occurs at the end of the learning process,
designed to measure/rank students
 A: Ongoing and actionable during the learning
process, designed to improve teaching/learning
 T: Evaluative—marks an end
 A: Diagnostic-opportunities to improve
Performance assessment
 Performance tasks challenge students to apply their
knowledge and skills to respond to complex real-world
problems.
 They can best be described as collections of questions
and activities that are coherently connected to a single
theme or scenario.
 These activities are meant to measure capacities such as
depth of understanding, writing and research skills and
complex analysis, which cannot be adequately assessed
with traditional assessment questions.
 No single correct answer
Interim/formative assessments
 Optional comprehensive and content-cluster assessment





to help identify specific needs of each student
Can be administered throughout the year
Provides clear examples of expected performance on
Common Core standards
Includes a variety of question types: selected response,
short constructed response, extended constructed
response, technology enhanced, and performance tasks
Aligned to and reported on the same scale as the
summative assessments
Fully accessible for instruction and professional
development
Summative assessments
 Assesses the full range of Common Core in English language
arts and mathematics for students in grades 3–8 and 11
(interim assessments can be used in grades 9 and 10)
 Measures current student achievement and growth
across time, showing progress toward college and career
readiness
 Can be given once or twice a year (mandatory testing window
within the last 12 weeks of the instructional year)
 Includes a variety of question types: selected response,
short constructed response, extended constructed response,
technology enhanced, and performance tasks
Item design
 All of the Common Core Standards will be clustered





into testable “claims”
Students will start with an entry level item aligned to
grade level standards
Items will get easier or harder depending on answer
Every student will walk away saying “this was a hard
test”
Every student will persist with testing until they get
half the items wrong
Clear, precise picture of student ability
Timeline
 School year 2011-12: test specifications, item
development, technology readiness tools
 March 2013: CMT/CAPT tests given for the last time
 Spring 2014: CSDE applying for waiver from
accountability testing in order to pilot versions of the
Smarter Balanced Assessments
 School year 2014-15: districts will be able to use
Smarter Balanced tools for formative and summative
assessments
Technology issues
 Testing will require access to Internet
 Decisions still pending on the variety of devices that
can be used (security, protocol issues,
authentications)
 There will be minimum hardware and software
requirements needed: attempt to make accessible to
older OS and less powerful machines
 Technology readiness tool will collect this
information
Technology Readiness Tool
 Degree to which hardware and software (devices) at
the school level meet the minimum requirements set
by each of the consortia.
 Adequate ratio of test taking devices to test takers
within a prescribed time table.
 Sufficient bandwidth to handle the expected volume
of traffic and content.
 Adequate number of trained personnel to support
the process.
System Architecture
 Resources and sample items will be open source
 Most of testing activity will occur in a cloud
environment
 Security will mimic protocols used by banking
industry
 Intent to deliver testing on tablets (either iPad or
Android), minimum 9.5” screen since many districts
are purchasing these-may need keyboards
 Tablets will be necessary for 25% of the math
assessment planned for 2016-17
New purchase specifications
 Hardware




1 GHz or faster processor
1 GB RAM or greater memory
9.5 inch (10 inch class) or larger screen size
1024 x 768 or better screen resolution
 Operating System






Windows 7
Mac 10.7
Linux (Ubuntu 11.10, Fedora 16)
Chrome OS
Apple iOS
Android 4.0
 Networking

Wired or wireless Internet connection
 Device Type

Desktops, laptops, netbooks, thin client, and tablets that meet the hardware, operating
system, and networking specifications
Security issues
 Every student’s test is likely to be somewhat different
 Test administration will likely take less time than a
prescribed window



Grade 3 Math 45-90 minute range for performance tasks and 45-60
minute range for extended response
Grade 4-11 Math 60-120 minute range for performance tasks and 4560 minute range for extended response
Tasks may cross days
 Whole group testing is unlikely; students are more likely
to be assessed in smaller groups over the 12 week window
for summative accountability testing
 Allowable teacher and peer interactions, group work
 Required tools and resources will be embedded within
online system; not free access to Internet for searches
Increasing rigor
 This is NOT CMT/CAPT given through a computer
 Because CAT changes the difficulty of questions, no
one will “get them all right”
 Assessments will yield a precise measurement of
what content students have mastered
 Scoring will be done with Artificial Intelligence
programs currently used for GRE, GMAT, and other
standardized assessments (only 10% human check)
Typical CMT test item
 http://www.csde.state.ct.us/public/cedar/assessmen
t/cmt/math_handbook.htm
 http://www.csde.state.ct.us/public/cedar/assessmen
t/cmt/l_a_handbook.htm
Typical Smarter Balanced test item
 http://www.smarterbalanced.org/wordpress/wp-
content/uploads/2012/02/DRAFT%20Math%20Ite
m%20Specifications.pdf
 http://www.smarterbalanced.org/wordpress/wpcontent/uploads/2012/02/DRAFT%20ELA%20Item
%20Specifications.pdf
 http://www.smarterbalanced.org/wordpress/wpcontent/uploads/2012/02/DRAFT%20Performance
%20Task%20Specifications.pdf
Sample question
 Old question: Randa ate 3/8 of a pizza, and Marvin ate 1/8 of the same







pizza. What fraction of the pizza did Randa and Marvin eat?
a. 5/8
b. 3/8
c. 1/4
d. 1/2
(Answer: d)
New question: Tito and Luis are stuffed with pizza! Tito ate one-fourth of
a cheese pizza. Tito ate three-eighths of a pepperoni pizza. Tito ate one-half
of a mushroom pizza. Luis ate five-eights of a cheese pizza. Luis ate the
other half of the mushroom pizza. All the pizzas were the same size. Tito
says he ate more pizza than Luis because Luis did not eat any pepperoni
pizza. Luis says they each ate the same amount of pizza. Who is correct?
Show all your mathematical thinking.
(Answer: Luis is right —
both ate 1 1/8 of a pizza).