Meeting NCLB Act: Students with Disabilities Who Are Caught in the Gap Martha Thurlow Ross Moen Jane Minnema National Center on Educational Outcomes http://education.umn.edu/nceo Sue Rigney U.S.

Download Report

Transcript Meeting NCLB Act: Students with Disabilities Who Are Caught in the Gap Martha Thurlow Ross Moen Jane Minnema National Center on Educational Outcomes http://education.umn.edu/nceo Sue Rigney U.S.

Meeting NCLB Act: Students
with Disabilities Who Are
Caught in the Gap
Martha Thurlow
Ross Moen
Jane Minnema
National Center on Educational Outcomes
http://education.umn.edu/nceo
Sue Rigney
U.S. Department of Education
http://www.ed.gov
CCSSO, 2004
Boston, MA
Session Plan
•
•
•
•
•
Background – NCLB and the “Gap”
Out-of-level testing realities
Current psychometric thinkings
What IS a state to do?!
Q & A Session
Purpose
“…to ensure that all children have a fair,
equal, and significant opportunity to obtain a
high-quality education and reach, at a
minimum, proficiency on challenging State
academic achievement standards and state
academic assessments”
Within a Content Area
AYP
combines
Other
Data
Test
Data
Safe
Harbor
Progress
Progress
over time
over time
% Proficient + Advanced
AND
95% tested
12 Years to 100% Proficient
Intermediate goals
Annual measurable objectives
States’ “Gap” Problem??
• General assessment without
accommodations
• General assessment with
accommodations
<<<< Something in between?>>>>
• Alternate assessment (≤ 1 proficient)
States’ Responses Have Been
Forms of
Out-of-Level Testing
• Off-Level
• Modified
Assessment
• Performance Level
• Challenge Down
• Levels Testing
• Alternate
Assessment
• Adaptive Testing
• Instructional Level
• Functional Level
• Alternative
Assessment
Definition Issues
• The administration of a test at a level that is
above or below the student’s grade level in
school
ASES SCASS (1999)
• Typically, only students with disabilities tested
below the grade in which their same-age peers
are enrolled
• States continue to use varying
definitions!
National Status of
Out-of-Level Testing
• For 2002 – 2003, 18 states tested out
of level in large-scale assessment
programs
New Pattern!
• Currently, 8 states have
discontinued or are doing so
One State’s Prevalence Data
Prevalence data available for both enrollment
grade and grade at which tested.
o Approximately 30% of special education
students were tested out of level in reading
and math, and approximately 20% in
writing
o How far below grade students were tested
spread as grade level increased:
Enrolled grade 8: 44% tested gr 6; 40% tested gr 4; 16%
tested gr 2
Enrolled grade 4: 68% tested gr 4; 32% tested gr 2
One State’s Data Interpretation
Performance data showed from 5% to 35%
of students performed at goal level on the
below grade level test – suggesting that
they probably should have been in a higher
grade level test
For example:
35% of grade 8 students tested on the grade 2 reading
test performed at goal level
5% of grade 8 students tested on the grade 6 math
test performed at goal level
One State’s Out-of-Level Test
Data Use
• Interpreted as TOO MANY tested at TOO
LOW of a test level
• Developed training with specific out-oflevel testing focus
• Challenged teachers to raise
assessment expectations!
NRT View at 50%
Performance Standards
8
7
6
5
4
3
3
4
5
6
7
Content Standards
8
NRT View at 25%
Performance Standards
8
7
6
5
4
3
3
4
5
6
7
Content Standards
8
Vertical Equating
Performance Standards
8
7
6
5
4
3
3
4
5
6
7
Content Standards
8
Vertical Equating Challenges
• Statistics
– Error variance: A score on one test equates
to a range of scores on another test.
• Content
– Construct differences: Could predict
mathematics scores from reading scores.
OOLT Alignment Study
• Nine states interviewed in Spring 2003
• Only two relate out-of-level test results to
enrolled grade performance
– One = psychometric support for single
grade difference
– One = human judgment with rubrics
Testing Without Aligning to
Enrolled Grade
• Cannot help meet AYP
• Is it a legitimate, humane way to meet
95% participation requirements or a
violation of NCLB intent?
• What can be done psychometrically?
Standards Based Assessment
Performance Standards
8
7
6
5
4
3
3
4
5
6
7
Content Standards
8
Performance Standards
Universal or Accommodated
Standards
Based
Assessment
8
7
6
5
4
3
3
4
5
6
7
Content Standards
8
Psychometric Limitations
• Psychometric improvements can only
remove barriers to seeing which students
are proficient
• Changing actual proficiency requires
something else
Research Findings Regarding
Out-of-Level Testing
• Number of research
studies conducted at
NCEO on out-of-level
testing
• The most recent:
–
–
–
–
Prevalence Study
Reporting Study
Alignment Study
Case Studies
NCLB and Out-of-Level Testing
• “In order to improve instruction and achievement
for all students with disabilities, the Department
expects States to assess as many students as
possible with academic assessments aligned to
regular achievement standards.”
• Out-of-level assessments aligned with alternate
achievement standards for students with the
most significant cognitive disabilities may be
considered alternate assessments (fall within 1.0
Percent Cap for proficient and advanced scores)
(Federal Register, 2003)
To Reduce the Use of Out-ofLevel Tests:
• Provide students access to the general
curriculum
• Develop universally designed
assessments
• Ensure that all teachers set high
expectations for all students and
understand the State’s academic
content standards
(Federal Register, 2003)
How do NCEO
recommendations align with
NCLB regulations?
• Introduce more appropriate assessments (i.e.,
universal design)
• Advocate for grade-level instruction and high
expectations for ALL students
• Clear and thorough reporting of all test results
and inclusion in accountability programs
• Call for reduction in out-of-level testing through
policy and implementation change
Easy to say …
… but what can we REALLY DO??!!
NCEO’s Responses at End of
Out-of-Level Testing Project
#1 – Identify students in “gap”
#2 – Restructure current thinking
#3 – Understand large-scale assessments
#4 – Use of accommodations
#5 – Raise teacher expectations
#6 – Provide high-quality training
#1 - States’ “Gap” Conundrum
• Who are these students?
• Can states identify these students?
• Similar subgroup of students across states?
• Change from year to year?
• Different students by grade, disability category,
general vs. special education?
#2 - Re-think at State, District,
and School Levels
 Not a child issue
 More than an assessment issue
 An INSTRUCTIONAL issue!
 Need to think critically about how to
augment instruction!
#3 - Understanding Purpose of
Large-Scale Assessments
Commonly used rationale for out-of-level testing
is the more precise and accurate use for
instructional decision making.
Thurlow & Minnema, 2001
But …
Time for a
Test!
NCLB does NOT require
student accountability
(e.g., graduation exams
to get diploma).
NCLB does require
SYSTEM level
accountability to
ensure all students
learn to high
levels.
#4 - Accommodations Use
 Anecdotal evidence of teacher
confusion about accommodations
 Little use of states’ accommodations
policies
 Now … real data!
Most Frequently Used in Testing
• Special education teachers said:
–
–
–
–
–
Extended (extra) time
Small group or individual administration
Test items read aloud
Directions read aloud
Alternate setting
However …
25% of teachers said they did not know
which accommodations were
considered nonstandard in their state.
n = ~ 750
% Not Knowing Standard vs.
Nonstandard by Accommodation
•
•
•
•
•
•
Read aloud (67%)
Calculator (67%)
Spell check or dictionary (46%)
Scribe (32%)
Visual cues on test (20%)
Extended/extra time (17%)
#5 - Raise Instruction AND
Assessment Expectations!
 Counter inconsistencies in practice
 Grade of enrollment drove out-of-level testing decision
 Language arts and math teachers used out-of-level test
criteria differently
 Think in terms of entire school system
Administrators
ALL school staff
Students
Family
#6 - Training, training, training!
 Content focused on basic, important information
“Our large-scale assessment program is a ‘blip’ on my professional
radar screen.”
Administrator, large urban school district, April, 2004
 Extend participant pool
 Don’t rely on “train the trainer” models only
 Experiment with new technologies!
 Combine general and special education +
practitioners + administrators
~ Thought shift happens! ~
 SEAs think critically about who these
students really are!
 Work across SEA divisions (e.g.,
assessment, special education, English
language learning)
 Update state policy in practitioner-friendly
format