Student Assessment and Data Analysis Oakland Schools MAEDSTammy L. Evans Why are educators so fired up about data?  How do we know if teachers are.

Download Report

Transcript Student Assessment and Data Analysis Oakland Schools MAEDSTammy L. Evans Why are educators so fired up about data?  How do we know if teachers are.

Student Assessment
and Data Analysis
Oakland Schools
MAEDS
2005
Tammy L. Evans
Why are educators
so fired up about data?
 How do we know if
teachers are teaching our
Superintendents ask…
curriculum?
 How do we maximize the
value of dollars spent for
assessment and data
management?
 Are all of our students
achieving at acceptable
levels?
MAEDS 6 October 2005
2
Professional learning communities ask
 What is it we want
our students to know
and be able to do?
 How will we know
when they have
learned it?
 What will we do
when students are not
learning?
MAEDS 6 October 2005
3
Why are educators so fired up
about “data”?
Improving
Student
Achievement!
MAEDS 6 October 2005
4
Creating some common language
about data in schools
 What are the major
systems?
 How are they
related?
 What have districts
done?
 Where do we want
to go?
MAEDS 6 October 2005
5
4 Major Data & Technology Systems in Schools
Oakland Schools focus is on Assessment and Analysis
Student Information Systems
Assessment Systems
Data warehouse
Data
analysis
systems
MAEDS 6 October 2005
6
(see Data warehouse PP on CD)
SAS DAT PURPOSE
Student Assessment System & Data Analysis Tool
 Improve teaching and increase learning
for all
 Useful reports for teachers, principals and
district administration
 Common assessments tied to GLCEs
 Item banks tied to GLCEs
 Multiple district on-ramps
MAEDS 6 October 2005
7
What is an Assessment System?
 Tool for gathering
achievement
information
 It is assessing what
is going on in
classrooms.
MAEDS 6 October 2005
8
Who needs what data?
A single assessment cannot meet all needs.
 Administrators, public,
legislators
– Evaluation
– Accountability
– Long range planning
 Teachers, parents, students
– Diagnosis
– Prescription
– Placement
– Short range planning
– Very specific ach info
e.g., What percent met standards
on 4th grade MEAP math?
Are students doing better this year
than they were doing last year?
Large Grain Size
MAEDS 6 October 2005
e.g., Who understood this
concept?
Why is Becky
having trouble reading?
Fine Grain Size
9
Oakland Schools’ Path
to Student Achievement
 Fall 2004 – Meetings with focus groups,
create RFP
 Oct 2004 – Meeting with Assessment,
Curriculum and Technology directors from
Oakland districts to discuss requirements,
including multiple “on ramps”
 June 2005 deadline
MAEDS 6 October 2005
10
The RFP
 Input gathered from LEA focus groups in
Curriculum, Assessment, Instruction and
Technology
 RFP authored at Oakland Schools through a
collaboration between Career Focused Education,
Learning Services, Research Evaluation and
Assessment, Purchasing, School Quality and
Technology Services.
 Draft copy provided to LEA Technology and
Assessment Directors for input.
 Click here for details of the RFP
 Click here for details of the vendor pricing
submitted
MAEDS 6 October 2005
11
The Committee
 OCSA charged Oakland Schools and LEAs to
move forward on acquisition of assessment and
analysis system.
 The RFP evaluation committee was formed,
consisting of ISD and LEA staff representing
Assessment, Curriculum and Technology.
 Representatives from OCREAC, Teaching and
Learning Council, Oakland County Technology
Directors, OCSA Instruction &Technology
subcommittee.
 Committee members were from Berkley, Huron
Valley, Lamphere, Lake Orion, Troy, Novi, South
Lyon, Walled Lake and West Bloomfield.
MAEDS 6 October 2005
12
ISD Collaboration
 Jan 2005 – Oakland Schools and Wayne RESA
met to review strategic goals around assessment
and data analysis.
 Joint RFP was created
 Wayne RESA joined RFP evaluation committee
 Wayne RESA and Oakland Schools separated
scoring and recommendation for individual needs
and approvals.
MAEDS 6 October 2005
13
The evaluation begins
 10 vendors responded to the RFP
 The committee met to review the
responses.
 The committee chose three vendors for
demonstrations
 Click here for the Debriefing Voting
Results.
MAEDS 6 October 2005
14
The demonstrations
• Vendors were asked to cover specific points.
• Half day demonstrations for each vendor were
held at Farmington Training Center on March 10
& 11, 2005.
 All Oakland Schools LEAs were invited to send
representatives to the demonstrations.
 Over 100 participants reviewed the products and
were asked to complete a survey.
– Click here for the Survey results.
MAEDS 6 October 2005
15
Further evaluation
 After the demonstrations, the committee met to
discuss the products and created a pros/cons list
for each vendor.
 Using an audience response system, the group
prioritized the functionality of the products and
rated each vendor on those functional areas. (see
SAS-DAT PP on CD for full presentation.)
 Click here for the Functionality Summary
MAEDS 6 October 2005
16
MAEDS 6 October 2005
17
MAEDS 6 October 2005
18
Vendor References
 A subcommittee was formed to conduct reference
interviews.
 Included committee members from Huron Valley,
South Lyon, Walled Lake and West Bloomfield
and Oakland Schools
 Plato – two references, EduSoft – two references,
Pearson – three references
 Click here for the Reference Questions
 The reference information was synthesized and
presented to the committee on April 11.
 Click here for the Reference Call Summary
MAEDS 6 October 2005
19
Further Analysis
 Reviewed goals of RFP
 Reviewed priority & ranking from vendor
demonstrations
 Reviewed vendor reference calls
 Reviewed pricing
MAEDS 6 October 2005
20
The Evaluation
 Filled out evaluation sheets
– Click here for the Evaluation Form
 Results tallied:
– Plato
– EduSoft
– Pearson
MAEDS 6 October 2005
4680
4350
5720
21
Site Visit
 May 4, 2005 – Putnam City Schools, OK
 Met with Curriculum Director, principals
to review product in use.
MAEDS 6 October 2005
22
Facilitated Product
Demonstration
 May 5, 2005 – Oakland Schools
 SAS-DAT Committee members were
invited to participate in a test drive of
Benchmark and Inform.
MAEDS 6 October 2005
23
Principal’s Dashboard
Key Feature
At school & classroom levels,
every bar in a graph links to
student names and information
Pearson Inform helps you
target assistance – provide
early2005
intervention 24
MAEDS 6 October
Teacher’s Dashboard
Key Feature
Pearson Inform provides “Concept Analysis”
at District, School, and Class Views …
A big help in planning
instruction, aligning
curriculum, and
identifying student needs
MAEDS 6 October 2005
25
Parent’s / Student’s Dashboard
Pearson Inform’s
Parent Access &
Family Views
MAEDS 6 October 2005
26
Oakland Schools Support
 Models defined to support diverse needs of
districts and multiple on-ramps
 Monetary support
 Curriculum, Item Banks, and Assessments
delivered to all districts
MAEDS 6 October 2005
27
The Partnership
 Created Benchmark “Lite”
– Host for Oakland Schools’
• Standard curriculum
• Units / Lesson plans
• Assessments
–
–
–
–
MCF – Michigan Curriculum Framework
Common assessments tied to GLCEs
Item banks tied to GLCEs
Allows districts to create assessments
 Benchmark “Full”
– administer tests (scan or web based)
– report scoring
 Inform
– Analyzes test responses down to the individual student
MAEDS 6 October 2005
28
Where we are now…
 Conversion for 27 of 29 districts
 Training
 Implementation! August 2005+
 Sharing experience with other MI
districts.
– Contract allows for state purchase
– Increased participation reduces cost for all
MAEDS 6 October 2005
29
MACUL 2006
Presentation will
cover…
 Success stories
 Lessons Learned
 Examples of
classroom
assessment
 Examples of analysis
 Website and
demonstration
MAEDS 6 October 2005
30
Questions
MAEDS 6 October 2005
31