Value-Added Systems Presentation to the ISBE Performance Evaluation Advisory Council Dr. Robert H.
Download ReportTranscript Value-Added Systems Presentation to the ISBE Performance Evaluation Advisory Council Dr. Robert H.
Value-Added Systems Presentation to the ISBE Performance Evaluation Advisory Council Dr. Robert H. Meyer Research Professor and Director Value-Added Research Center University of Wisconsin-Madison February 25, 2011 Attainment and Gain • Attainment – a “point in time” measure of student proficiency – compares the measured proficiency rate with a predefined proficiency goal. • Gain – measures average gain in student scores from one year to the next Attainment versus Gain Grade 3 Grade 4 Grade 5 Grade 6 Grade 7 Grade 8 Growth: Starting Point Matters Reading results of a cohort of students at two schools School 2006 Grade 4 Scale Score Avg. 2007 Grade 5 Scale Score Avg. Average Scale Score Gain A 455 465 10 B 425* 455* 30 Grade 4 Proficient Cutoff 438 Grade 5 Proficient Cutoff 463 *Scale Score Average is below Proficient Example assumes beginning of year testing 4 Value-Added • A kind of growth model that measures the contribution of schooling to student performance on the standardized tests • Uses statistical techniques to separate the impact of schooling from other factors that may influence growth • Focuses on how much students improve on the tests from one year to the next as measured in scale score points Value-Added Model Definition • A value-added model (VAM) is a quasi-experimental statistical model that yields estimates of the contribution of schools, classrooms, teachers, or other educational units to student achievement, controlling for non-school sources of student achievement growth, including prior student achievement and student and family characteristics. • A VAM produces estimates of productivity under the counterfactual assumption that all schools serve the same group of students. This facilitates apples-to-apples school comparisons rather than apples-to-oranges comparisons. • The objective is to facilitate valid and fair comparisons of productivity with respect to student outcomes, given that schools may serve very different student populations. A More Transparent (and Useful) Definition of VA • Value-added productivity is the difference between actual student achievement and predicted student achievement. • Or, value-added productivity is the difference between actual student achievement and the average achievement of a comparable group of students (where comparability is defined by a set of characteristics such a prior achievement, poverty and ELL status). In English Posttest = Post-on-Pre x Pretest Link Student School Unobserved + + + Characteristics Effects Factors Value Added VARC Philosophy • Development and implementation of a valueadded system should be structured as a continuous improvement process that allows for full participation of stakeholders • Model Co-Build; Complete customization – Analysis – Reporting • Value–added is one tool in a toolbox with multiple indicators VARC Value-Added Partners • • • • • • • • • • • • • • • • • Design of Wisconsin State Value-Added System (1989) Minneapolis (1992) Milwaukee (1996) Madison (2008) Wisconsin Value-Added System (2009) Milwaukee Area Public and Private Schools (2009) Racine (2009) Chicago (2006) Department of Education: Teacher Incentive Fund (TIF) (2006 and 2010) New York City (2009) Minnesota, North Dakota & South Dakota: Teacher Education Institutions and Districts (2009) Illinois (2010) Hillsborough County , FL (2010) Broward County, FL (2010) Atlanta (2010) Los Angeles (2010) Tulsa (2010) Districts and States working with VARC Minneapolis Milwaukee Madison Racine Chicago Los Angeles New York City Tulsa Atlanta Hillsborough County Broward County Measuring knowledge • Many factors influence what a student learns and how their knowledge is measured • A variety of measures, including (but not limited to) assessments, tell us what a student knows at a point in time. • What are some ways we measure knowledge? Measuring knowledge Large scale assessments MAP WKCE Daily teacher assessments Daily Journal Unit Project Local assessments used by the district Diagnostic Test End-of-course Exam Observations After-school Activities Hands-on Project The Simple Logic of Value-Added Analysis • School Value-Added Report – School specific data – Grade level value-added • Comparison Value-Added Reports – Compare a school to other schools in the district, CESA, or state – Also allows for grade level comparisons • Tabular Data available for School Report and Comparison Reports Attainment and Value-Added How complex should a value-added model be? • Rule: "Simpler is better, unless it is wrong.“ • Implies need for “quality of indicator/ quality of model” diagnostics. Model Features • • • • • • • • Demographics Posttest on pretest link Measurement error Student mobility: dose model Classroom vs. teacher: unit vs. agent Differential effects Selection bias mitigation: longitudinal data Test property analysis MAP vs. ISAT • MAP dates: September, January, May • MAP: uses Rasch equating – ISAT: 3PL • MAP: slightly higher reliability - ~0.96 in math, ~0.94 in reading – ISAT math ~0.93, reading ~0.9 • Cut scores on MAP are determined by equipercentile equating to ISAT Hillsborough FCAR Average Math Developmental Scale Scores means 2100 2000 1900 1800 1700 1600 1500 1400 1300 1200 1100 1000 02 03 04 05 06 07 08 09 10 11 GRADE_LEVEL cohort Cohort year is fall of first grade 1996 2002 1997 2003 1998 2004 1999 2005 2000 2006 2001 2007 12 Hillsborough Average FCAT Reading Developmental Scale Scores means 2000 1900 1800 1700 1600 1500 1400 1300 1200 1100 1000 900 800 700 02 03 04 05 06 07 08 09 10 11 GRADE_LEVEL cohort Cohort year is fall of first grade 1996 2002 1997 2003 1998 2004 1999 2005 2000 2006 2001 2007 12 Hillsborough Standard Deviation of FCAT Math Developmental Scale Scores stdevs 340 330 320 310 300 290 280 270 260 250 240 230 220 210 200 190 180 170 160 02 03 04 05 06 07 08 09 10 11 GRADE_LEVEL cohort Cohort year is fall of first grade 1996 2002 1997 2003 1998 2004 1999 2005 2000 2006 2001 2007 12 Hillsborough Standard Deviation of FCAT Reading Developmental Scale Scores stdevs 410 400 390 380 370 360 350 340 330 320 310 300 290 280 270 260 250 240 230 02 03 04 05 06 07 08 09 10 11 GRADE_LEVEL cohort Cohort year is fall of first grade 1996 2002 1997 2003 1998 2004 1999 2005 2000 2006 2001 2007 12 Minimal correlation between initial status and value-added Grade-Level Statewide Results Grade Mean Score N Reliability of ValueAdded SD of Score Subject State Math MN 3 59460 200.0 13.9 0.901 Math MN 4 58346 210.8 14.6 0.916 Math MN 5 57053 219.9 16.2 0.907 Math MN 6 52400 226.7 16.5 0.873 Math MN 7 47985 232.1 17.4 0.883 Math MN 8 44227 236.4 17.9 0.823 Math MN 9 26512 238.8 18.2 0.826 Grade-Level Statewide Results Grade Mean Score N Reliability of ValueAdded SD of Score Subject State Math WI 3 43289 199.9 13.2 0.820 Math WI 4 44140 209.3 13.7 0.842 Math WI 5 43822 217.3 14.8 0.849 Math WI 6 47004 222.7 15.2 0.836 Math WI 7 44549 228.4 16.0 0.837 Math WI 8 43246 233.1 16.8 0.865 Math WI 9 26427 234.0 17.7 0.862 Grade-Level Statewide Results Grade Mean Score N Reliability of ValueAdded SD of Score Subject State Reading WI 3 43139 194.8 15.1 0.736 Reading WI 4 43671 202.9 14.4 0.780 Reading WI 5 43668 209.7 13.8 0.737 Reading WI 6 46233 214.0 14.2 0.719 Reading WI 7 44616 218.3 14.0 0.792 Reading WI 8 43251 221.7 14.1 0.826 Reading WI 9 28066 223.3 14.4 0.843 MPS and MMSD Value-Added compared to Wisconsin 6th to 7th Grade (Nov 2006 – Nov 2007) Mathematics – State VA Model School Effects MPS School Effects MMSD School Effects School/District VA Productivity Parameters in WKCE Scale Score Units (Relative to State) Visit the VARC Website http://varc.wceruw.org/ for more information about VARC and value-added