Closing the Loop Workshop, by Raymond Zurawski

Download Report

Transcript Closing the Loop Workshop, by Raymond Zurawski

Closing the Loop

What to do when your assessment data are in September 21, 2005 Raymond M. Zurawski, Ph.D.

1

Step 7: Close the Loop (Use the Results) Step 8: Revise the Assessment Plan and Continue the Loop Step 1: Identify Program Goals Step 6: Report Findings And Conclusions

Cycle of Assessment

Step 2: Specify Intended Learning Outcomes (Objectives) Step 5: Analyze and Interpret the Data: (Make Sense of It All) Step 4: Implement: Data Collection Step 3: Select Assessment Methods

Assessment Methods

September 21, 2005 Raymond M. Zurawski, Ph.D.

3

Assessment Methods Used at SNC

    Examination of student work      Capstone projects Essays, papers, oral presentations Scholarly presentations or publications Portfolios Locally developed examinations Major field or licensure tests Measures of professional activity   Performance at internship, placement, sites Supervisor evaluations Miscellaneous Indirect Measures   Satisfaction/evaluation questionnaires Placement analysis (graduate or professional school, employment) September 21, 2005 Raymond M. Zurawski, Ph.D.

4

Other Methods Used

 Faculty review of the curriculum   Curriculum audit Analysis of existing program requirements  External review of curriculum  Analysis of course/program enrollment, drop out rates September 21, 2005 Raymond M. Zurawski, Ph.D.

5

What to Know About Methods

 Notice that different assessment methods may yield different estimates of program success  Measures of student self-reported abilities and student satisfaction may yield different estimates of program success than measures of student knowledge or student performance 

What are your experiences here at SNC?

 Good assessment practice involves use of multiple methods; multiple methods provide greater opportunities to use findings to improve learning September 21, 2005 Raymond M. Zurawski, Ph.D.

6

What to Know About Methods

 Even if the question is simply…  Are students performing…    …way better than good enough ?

… good enough ?

…NOT good enough ?

 The answer may depend on the assessment method used to answer that question September 21, 2005 Raymond M. Zurawski, Ph.D.

7

Implementation

September 21, 2005 Raymond M. Zurawski, Ph.D.

8

Implementation

 Common Problems     Methodological problems  Instrument in development; method misaligned with program goals Human or administrative error Response/participation rate problems  Insufficient numbers (few majors; reliance on volunteers, convenience sample; poor response rate); insufficient incentives, motivation High “costs” of administration  “Other” ( no assessment, no rationale)  NOTE: Document the problems; provides one set of directions for ‘closing the loop’ September 21, 2005 Raymond M. Zurawski, Ph.D.

9

Document Your Work!

 “If you didn’t document it, it never happened…” The clinician’s mantra 10 September 21, 2005 Raymond M. Zurawski, Ph.D.

Analyzing and Interpreting Data

September 21, 2005 Raymond M. Zurawski, Ph.D.

11

Analyzing and Interpreting Data

 General Issues   Think about how information will be examined, what comparisons will be made, even before the data are collected Provide Descriptive information     Percentages (‘strongly improved’, ‘very satisfied’) Means, medians on examinations Summaries of scores on products, performances Provide Comparative information    External norms, local norms, comparisons to previous findings Comparisons to Division, College norms Subgroup data (students in various concentrations within program; year in program) September 21, 2005 Raymond M. Zurawski, Ph.D.

12

Interpretations

   Identify patterns of strength Identify patterns of weakness Seek agreement about innovations, changes in educational practice, curricular sequencing, advising, etc.

that program staff believe will improve learning September 21, 2005 Raymond M. Zurawski, Ph.D.

13

Ways to Develop Targeted Interpretations

 What questions are most important to you? What’s the story you want to tell?

 helps you decide how you want results analyzed  Seek results reported against your criteria and standards of judgment so you can discern patterns of achievement September 21, 2005 Raymond M. Zurawski, Ph.D.

14

September 21, 2005

Interpreting Results in Relation to Standards

 Some programs establish target criteria  Examples   If the program is effective, then 70% of portfolios evaluated will be judged “Good” or “Very good” in design The average alumni rating of the program’s overall effectiveness will be at least 4.5 on a 5.0 point scale Raymond M. Zurawski, Ph.D.

15

Standards and Results: Four Basic Relationships

 Four broad relationships are possible:  A standard was established that students met  A standard was established that students did not meet   No standard was established The planned assessment was not conducted or not possible  Some drawbacks to establishing target criteria    Difficulties in picking the target number Results exceeding standard do not justify inaction Results not meeting standard do not represent failure September 21, 2005 Raymond M. Zurawski, Ph.D.

16

Reporting Results

September 21, 2005 Raymond M. Zurawski, Ph.D.

17

Reporting Assessment Findings

 Resources  An Assessment Workbook  Ball State University  Another Assessment Handbook  Skidmore College  An important general consideration:  Who is your audience?

September 21, 2005 Raymond M. Zurawski, Ph.D.

18

Sample Report Formats

      Skidmore College Old Dominion University Ohio University George Mason University Montana State University (History)  Other programs Institutional Effectiveness Associates, Inc.

September 21, 2005 Raymond M. Zurawski, Ph.D.

19

Local Examples of Assessment Reports

   Academic Affairs Divisions    Division of Humanities and Fine Arts Division of Natural Sciences Division of Social Sciences Student Life Mission and Heritage 20 September 21, 2005 Raymond M. Zurawski, Ph.D.

OK, but just

HOW

do I report…

 Q: How to report….

 Survey findings, Major Field Test data, Performance on Scoring Rubrics, etc.

 A: Don’t Reinvent the Wheel  Consult local Assessment and Program Review reports for examples September 21, 2005 Raymond M. Zurawski, Ph.D.

21

Closing the Loop: Using Assessment Results

September 21, 2005 Raymond M. Zurawski, Ph.D.

22

Closing the Loop: The Key Step

   To be meaningful, assessment results must be studied, interpreted, and used Using the results is called “closing the loop” We conduct outcomes assessment because the findings can be used to improve our programs Raymond M. Zurawski, Ph.D.

23 September 21, 2005

September 21, 2005

Closing the Loop

 Where assessment evaluation together… come   and Assessment:  Gathering, analyzing, and interpreting information about student learning Evaluation  Using assessment findings to improve institutions, divisions, and departments  Upcraft and Schuh Raymond M. Zurawski, Ph.D.

24

Why Close the Loop?

    To Inform Program Review To Inform Planning and Budgeting To Improve Teaching and Learning To Promote Continuous Improvement (rather than ‘inspection at the end’) 25 September 21, 2005 Raymond M. Zurawski, Ph.D.

Steps in Closing the Assessment Loop

     Briefly report methodology for each outcome Document where the students are meeting the intended outcome Document where they are not meeting the outcome Document decisions made to improve the program and assessment plan Refine assessment method and repeat process after proper time for implementation September 21, 2005 Raymond M. Zurawski, Ph.D.

26

Ways to Close the Loop

    Curricular design and sequencing Restriction on navigation of the curriculum Weaving more of “x” across the curriculum Increasing opportunities to learn “x” September 21, 2005 Raymond M. Zurawski, Ph.D.

27

Additional Ways to Close the Loop

    Strengthening advising Co-designing curriculum and co-curriculum Development of new model of teaching and learning based on research or others’ practice Development of learning modules or self paced learning to address typical learning obstacles September 21, 2005 Raymond M. Zurawski, Ph.D.

28

And don’t forget…

 A commonly reported use of results is to refine the assessment process itself    New or refined instruments Improved methods of data collection (instructions, incentives, timing, setting, etc.) Changes in participant sample  Re-assess to determine the efficacy of these changes in enhancing student learning.

September 21, 2005 Raymond M. Zurawski, Ph.D.

29

A Cautionary Tale

 Beware the Lake Woebegone Effect  …where all the children are above average… 30 September 21, 2005 Raymond M. Zurawski, Ph.D.

September 21, 2005

A Cautionary Tale

  When concluding that…

no changes are necessary at this time…

Standards may have been met but…   There may nonetheless be many students failing to meet expectations  How might they be helped to perform better?

There may nonetheless be ways to improve the program Raymond M. Zurawski, Ph.D.

31

Facilitating Use of Findings

 Laying Appropriate Groundwork     Assessment infrastructure Conducive policies Linking assessment to other internal proceses  (e.g., planning, budgeting, program review, etc,) Establish an annual assessment calendar September 21, 2005 Raymond M. Zurawski, Ph.D.

32

Factors that Discourage Use of Findings

   Failure to inform relevant individuals about purposes and scope of assessment projects Raising concerns and obstacles over unimportant issues Competing agendas and lack of sufficient resources September 21, 2005 Raymond M. Zurawski, Ph.D.

33

What

You

Can Do (Fulks, 2004)

      Schedule time to record data directly after completing the assessment.

Prepare a simple table or chart to record results.

Think about the meaning of these data and write down your conclusions.

Take the opportunity to share your findings with other faculty in your area as well in those in other areas.

Share the findings with students, if appropriate.

Report on the data and what you have learned at discipline and institutional meetings. September 21, 2005 Raymond M. Zurawski, Ph.D.

34

Group Practices that Enhance Use of Findings

   Disciplinary groups’ interpretation of results Cross-disciplinary groups’ interpretation of results (library and information resource professionals, student affairs professionals) Integration of students, TAs, internship advisors or others who contribute to students’ learning September 21, 2005 Raymond M. Zurawski, Ph.D.

35

External Examples of Closing the Loop

    University of Washington Virginia Polytechnic University St. Cloud State University Montana State University (Chemistry) 36 September 21, 2005 Raymond M. Zurawski, Ph.D.

Closing the Loop: Good News!

Many programs at SNC have used their results to make program improvements or to refine their assessment procedures

September 21, 2005 Raymond M. Zurawski, Ph.D.

37

Local Examples of Closing the Loop

  See HLC Focused Visit Progress Report Narrative on OIE Website See Program Assessment Reports and Program Review Reports on OIE Website 38 September 21, 2005 Raymond M. Zurawski, Ph.D.

Ask your colleagues in … about their efforts to close the loop

          Music Religious Studies Chemistry Geology Business Administration Economics Teacher Education Student Life Mission and Heritage Etc.

September 21, 2005 Raymond M. Zurawski, Ph.D.

39

One Example of Closing the Loop

 Psychology     Added capstone in light of curriculum audit Piloting changes to course pedagogy to improve performance on General Education assessment Established PsycNews in response to student concerns about career/graduate study preparation Replaced pre-test Major Field Test administration with a lower cost, reliable and valid externally developed test September 21, 2005 Raymond M. Zurawski, Ph.D.

40

Conclusions

September 21, 2005 Raymond M. Zurawski, Ph.D.

41

Conclusions

 Programs are relatively free to choose which aspects of student learning they wish to assess  Assessing and reporting matter, but . . .

 Taking action on the basis of good information about real questions is the best reason for doing assessment

Conclusions

 The main thing…  …is to keep the main thing…  … the main thing!

September 21, 2005 Douglas Eder, SIU-E Raymond M. Zurawski, Ph.D.

43

Conclusions

    It may be premature to discourage the use of any method It may be premature to establish a specific target criteria It may be premature to require strict adherence to a particular reporting format Remember that sample reports discussed here are examples not necessarily models September 21, 2005 Raymond M. Zurawski, Ph.D.

44

Oh, and by the way…

Document Your Work!

September 21, 2005 Raymond M. Zurawski, Ph.D.

45

Additional Resources

 Internet Resources for Higher Education Outcomes Assessment (at NC State) September 21, 2005 Raymond M. Zurawski, Ph.D.

46

Concluding Q & A:

A One-Minute paper  What remains most unclear or confusing to you about closing the loop at this point?

47 September 21, 2005 Raymond M. Zurawski, Ph.D.