OSCEs - IDEAL Consortium
Download
Report
Transcript OSCEs - IDEAL Consortium
OSCE
Kathy Boursicot
Train the Trainer Assessment Workshop
October 29, 2003
Hong Kong International Consortium
OSCE
Format
Purpose
Advantages
Writing principles
Training observers
Scoring considerations
What is an OSCE?
Objective
Structured
Clinical
Examination
Harden RG and Gleeson FA Assessment of clinical competence using an
objective structured clinical examination (OSCE) Medical Education,1979,
Vol 13: 41-54
OSCE test design
Observed Stations:
clinician examiners
Varieties of OSCEs
Patient-based
Traditional OSCE
SP-based test
Station couplets
Integral consultations
Clinical task
Written task
….
….
….
….
OSCE
Format
Purpose
Advantages
Writing principles
Training observers
Scoring considerations
Professional authenticity
Simple model of competence
Does
Shows how
Knows how
Knows
Miller GE. The assessment of clinical
skills/competence/performance. Academic
Medicine (Supplement) 1990; 65: S63-S7.
Testing formats
Behaviour~
attitude/skills
Professional practice
Does
Shows how
Cognition~
knowledge
Knows how
Knows
OSCEs
EMQs, SEQs
MCQs
OSCE - Objective
All the candidates are presented with the
same test
Specific skill modalities are tested at each
station
History taking
Explanation
Clinical examination
Procedures
OSCE - Structured
The marking scheme for each station is
structured
Structured interaction between examiner
and student
OSCE – Clinical Examination
Test of performance of clinical skills
candidates have to demonstrate their skills,
not just describe the theory
OSCE
Format
Purpose
Advantages
Writing principles
Training observers
Scoring considerations
Characteristics of assessment instruments
Utility =
Reliability
Validity
Educational impact
Acceptability
Feasibility
Van der Vleuten, C. The assessment of professional competence:
developments, research and practical implications,
Advances in Health Science Education, 1996, Vol 1: 41-67
Test characteristics
Reliability of a test / measure
reproducibility
of scores across
raters, questions, cases, occasions
capability
to differentiate consistently
between good & poor students
Sampling
Domain of Interest
Test Sample
Test Sample
Reliability
Competencies are highly domain-specific
Broad sampling is required to obtain
adequate reliability
across content, i.e., range of cases/situations
across other potential factors that cause error
variance, i.e.,
testing time, number of cases, examiners, patients,
settings, facilities
Test characteristics
Validity of a test / measure
The content is deemed appropriate by the
relevant experts
The test measures the characteristic (e.g.
knowledge, skills) that it is intended to
measure
The performance of a particular task predicts
future performance
Test characteristics
Validity of a test / measure
OSCE
Format
Purpose
Advantages
Writing principles
Training observers
Scoring considerations
Advantages of using OSCEs in clinical assessment
Careful specification of content = Validity
Observation of wider sample of activities =
Reliability
Structured interaction between examiner & student
Structured marking schedule
Each student has to perform the same tasks =
Acceptability
OSCE
Format
Purpose
Advantages
Writing principles
Training observers
Scoring considerations
OSCE Station Writing
How to start
Decide what tasks you
want to
can
should
test in an OSCE format
OSCEs test performance, not knowledge
Constructive alignment
Need to know the learning objectives of
your course / programme
Map these across :
Subject areas
Knowledge areas
Skill areas
Blueprinting
Content of the assessment should align
with the learning objectives of the course
Blueprinting
allows mapping of test items to specific
learning outcomes
ensures adequate sampling across subject
area and skill domains
OSCE blueprint: systems-based
Hx taking
(incl. diag)
CVS
Endocrine
Gastro
H&N
Haem & LN
Musculoskl
etc
Phys exam
(incl. diag)
Procedures
Counseling/E
ducation
Ordering
investigs
OSCE blueprint: discipline-based
Hx taking
(incl. diag)
Anaes & CC
Clin Pharm
Comm Health
Emergency med
Family med
Musculoskel
etc
Phys exam
(incl. diag)
Procedures
Counseling/E
ducation
Ordering
investigs
Key features of success in designing OSCEs
Feasibility
Congruence
Feasibility
Is it a reasonable task to expect the
candidates to perform?
Can the task be examined at an OSCE
station?
Can the task be performed in the time
allowed?
Feasibility
Is it a reasonable task to expect the candidates to
perform? Is it authentic?
Can the task be examined at an OSCE station?
Match clinical situations as closely as possible
Some tasks may require simulated patients
Some tasks may require manikins
Some tasks simply cannot be examined in this format
Feasibility
Can task be performed in time allowed?
Pilot the stations to see if they are feasible
Check equipment /helpers/practicalities
Congruence
Is it testing what you want it to test?
Station construct: describe what station
is testing
Congruence
Ensure that all parts of station coordinate
Candidate instructions
Marking schedule
Examiner instructions
Simulated patient instructions
Equipment
Station construct
This station tests the candidates ability
to …………………………
Candidate instructions
State circumstances: e.g. outpatient clinic, ward,
A & E, GP surgery
Specify the task required of the candidate: e.g.
take a history, perform a neurological
examination of the legs, explain a diagnosis
Specify tasks NOT required
Instruct on summing up: e.g. tell the patient, tell
the examiner
Examiner instructions
Copy of candidate instructions
Specific instructions appropriate to the task:
e.g., do not prompt, explicit prompts, managing
equipment
Simulated patient instructions
Give as much detail as possible so they can be
consistent
try to leave as little as possible for them to ad lib!
Give enough information to enable them to
answer questions consistently
Be specific about affect in each role
Specify patient demographics
i.e., gender, age, ethnicity, social class, etc.
Marking schedule
Ensure marks are allocated for tasks the
candidates are asked to perform
Decide relative importance of diagnosis vs
process (history taking, examination)
Separate checklist for process skills
Equipment
Be detailed
Think of
Chairs + table / couch / bench
Manikins - specify
Medical equipment
Stethoscope, ophthalmoscope, sphyg,
suturing materials, etc
Designing stations
Use your blueprint
Be clear what you are testing: define the
construct
Check for congruence
Pilot for feasibility
OSCE
Format
Purpose
Advantages
Writing principles
Training observers
Scoring considerations
Training observers
Understand the principles of OSCEs
Enhance inter-rater consistency
Techniques
Examiners must train together
Videos
‘live’ stations
Discussion of marking inconsistencies
Training observers
General training
Station-specific training
OSCE
Format
Purpose
Advantages
Writing principles
Training observers
Scoring considerations
Scoring considerations
Global vs checklist scoring
Weighting
Standard setting
Checklist scoring
Advantages
Helps examiner know what the station setters are
looking for
Helps the examiner be objective
Facilities the use of non-expert examiners
Disadvantages
Can just reward process/thoroughness
May not sufficiently reward the excellent candidate
Ignores the examiners expertise
Global scoring
Advantages
Utilises the expertise of the examiners
They are in a position to make a (global) judgement
about the performance
Disadvantages
Examiners have to be expert examiners i.e. trained
Examiners must be familiar with expected standards
for the level of the test
Weighting
In a checklist, some items may be weighted
more than others
More complicated scoring system
Makes no difference to very good & very bad
candidates
Can enhance discrimination at the cut score
Standard setting
No perfect method!
Should be criterion-referenced method
e.g. Angoff, Ebel, etc.
But
are these suitable for performance based tests?
Performance-based standard setting methods
Borderline group method
Contrasting group method
Regression based standard method
Kramer A, Muijtjens A, Jansen K, Düsman H, Tan L, van der Vleuten C
Comparison of a rational and an empirical standard setting procedure for an OSCE,
Medical Education, 2003 Vol 37 Issue 2, Page 132
Borderline method
Test score distribution
Checklist
1. Hs shjs sjnhss sjhs sjs sj
2. Ksks sksmsiqopql qlqmq q q qkl
3. Lalka kdm ddkk dlkl dlld
4. Keyw dd e r rrmt tmk
5. Jfjfk dd
6. Hskl;s skj sls ska ak akl ald
7. Hdhhddh shs ahhakk as
TOTAL
Borderline score distribution
Pass, Fail, Borderline P/B/F
Passing score
Contrasting groups method
Test score distribution
Checklist
1. Hs shjs sjnhss sjhs sjs sj
2. Ksks sksmsiqopql qlqmq q q qkl
3. Lalka kdm ddkk dlkl dlld
4. Keyw dd e r rrmt tmk
5. Jfjfk dd
6. Hskl;s skj sls ska ak akl ald
7. Hdhhddh shs ahhakk as
TOTAL
Fail
Pass
Pass, Fail, Borderline P/B/F
Passing score
Regression based standard
Checklist
X = passing score
1. Hs shjs sjnhss sjhs sjs sj
2. Ksks sksmsiqopql qlqmq q q qkl
3. Lalka kdm ddkk dlkl dlld
Checklist
Score
4. Keyw dd e r rrmt tmk
5. Jfjfk dd
6. Hskl;s skj sls ska ak akl ald
7. Hdhhddh shs ahhakk as
X
TOTAL
Overall rating 1 2 3 4 5
1
2
3
4
5
=
=
=
=
=
Clear fail
Borderline
Clear pass
v good pass
excellent pass
1
2
3
Clear
fail
Borderline
Clear
pass
4
5
v good pass excellent pass
Borderline/contrasting/regression based methods
“Panel” equals examiners
Reliable: cut-off score based on large sample of
judgments (no. of stations x no. of candidates)
Credible: based on expert judgment in direct
observation
Passing score not known in advance
(as with all
examinee centered methods)
Judgments not independent of checklist scoring