Transcript Slide 1

Postgraduate Medical Training – Evaluation and
Audit
Copenhagen Nov 2013
Professor Wendy Reid
Medical Director Health Education England
Past- Vice-President, Education, RCOG
© Royal College of Obstetricians and Gynaecologists
UK Specialty Training &
Education Programme
Professional authenticity
A model of clinical competence
Does
Shows how
Knows how
Knows
Behaviour =
Skills +
attitude
Cognition =
knowledge
Miller GE. The assessment of clinical skills/competence/performance.
Academic Medicine (Supplement) 1990; 65: S63-S67.
UK Specialist training programme
• Basic – years 1&2, part 1 MRCOG
• Intermediate – years 3,4&5, part 2 MRCOG
• Advanced – years 6&7, requires 2 ATSMs
minimum, career development and
‘independent’ competencies
• 19 core modules, subject based, includes
professional skills and leadership
Basic Training
• Exposure to the specialty
• Basic emergency obstetric and gynaecology
skills
• Understanding role of the doctor
• Team work – multi professional, develop
leadership skills
• Pass Part 1 MRCOG
Intermediate training
• Builds on basic skills
• Leadership – clinical, administrative
• Competences for normal practice i.e. day to day
obstetrics, emergency gynae and core gynae
skills
• Pass Part 2 MRCOG
• Workplace-based assessments
• More clinical responsibility, labour ward
leadership and acute gynaecology, develop
interests and choose advanced modules
Advanced training
• Core continues throughout programme!
• Advanced Training Skills Modules
(minimum x2)
• Designed to produce a workforce for the
service and give individuals scope to
develop clinical expertise in specific area
• New ATSMs in development, some
academic, some ‘professional’
Advanced training skills
modules (ATSMs)
• Fetal Medicine
• Benign Vaginal Surgery
• Advanced Labour Ward
Practice
• Advanced Lap surgery for the
excision of benign disease
• Benign Gynaecological
Surgery: Laparoscopy
• Labour Ward Lead
• Benign Gynaecological
Surgery: Hysteroscopy
• Maternal Medicine
• Colposcopy
• Advanced Antenatal Practice
• Vulval Disease
• Acute Gynaecology and Early
Pregnancy
• Abortion Care
• Gynaecological Oncology
• Sexual Health
• Subfertility and Reproductive
Endocrinology
• Menopause
• Urogynaecology
• Paediatric and Adolescent
Gynaecology
• Benign Abdominal Surgery
• Medical Education
• Domestic violence
Workplace Based Assessments
• All trainee grades in UK
• Varied names but similar principles
• Ongoing challenge of ‘formative vs
summative’
• Monitoring through Royal Colleges/Faculties
Professional authenticity
Testing formats
Does
Performance/
hands on
assessment
Shows how
Knows how
Knows
Written/
Computer
based
assessment
Miller GE. The assessment of clinical skills/competence/performance.
Academic Medicine (Supplement) 1990; 65: S63-S67.
Drivers for WBA
• New curricula – trainees need to prove
‘competence’
• GMC- the regulator ( and the public) want
explicit evidence of competence
• Professional examinations do not test ‘real
life’ skills and performance
• Learning from other systems
• One way of evaluating quality of training
UK experience of WBA
• Began with Foundation Programme (years 1&”
after graduation)
• Launched 2005, integrated assessment
process
• Regardless of post or geography
• Outcomes collated by Sheffield University
• Each training area (Deanery) informed of
‘outliers’
• Large cohort
• Centralised faculty training
Specialty training
•
•
•
•
•
From end of F2 to CCT
New curricula, launched August 2007
Assessment tools based on FP
Many ‘in development’ and specialty specific
Trainees in mixed programmes, mostly using log
books to capture evidence of progress
• Most curricula mandate ‘minimum numbers of
assessments’
• Summarised annually in ARCP (previously RITA)
Challenges of WBA in Specialty Training
• Does it really measure the doctors?
• Are we sure we are measuring the right
things?
• How often do they need to be done?
• Are they a good measure of continued
competence?
• How do we involve patients?
• How do we ensure trainers are trained and
have the time to do WBAs properly?
• To provide QA takes large numbers – poor
reliability
Other tools for QA of Training
• Longitudinal analysis of MRCOG results –
cohort comparison, demographic data
required
• Trainee doctor ‘user’ surveys
• Trainee feedback at end of training episodes
• Population wide survey of trainee doctors by
the GMC
Whole QA system
Formal requirements in UK for
Training QA
• Royal College annual report to GMC –
specialty specific
• Deanery (regional) annual report of Education
and Training – all specialties
• Trainees must complete the Annual GMC
survey
• All curriculum changes assessed by GMC
• All examination changes and examination data
submitted to the GMC
Whole QA system
• “The GMC expects medical schools and deaneries to
demonstrate compliance with the standards and
requirements that it sets. To do this, they will need to work in
close partnership with the medical Royal Colleges and
Faculties, NHS trusts and health boards and other LEPs. This
means that QM should be seen as a partnership between
those organisations because it is only through working
together that medical schools, deaneries, Royal Colleges and
Faculties, with LEPs, can deliver medical education and
training to the standards required.” (GMC Quality
Improvement Framework, para. 29)
Whole QA system
• “The GMC quality assures medical education and training
through the medical schools and deaneries but day-to-day
delivery is at LEP level. This delivery involves medical staff,
medical education managers, undergraduate and
postgraduate medical centre staff, other health professions
and employers. Clinical placements, student assistantships,
individual foundation programme and specialty including GP
training are delivered through careful supervision and
assessment by specialists in the relevant discipline advised
and overseen by regional and local staff from the UKFPO, the
Academy of Medical Royal Colleges and the relevant medical
Royal College or Faculty.” (GMC Quality Improvement
Framework, para. 46)
Role of medical royal colleges
• Set curriculum and workplace-based assessments for trainee
doctors according to GMC standards
• Set criteria for progression between stages of training
• Engage with a range of stakeholders to assure quality of
training, particularly 16 UK deaneries
• Provide fora for making policy, sharing best practice and
developing training requirements as clinical practice develops
• Provide specialist faculty development
• Assure the quality of individual trainees (recommendation for
CCT/CESR, MRCOG)
Role of medical royal colleges
• Colleges can also raise concerns about
patient or trainee safety directly with the
GMC or CQC
• Colleges work together on national
medical education policy through
Academy
Governance
• College committees agree national policy on various
aspects of specialty education (e.g. exams,
curriculum, ARCP)
• Network of College Tutors coordinate training in
individual Trusts
• Specialist educational management and leadership
roles created in Colleges (e.g. committee chair)
• Heads of Deanery Specialty Schools jointly appointed
with Colleges
• Colleges report to GMC via Annual Report
QA processes
• ARCP – colleges send specialist assessor to assure
deanery process for progressing trainees
• Quality visits – colleges provide specialist assessor on
request to join deanery visit team
• CCT/CESR(CP) and CESR – recommendation of
individual doctor to GMC for inclusion on specialist
register
• Examination – standard-setting
• Curriculum approval – changes to curricula approved
by GMC
Data on quality
• GMC Trainee Survey
• ARCP outcome data – summary of achievements
annually for every trainee
• Examination data
• Colleges’ own surveys (e.g. Training Evaluation Form)
• Reports from external assessors on local/regional QA
processes (ARCP and quality visits)
• Increasingly linked with quality of care and patient
safety reviews
GMC Trainees’ survey – O&G
Perspectives
• Three specific elements o How O&G trainees compare with other
specialties.
o How the results from this year for specific
questions compare with those in previous years
(looking at the areas previously considered).
o Specialty Specific Questions
Total number of trainees responding 49000 (95%)
Trainee Evaluation Forms
• Not mandatory
• Might work effectively if based on MSF ‘360’
feedback
• Should be real-time tool for local training
quality management
• Best discriminator is ‘would you recommend
this job to a friend?’
Programme Groups
National
Programme Group
ACCS
Mean
This Report
Min Q1 Median
Q3 Max Lower CI
Upper CI
N
Mean
Lower CI
Upper CI
N
79.66
24 72
80 92
100
78.75
80.57 1114
79.66
78.75
80.57 1114
Acute Internal Medicine
81.72
20 76
84 92
100
81.32
82.12 4766
77.6
75.97
79.24
302
Allergy
81.72
20 76
84 92
100
81.32
82.12 4766
82.8
74.7
90.9
10
Anaesthetics
82.68
20 76
84 92
100
82.16
83.2 2409
82.58
82.05
Anaesthetics F1
75.46
20 68
76 84
100
75.12
75.79 7077
89.92
88.5
91.33
198
Anaesthetics F2
78.67
20 72
80 88
100
78.33
79.01 7138
87.79
86.32
89.27
232
Audio vestibular medicine
81.72
20 76
84 92
100
81.32
82.12 4766
81.6
74.11
89.09
15
Cardiology
81.72
20 76
84 92
100
81.32
82.12 4766
81
79.75
82.26
550
Cardio-thoracic surgery
83.67
20 76
84 96
100
83.19
84.14 3514
82.78
79.28
86.28
95
Chemical pathology
84.93
20 80
84 96
100
83.93
85.92
672
80.45
76.81
84.09
62
Child and adolescent psychiatry
86.46
20 80
88 96
100
85.75
87.18 1232
87
85.32
88.67
211
Clinical genetics
81.72
20 76
84 92
100
81.32
82.12 4766
86.78
83.84
89.73
46
Clinical neurophysiology
81.72
20 76
84 92
100
81.32
82.12 4766
85.22
80.66
89.77
23
Clinical oncology
84.53
20 76
84 96
100
83.85
85.2 1332
82.3
80.79
83.81
285
Clinical pharmacology and therapeutics
81.72
20 76
84 92
100
81.32
82.12 4766
78.86
71.5
86.22
21
Clinical radiology
84.53
20 76
84 96
100
83.85
85.2 1332
85.13
84.38
85.89 1047
CMT
74.55
20 64
76 84
100
74
75.11 2730
74.55
74
75.11 2730
Community Sexual and Reproductive Health
78.59
20 68
80 88
100
77.93
79.25 1915
78.46
71.5
Core Anaesthetics
85.28
20 80
88 96
100
84.48
86.08 1052
85.28
84.48
86.08 1052
CPT
81.77
24 76
84 92
100
81.06
82.47 1529
81.77
81.06
82.47 1529
CST
74.52
20 64
76 88
100
73.67
75.38 1463
74.52
73.67
75.38 1463
Dermatology
81.72
20 76
84 92
100
81.32
82.12 4766
84.29
82.38
86.21
191
Emergency medicine
80.15
28 72
80 92
100
78.98
81.32
550
80.15
78.98
81.32
550
Emergency Medicine F1
75.46
20 68
76 84
100
75.12
75.79 7077
87.81
86.05
89.57
169
Emergency Medicine F2
78.67
20 72
80 88
100
78.33
79.01 7138
82.45
81.75
83.14 1199
83.11 2358
85.43
13
O&G Programme Group
Comparison
Supervision (1)
How would you rate the
quality of (clinical) supervision
in this post?
Very Poor
Poor
Fair
0.26%
0.91%
1.31%
1.38%
3.00%
3.81%
5.09%
4.31%
O&G Trainees 2012
15.83%
18.02%
20.08%
22.61%
O&G Trainees 2010
50.05%
48.95%
49.78%
49.22%
Good
Excellent
O&G Trainees 2011
O&G Trainees 2009
30.86%
28.31%
23.74%
22.49%
Did you have a designated educational supervisor (the person responsible for your appraisal) in this post?
Yes
2012 – 99.3% (2011 – 99.5%, 2010 – 99.5%, 2009 – 99.8%)
In this post did you have a training/learning agreement with your educational supervisor, setting out your respective responsibilities?
Yes
2012 – 86.4% (2011 – 91.9%, 2010 – 92.6%, 2009 – 91.1%)
In this post did you use a learning portfolio?
Yes
2012 – 92.4% (2011 – 94.7%, 2010 – 89.9%, 2009 – 91.2%)
In this post were you told who to talk to in confidence if you had concerns, personal or educational?
Yes
2012 – 71.2% (2011 – 77.7%, 2010 – 72.2%, 2009 – 68.8%)
Supervision (2)
Did you have a formal meeting with your supervisor to
talk about your progress in this post?
Did you have a formal assessment of your performance
in the workplace in this post?
O&G Trainees 2012
O&G Trainees 2011
O&G Trainees 2012
O&G Trainees 2011
O&G Trainees 2010
O&G Trainees 2009
O&G Trainees 2010
O&G Trainees 2009
No, but I would like to
No, but this will happen
Yes, but it wasn't useful
Yes, and it was useful
1.63%
1.71%
1.81%
3.79%
No, but I would like to
6.47%
4.17%
6.65%
13.83%
10.15%
11.94%
11.82%
10.04%
81.13%
82.17%
79.71%
72.34%
4.84%
5.17%
6.94%
8.27%
No, but this will happen
8.94%
6.50%
10.30%
20.22%
Yes, but it wasn't useful
7.26%
8.94%
8.17%
7.60%
Yes, and it was useful
76.97%
79.38%
74.59%
63.91%
Access to Training (1)
How would you rate the practical
experience you were receiving in
this post?
Very Poor
Poor
O&G Trainees 2012
1.05%
1.02%
1.64%
1.79%
O&G Trainees 2011
O&G Trainees 2010
5.10%
4.09%
7.11%
7.60%
Fair
O&G Trainees 2009
22.50%
25.13%
26.53%
30.86%
42.95%
47.53%
44.69%
42.34%
Good
Excellent
Never
Rarely
Monthly
O&G Trainees 2012
2.05%
3.92%
4.98%
3.89%
O&G Trainees 2011
O&G Trainees 2010
21.50%
28.03%
28.17%
27.45%
O&G Trainees 2009
16.82%
16.43%
17.72%
16.15%
38.15%
36.71%
37.44%
Weekly
Daily
28.39%
22.23%
20.02%
17.40%
12.72%
13.47%
12.42%
15.07%
46.90%
In this post, how often have you
worked beyond your rostered
hours?
Access to Training (2)
O&G Trainees 2009
2.63%
O&G Trainees 2010
O&G Trainees 2011
O&G Trainees 2012
16.81%
20.63%
12.62%
19.04%
18.27%
10.89%
2.90%
22.85%
17.57%
9.10%
1.53%
26.60%
16.40%
7.89%
1.68%
47.31%
48.91%
How confident are you that this post will help you
acquire the competencies you needed at that
particular stage of your training?
Very confident
Fairly confident
Neutral
48.95%
Not very confident
Not at all confident
47.42%
How good or poor was access to each of the following
in your post? (2012 question only)
Simulation facilities
Equipped rooms for group teaching
Space for private study
Internet access
E-learning resources
Very
Good
Good
Online journals
Library
9.41%
31.91%
11.72%
53.68%
6.36%
31.91%
22.77%
14.14%
12.78%
24.24%
51.79%
53.21%
46.32%
46.90%
Working Beyond Competence (1)
In this post how often did you feel forced to cope with clinical problems beyond your competence or experience?
26.67%
O&G Trainees 2009
55.74%
10.11%
6.16%
1.32%
27.90%
O&G Trainees 2010
10.28%
6.24%
0.98%
O&G Trainees 2012
O&G Trainees 2010
O&G Trainees 2011
O&G Trainees 2012
4.19%
2.81%
0.72%
4.38%
2.13%
0.82%
3.81%
1.82%
0.68%
3.36%
1.58%
0.42%
32.66%
57.70%
8.93%
6.03%
0.85%
44.11%
48.37%
4.89%
2.21%
0.42%
In this post how often have you been expected to obtain
consent for procedures where you feel you do not understand
the proposed interventions and its risks?
59.63%
O&G Trainees 2009
33.64%
32.58%
59.03%
Never
Rarely
61.11%
O&G Trainees 2010
Monthly
Weekly
O&G Trainees 2011
Daily
28.50%
Weekly
Daily
In this post how often, if ever, were you supervised by
someone who you felt wasn't competent to do so?
O&G Trainees 2009
Rarely
Monthly
26.49%
O&G Trainees 2011
Never
54.60%
33.01%
34.83%
11.75%
17.93%
2.48%
2.06%
0.61%
0.11%
1.85%
0.64%
0.06%
66.14%
O&G Trainees 2012
1.10%
0.21%
0.11%
22.20%
20.28%
16.56%
75.01%
Never
Rarely
77.17%
Monthly
Weekly
79.71%
Daily
Working Beyond Competence (2)
O&G Trainees 2009
O&G Trainees 2010
No, there was no-one I could contact
O&G Trainees 2011
In this post did you always know who
was providing your clinical supervision
when you were working? (2009 – 2011
inclusive)
0.00%
0.55%
0.23%
No, but there was usually someone I
could contact
6.96%
8.34%
6.03%
Yes, but they were not easy to access
6.72%
8.34%
6.60%
Yes and they were accessible
86.32%
82.78%
87.14%
3.89%
0.37%
0.16%
6.89%
In this post did you always know who
your available senior support was during
on call (2012)
Yes - accessible
Yes - not easy access
No - but someone to
contact
No - no one to contact
88.70%
N/A
Undermining – 2012 (1)
0.63%
1.74%
How often, if at all, have you been the victim of bullying and
harassment in this post?
2.00%
3.73%
18.82%
64.51%
8.57%
Every Day
0.47%
At least once per week
2.58%
How often, if at all, have you witnessed someone else being the
victim of bullying and harassment in this post?
At least once per fortnight
2.42%
At least once per month
7.26%
24.76%
Less often than once per month
51.95%
10.57%
Prefer not to answer
0.53%
2.00%
In this post, how often if at all, have you experienced behaviour
from a consultant/GP that undermined your professional
confidence and/or self esteem?
2.79%
5.47%
26.81%
54.15%
8.25%
Never
Undermining 2012 (2)
Overall
•
96.0% of trainees said they had never
been bullied and/or harassed in their post,
or if they had, it happened less than once
a month. 1.1% said it happened every day
or at least once per week (n=48,512).
•
1.6% said they had witnessed someone
else being the victim of bullying and/or
harassment in their post every day or at
least once per week (n=48,464).
•
92.4% said they had never experienced
behaviour from a consultant or GP that
undermined their professional confidence
and/or self-esteem or, if they had, it
happened less than once a month. 1.7%
said it happened every day or at least
once per week (n=48,785).
O&G
•
The equivalent figures for O&G are
83.33% and 2.37%.
•
The equivalent figure for O&G is
3.05%.
•
The equivalent figures for O&G are
80.96% and 2.53%.
(1902 respondents for O&G)
O&G versus other specialties
Next Steps
• Specialty specific questions to be further analysed
• Breakdown by training level may also be available –
need to discuss with GMC.
• Will be involved with preparation for 2014 survey.
• Must publish more quickly
• An updated trainers survey has been discussed –
believe something may be in place within 12 months.
• TEF – a potential method of triangulation?
• RCOG has appointed Workplace Advisory Officer to
combat undermining
Who does what in governance
of training?
• Education Board -RCOG
• Heads of Schools – joint between college and
local regional Postgraduate Dean
• Local Education and Training Boards
(Deaneries)
• Individual hospitals (Local Education
Providers, LEPs) – DMEs or Clinical Tutors
• Individual doctors through trusts or
organisations
Future developments
• Outcome of review of QA system by GMC in Autumn 2013
(note: GMC became regulator in 2010 for PG medical
education)
• Growing recognition of need to clarify role of colleges in QA
• Increased focus on sharing data between deaneries, colleges
and GMC
• Increased emphasis on the role of educational and clinical
supervisors/trainers with consequent impact on service
delivery
• Impact of national policy changes, e.g. Shape of Training, new
English healthcare structure
• Role of HEE, relationship with Colleges, GMC, devolved
nations