Using Evaluation to Optimize the Responsiveness of HIV Clinical Training and Capacity-Building Moderator: Janet Myers, PhD, MPH Presenters: Mi-Suk Kang-Dufour, PhD, MPH Kevin Khamarko, MA 2012 Ryan White.
Download ReportTranscript Using Evaluation to Optimize the Responsiveness of HIV Clinical Training and Capacity-Building Moderator: Janet Myers, PhD, MPH Presenters: Mi-Suk Kang-Dufour, PhD, MPH Kevin Khamarko, MA 2012 Ryan White.
Using Evaluation to Optimize the Responsiveness of HIV Clinical Training and Capacity-Building Moderator: Janet Myers, PhD, MPH Presenters: Mi-Suk Kang-Dufour, PhD, MPH Kevin Khamarko, MA 2012 Ryan White All Grantees Meeting Workshop Overview • • • • Brief description of the NEC Overview of AETC process & cross-site data sources How to use data to target training Examples of evaluations from specific initiatives: – HIV Testing Initiative – Minority AIDS Initiative – US/Mexico Border Initiative • Future directions for cross region evaluation Background on the AETC National Evaluation Center AETC NEC Focus • To provide leadership in the development, design, testing, and dissemination of effective evaluation models with emphasis on outcomes. • To determine the outcomes of AETC clinical education and training programs with respect to changes in provider behavior and clinical practice. • We evaluate all aspects of the training process, from pre-training conditions to training outcomes. Training Evaluation Framework Before Training Pre-Training Conditions: Tailoring Training: Individual Characteristics, Training Motivation, Pre-training Environment (Barriers and Facilitators Project Measures) Curriculum Adequacy Competency Assessment (Curriculum Adequacy Checklist, Professional Gap Assessment) UCSF NEC 2 2007 - 2012 UCSF NEC 3 2012 - 2016 During Training Training Process Evaluation: Number and Type of Activities (Event Record) and Trainees (Participant Information Form) Columbia NEC 1999 - 2004 After Training Training Outcome Evaluation: Training Impact Evaluation: Trainee Reaction (ACRE Immediate- Post) Learning and Behavior Change (ACRE Follow-Up) Patient Outcomes (Chart Review) UCSF NEC 1, 2 & 3 2004-2007 UCSF NEC 2 2007 - 2012 AETC NEC Services Serves as resource to AETC regions and centers: Evaluation training: web-based modular curriculum Evaluation planning & design IRB (human subjects) TA Web-based data collection system Survey creation Analysis Dissemination AETC Standardized Process and Cross-Regional Evaluation Data Sources AETC Process Data Forms • OMB approved forms required by HRSA – Event record (ER) collects information about the trainings – Participant Information Form (PIF) collects demographic and practice characteristics of training participants PIF and ER Datasets Total # of trainings Total # of trainees • • • FY 2008/09 FY 2009/10 FY 2010/11 18,070 17,190 17,357 139,496 141,751 132,692 There are PIFs without unique IDs, particularly from some regions In FY 10/11, there were 2.7% repeat PIFs Across 3 years, about 22% of trainees attended more than one training AETC Cross-Region Evaluation (ACRE) • Purpose & Significance: • To implement a select list of standardized evaluation questions across the AETC network to assess the effectiveness of training programs • Standardization Process: • Questions were vetted and piloted over the course of three years • Full implementation was required in fourth year (FY 10/11) Immediate Post: ACRE IP • Goal: To implement a standardized survey across the AETC network to assess trainees’: • knowledge change • intent to apply training content • Four standard questions developed in collaboration with the AETC evaluation community in late 2007 and finalized in July 2008. ACRE IP Questions 1. How would you rate your level of knowledge about this content before the training program? (Novice to Expert) 2. How would you rate your level of knowledge about this content after the training program? (Novice to Expert) 3. How would you rate the overall quality of the training program? (Poor to Excellent) 4. I can apply the information learned in my practice/service setting. (Disagree Strongly to Agree Strongly) 6 week Follow-up: ACRE FUP • Goal: to implement a standard quantitative follow up instrument measuring changes in knowledge and behavior at 6 weeks after a training • Qual pilot site findings & a draft quantitative survey were shared with AETCs in early 2010 • All regions began implementing the follow-up survey by April 1st, 2011 ACRE FUP Pilot Questions • The ACRE Follow-up pilot questionnaire focused on 4 main areas: 1. Information sharing (Yes/No) 2. Changes in abilities and/or skills (Scale, 1-5) 3. Changes in practice behavior (Scale, 1-5) 4. Implementation barriers encountered (Yes/No) ACRE Protocol AETC trainings are categorized into 5 levels: – Level 1: Didactic presentations – Level 2: Skills building – Level 3: Clinical Training – Level 4: Clinical consultations (group or individual) – Level 5: Technical Assistance ACRE Protocol • ACRE IP questions are collected immediately following the event for all Level 1, 2, and 3 trainings • ACRE FU questions are collected through web-based survey 6 weeks following all level 2 trainings that include clinical topics • Some exceptions made for joint trainings with other agencies or CME courses ACRE IP Response Rates** FY 08/09 FY 09/10 FY 10/11 Total PIFs collected* (levels 1-3) Total ACRE IP surveys completed % PIFs from (levels 1-3) with a matched ACRE IP * Not able to exclude exceptions for this analysis 42,467 88,765 89,550 14,430 20,725 40,537 28% 24% 48% ** FYs 08/09 and 09/10 included pilot regions only Number of IP records collected by quarter 14000 12000 10000 8000 6000 4000 2000 08/09 Fiscal Year 09/10 Fiscal year quarter 4 quarter 3 quarter 2 quarter 1 quarter 4 quarter 3 quarter 2 quarter 1 quarter 4 quarter 3 quarter 2 quarter 1 0 10/11 Fiscal year ACRE FU Pilot Implementation (1/1/2011 – 6/30/2011) Records ACRE Follow-up Event Records (ER) Participant Information Forms (PIF) N 1501 1363 866 Number of FUP records collected by quarter FY 2010/2011 600 500 400 300 200 100 0 quarter 1 quarter 2 quarter 3 09/10 Fiscal year quarter 4 General Findings from ACRE Evaluations Overall Means For Each IP Question 5.00 +0.98 4.00 +0.99 +0.76 3.00 08/09 09/10 10/11 2.00 1.00 0.00 Overall quality Ability to apply Knowledge the information before training Knowledge after training Means by Topic of Training (FY 10/11) Clinical Prevention Organization & Targeted management/ & behavior Psycho-social delivery population treatment change Overall quality Ability to apply the information Knowledge before training Knowledge after training Change in knowledge 4.18 4.24 4.20 4.23 4.32 4.20 4.26 4.23 4.27 4.32 3.22 3.27 3.30 3.26 3.26 3.94 4.03 4.04 4.01 4.09 0.72 0.76 0.74 0.75 0.83 Regression Findings: 09/10 Trainings • Compared to didactic trainings, trainings that incorporated skills building or technical assistance significantly increased perceived ability to apply the training content (both p<0.05). • Technical assistance was about twice as likely as skills building to improve perceived ability. Regression Findings: 09/10 Trainees • Direct HIV service providers were more likely to report higher gains in knowledge and ability to apply knowledge (p<0.05). • Compared to 08/09, in 09/10, direct service providers reported a larger gain in knowledge (p<0.05). Regression Findings: 09/10 Trainees cont. • Compared to white providers, non-white providers were more likely to report larger increases in knowledge (p<0.01). • In 09/10, in particular, African American providers demonstrated the greatest gains in knowledge among all groups. Using Evaluation Data to Target Training Example: Responding to National AIDS Strategy Goals NHAS has 4 overall goals: 1. Reducing new HIV infections 2. Increasing access to care and improving health outcomes for people living with HIV/AIDS 3. Reducing HIV-related disparities and health inequities 4. Achieving a more coordinated national response to the HIV epidemic Example: Responding to National AIDS Strategy Goals Achieving the NHAS goals depends on having a well trained cadre of health providers – a key function of the AETCs Training Providers is explicitly included in the NHAS plan (Goal 2 step 2): Take deliberate steps to increase the number and diversity of available providers of clinical care and related services for people living with HIV. Example: Responding to National AIDS Strategy Goals FY 08/09 FY 09/10 FY10/11 Novice HIV providers (<1 year) 11% 10% 8% New HIV providers (<2 years) 12% 12% 13% Providers in rural settings 19% 20% 20% Providers who are racial/ethnic minorities themselves 44% 43% 44% Providers who serve HIV infected minorities 54% 55% 55% Providers who serve mostly (>50%) HIV infected minorities 30% 31% 29% Evaluating Specific Initiatives 8000 Number of Training Events for specific initiatives, by funding year 7000 6000 5000 08/09 09/10 10/11 4000 3000 2000 1000 0 HIV testing (by topic) MAI UMBAST HIV Testing Myers JJ, Bradley-Springer L, Kang Dufour MS, Koester KA, Beane S, Warren N, Beal J, Frank LR. Supporting the integration of HIV testing into primary care settings. Am J Public Health. 2012 Jun;102(6):e25-32.. Routine HIV testing In 2008, the CDC provided supplemental funding to the AETC program to enhance delivery of intensive, clinic-based education, training, and technical assistance activities to support the integration of HIV testing into primary care settings 5 Routine HIV Testing Trainings: ACRE IP Outcomes FY 10/11 4 +0.97 3 2 1 0 Overall quality Ability to apply Knowledge Before Knowledge After HIV Testing Trainings: Findings • Compared with other AETC training, HIV testing training was longer and used a broader variety of strategies to educate more providers per training. • During education, providers were able to understand their primary care responsibility to address public health concerns through HIV testing. HIV Testing Case Strategies • Intensive long-term trainings focused on developing organization-level systems to help health professionals deliver testing. • Technical assistance concentrated on establishing or revising policies and procedures for testing and linkage to care for newly diagnosed patients. Minority AIDS Initiative Shade SB, Sackett N, Khamarko K, Koester KA, Bie J, Newberry J, Beal J, Culyba R, Jacobson K, Kinder A, Nusser J, Myers JJ. Quality of Comprehensive HIV Care in Underserved Communities: Does Clinical Training Lead to Improvement. Am J Med Qual. 2012 Aug 14. [Epub ahead of print] MAI Chart Review Project • Chart abstraction and feedback (34 clinics; N=530) was used to assess adherence to clinical practice guidelines. • We identified training needs and assessed change in clinical practice (14 clinics, N=271). MAI Chart Review Project • 49% (95% CI=44, 53) adherence to clinical practice guidelines at baseline • Gave feedback associated with chart review and provided clinical trainings • 11% increase (95% CI=7, 16) in adherence to clinical practice guidelines at follow up re e ... n * ca na tio g/ Ed u cc i tin g Monitoring ni n Va es g * Sc ST IT ha rt in g t * C ito r in Vi si in g Treatment M on tia l Te st xi s * In i b hy la RT l 20 La ro p A ra l 10 O IP O ve % Change in Adherence to Guidelines 25 Results Prevention * * 15 * * 5 0 * p<0.01 MAI Funded Trainings: findings from process and ACRE data • Among those providing direct services to HIV infected clients – Most provider services to minorities – Providers who are themselves minorities are more likely to have a practice that serves predominantly minority clients 60% Participants characteristics MAI Funded vs. Other trainings 50% 40% 30% 20% 10% 0% Minority providers Provide direct HIV care Provide direct HIV care to minority clients to mostly (>50%) minority clients MAI Funded Other funding Percent of providers reporting that they serve minority populations with HIV, by race/ethnicity of provider 70.0% 60.0% 50.0% 40.0% White Black Hispanic Other 30.0% 20.0% 10.0% 0.0% No Minority patients Less than Half of HIV patients are minorities More than half of HIV patients are minorities Based on PIF data from providers of direct services to HIV infected patients, aggregated across funding years 08/09 to 10/11 MAI Funded Trainings: findings from process and ACRE data • MAI trainings were well received across all years • Knowledge increased a similar amount across all funding years ACRE IP results for MAI funded trainings 5 +0.68 4 +0.67 +0.69 3 FY 08/09 FY 09/10 FY 10/11 2 1 0 Overall quality Ability to apply Knowledge Before Knowledge After MAI Funded Trainings: findings from process and ACRE data • Participants reported increased skills and abilities and implementing those skills at 6weeks post training • Skills increases and implementation were seen across training modes MAI Training Outcomes by Mode Chart/Case review Computer based Conference call/telephone Lecture/Workshop Preceptorship/mini-residency Role play/simulation Self-study Telemedicine 0 Implemented Skills and Abilities 1 2 3 4 Improved Skills and Abilities 5 U.S.-Mexico Border AETC Steering Team (UMBAST) UMBAST Background • 3 regional AETCs (Mountain-Plains, Pacific, and Texas/Oklahoma) provide targeted training and capacity building assistance to health care providers serving border communities – Border is defined as U.S. communities 61.5 miles north of the Mexican border UMBAST Background UMBAST Training Trends FY 2008 2009 FY 2009 2010 FY 2010 2011 FY 2011-2012 UMBAST Other UMBAST Other UMBAST Other UMBAST Other Total Events 104 2985 104 3389 68 3808 66 3509 Trainees with PIFs 782 31999 1002 35492 1241 31564 1078 25973 Mean # of PIFs 9.41 9.33 11.28 8.57 22.75 7.75 15.71 7.48 UMBAST Training Topics Health Care Clinical Organization Management and Delivery Prevention and Behavior Change Psycho - Targeted Immigrant/ Racial/Ethnic social Populations Border Minorities UMBAST FY 08/09 Other 78.85* 85.57* 41.35 20.19 27.88* 77.88 43.26* 89.78* 42.48* 41.41 24.49 16.34* 14.97 30.01* UMBAST FY 09/10 Other 70.19 79.81 46.15 25 24.03 79.81 50.96 84.63 41.13 38.09 27.91 15.23 16.4 33.17 UMBAST FY 10/11 Other 85.29 86.76* 70.59 30.88 32.35* 76.47 64.71* 87.73 33.64* 31.28 22.53 10.19* 11.84 26.65* UMBAST FY 11/12 Other 56.06 90.91 48.48 12.12 83.33 65.15 40.91 83.67 38.33 36.65 23.94 47.34 12.65 27.61 UMBAST Trends by Level FY 2008 2009 FY 2009 2010 FY 2010 2011 FY 2011-2012 UMBAST Other UMBAST Other UMBAST Other UMBAST Other Level 1- Didactic 32.69 24.79 24.03 17.17 42.64 14.29 34.85 11.91 Level 2- Skills Building 9.61 23.01 26.92 21.74 22.05 21.03 36.36 25.71 Level 3- Clinical Training 0.96* 11.29* 8.65 14.48 1.47 17.01 0 13.48 Level 4- Group Consult 7.69 8.07 11.53 8.76 48.53 8.95 30.3 9.52 Level 4- Clinical Consult 3.84* 23.55* 3.84 26.61 5.88 28.23 0 29.1 Level 5- TA 18.26* 66.35 17.41 20.59 14.62 31.82 13.57 65.38* UMBAST Training Trends FY 2008 2009 FY 2009 2010 FY 2010 2011 FY 2011-2012 UMBAST Other UMBAST Other UMBAST Other UMBAST Other % of Clinical Trainees 61.3 75.18 63.86 75.1 68.76 74.84 65.2 70.06 % >1 Year Experience 80.24 78.06 85.5 85.18 72.75 77.55 82.31 81.42 % Urban 53.64 69.22 56.99 69.07 63.89 72.32 62.81 67.9 Capacity Building Evaluation and Future Directions What is Capacity Building? Expert HIV Educator and Provider Longitudinal Training and Mentoring Model (LTMM)* Advanced HIV Clinical Care Intermediate HIV Clinical Care Basic HIV Clinical Care HIV Screening & Diagnosis Pre-HIV Screening, Diagnosis and Care *Adapted with permission from materials created by the Northwest AETC • CBA definition from the MAI project: “a program of multiple and distinct efforts aiming to improve the capacity of MAI-targeted clinics and providers to serve HIVinfected patients.” • As an example, an AETC might work with a clinic to conduct needs assessment and then plan a multi-month effort of engagement with a clinic including: – several didactic trainings for all clinic staff; – preceptorship opportunities for primary care providers; – one-on-one assistance in implementing a new patient flow system; and assistance developing a referral network for complex HIV care. SUPPORTED HIV CARE Stage 2 COLLABORATIVE HIV CARE Stage 1 EXTERNAL HIV CARE Foundation •Provide HIV testing (routine or targeted) but no medical management for patients with positive test results •Onsite HIV care integrated with primary care through extensive consultation from clinical mentor for initiation of ARV regimen and for management of more complex patients •Onsite HIV care integrated with primary care, supported by consultation with clinical mentor as needed for management of treatment failure and/or co-morbid conditions complicated by HIV disease and/or treatment COMPREHENSIVE HIV CARE MANAGEMENt Stage 3 •Onsite integrated HIV care with primary care, including management of complex comorbid conditions, PMTCT, and complex OI's with access to consultation from clinical mentors as needed NCHCMC’s Creating the Community Care Continuum Practice-based questions in the newest version of ACRE: From the list below, check the types of HIV clinical care services you provide to your patients/clients: I conduct HIV testing I provide primary care for HIV-infected patients I monitor HIV-specific lab tests I initiate antiretroviral therapy I conduct adherence counseling and monitor adherence I provide prophylaxis and treatment for opportunistic infections I manage treatment when drug resistance is present I initiate care to prevent and treat co-morbid conditions I provide clinical consultation to other clinicians regarding HIV care Practice-based questions in the newest version of ACRE: From the list below, check the types of situations in which you refer HIVinfected patients to other care settings: I refer patients for all HIV-related care and treatment after diagnosis I refer when I think the patient needs to start on antiretroviral therapy I refer when antiretroviral treatment fails I refer patients with co-infections/co-morbidities I refer patients with complex treatment issues I do not refer , but I do consult with HIV care specialists when I have questions I do not refer, I am an HIV care specialist I refer patients for all HIV-related care and treatment after diagnosis I refer when I think the patient needs to start on antiretroviral therapy Practice-based questions in the newest version of ACRE: How knowledgeable are you about where to refer an HIV-infected patient for care? I do not have any referral sources for HIV-specific patient care issues I have limited referral sources for HIV-specific patient care issues I have referral sources for some HIV-specific patient care issues I have referral sources for most HIV-specific patient care issues I have referral sources for all HIV-specific patient care issues Practice-based questions in the newest version of ACRE: In your practice, how often do you provide HIV testing to patients with unknown HIV status? To all existing patients To all new patients at intake To patients based on risk factors When a patient requests it I refer patients elsewhere for HIV testing I am not involved in HIV testing in any capacity To all existing patients To all new patients at intake To patients based on risk factors THANK YOU! Questions?