User login
“Thank You for Not Letting Me Crash and Burn”: The Imperative of Quality Physician Onboarding to Foster Job Satisfaction, Strengthen Workplace Culture, and Advance the Quadruple Aim
From The Ohio State University College of Medicine Department of Family and Community Medicine, Columbus, OH (Candy Magaña, Jná Báez, Christine Junk, Drs. Ahmad, Conroy, and Olayiwola); The Ohio State University College of Medicine Center for Primary Care Innovation and Transformation (Candy Magaña, Jná Báez, and Dr. Olayiwola); and The Ohio State University Wexner Medical Center (Christine Harsh, Erica Esposito).
Much has been discussed about the growing crisis of professional dissatisfaction among physicians, with increasing efforts being made to incorporate physician wellness into health system strategies that move from the Triple to the Quadruple Aim.1 For many years, our health care system has been focused on improving the health of populations, optimizing the patient experience, and reducing the cost of care (Triple Aim). The inclusion of the fourth aim, improving the experience of the teams that deliver care, has become paramount in achieving the other aims.
An area often overlooked in this focus on wellness, however, is the importance of the earliest days of employment to shape and predict long-term career contentment. This is a missed opportunity, as data suggest that organizations with standardized onboarding programs boast a 62% increased productivity rate and a 50% greater retention rate among new hires.2,3 Moreover, a study by the International Institute for Management Development found that businesses lose an estimated $37 billion annually because employees do not fully understand their jobs.4 The report ties losses to “actions taken by employees who have misunderstood or misinterpreted company policies, business processes, job function, or a combination of the three.” Additionally, onboarding programs that focus strictly on technical or functional orientation tasks miss important opportunities for culture integration during the onboarding process.5 It is therefore imperative to look to effective models of employee onboarding to develop systems that position physicians and practices for success.
Challenges With Traditional Physician Onboarding
In recent years, the Department of Family and Community Medicine at The Ohio State University College of Medicine has experienced rapid organizational change. Like many primary care systems nationwide responding to disruption in health care and changing demands on the clinical workforce, the department has hired new leadership, revised strategic priorities, and witnessed an influx of faculty and staff. It has also planned an expansion of ambulatory services that will more than double the clinical workforce over the next 3 years. While an exciting time, there has been a growing need to align strategy, culture, and human capital during these changes.
As we entered this phase of transformation, we recognized that our highly individualized, ad hoc orientation system presented shortcomings. During the act of revamping our physician recruitment process, stakeholder workgroup members specifically noted that improvement efforts were needed regarding new physician orientation, as no consistent structures were previously in place. New physician orientation had been a major gap for years, resulting in dissatisfaction in the first few months of physician practice, early physician turnover, and staff frustration. For physicians, we continued to learn about their frustration and unanswered questions regarding expectations, norms, structures, and processes.
Many new hires were left with a kind of “trial by fire” entry into their roles. On the first day of clinic, a new physician would most likely need to simultaneously see patients, learn the nuances of the electronic health record (EHR), figure out where the break room was located, and quickly learn population health issues for the patients they were serving. Opportunities to meet key clinic site leadership would be at random, and new physicians might not have the opportunity to meet leadership or staff until months into their tenure; this did not allow for a sense of belonging or understanding of the many resources available to them. We learned that the quality of these ad hoc orientations also varied based on the experience and priorities of each practice’s clinic and administrative leaders, who themselves felt ill-equipped to provide a consistent, robust, and confidence-building experience. In addition, practice site management was rarely given advance time to prepare for the arrival of new physicians, which resulted in physicians perceiving practices to be unwelcoming and disorganized. Their first days were often spent with patients in clinic with no structured orientation and without understanding workflows or having systems practice knowledge.
Institutionally, the interview process satisfied some transfer of knowledge, but we were unclear of what was being consistently shared and understood in the multiple ambulatory locations where our physicians enter practice. More importantly, we knew we were missing a critical opportunity to use orientation to imbue other values of diversity and inclusion, health equity, and operational excellence into the workforce. Based on anecdotal insights from employees and our own review of successful onboarding approaches from other industries, we also knew a more structured welcoming process would predict greater long-term career satisfaction for physicians and create a foundation for providing optimal care for patients when clinical encounters began.
Reengineering Physician Onboarding
In 2019, our department developed a multipronged approach to physician onboarding, which is already paying dividends in easing acculturation and fostering team cohesion. The department tapped its Center for Primary Care Innovation and Transformation (PCIT) to direct this effort, based on its expertise in practice transformation, clinical transformation and adaptations, and workflow efficiency through process and quality improvement. The PCIT team provides support to the department and the entire health system focused on technology and innovation, health equity, and health care efficiency.6 They applied many of the tools used in the Clinical Transformation in Technology approach to lead this initiative.7
The PCIT team began identifying key stakeholders (department, clinical and ambulatory leadership, clinicians and clinical staff, community partners, human resources, and resident physicians), and then engaging those individuals in dialogue surrounding orientation needs. During scheduled in-person and virtual work sessions, stakeholders were asked to provide input on pain points for new physicians and clinic leadership and were then empowered to create an onboarding program. Applying health care quality improvement techniques, we leveraged workflow mapping, current and future state planning, and goal setting, led by the skilled process improvement and clinical transformation specialists. We coordinated a multidisciplinary process improvement team that included clinic administrators, medical directors, human resources, administrative staff, ambulatory and resident leadership, clinical leadership, and recruitment liaisons. This diverse group of leadership and staff was brought together to address these critical identified gaps and weaknesses in new physician onboarding.
Through a series of learning sessions, the workgroup provided input that was used to form an itemized physician onboarding schedule, which was then leveraged to develop Plan-Do-Study-Act (PDSA) cycles, collecting feedback in real time. Some issues that seem small can cause major distress for new physicians. For example, in our inaugural orientation implementation, a physician provided feedback that they wanted to obtain information on setting up their work email on their personal devices and was having considerable trouble figuring out how to do so. This particular topic was not initially included in the first iteration of the Department’s orientation program. We rapidly sought out different ways to embed that into the onboarding experience. The first PDSA involved integrating the university information technology team (IT) into the process but was not successful because it required extra work for the new physician and reliance on the IT schedule. The next attempt was to have IT train a department staff member, but again, this still required that the physician find time to connect with that staff member. Finally, we decided to obtain a useful tip sheet that clearly outlined the process and could be included in orientation materials. This gave the new physicians control over how and when they would work on this issue. Based on these learnings, this was incorporated as a standing agenda item and resource for incoming physicians.
Essential Elements of Effective Onboarding
The new physician onboarding program consists of 5 key elements: (1) 2-week acclimation period; (2) peer learning and connection; (3) training before beginning patient care; (4) standardization, transparency, and accountability in all processes; (5) ongoing feedback for continued program improvement with individual support (Figure).
The program begins with a 2-week period of intentional investment in individual success, during which time no patients are scheduled. In week 1, we work with new hires to set expectations for performance, understand departmental norms, and introduce culture. Physicians meet formally and informally with department and institutional leadership, as well as attend team meetings and trainings that include a range of administrative and compliance requirements, such as quality standards and expectations, compliance, billing and coding specific to family medicine, EHR management, and institutionally mandated orientations. We are also adding implicit bias and antiracism training during this period, which are essential to creating a culture of unity and belonging.
During week 2, we focus on clinic-level orientation, assigning new hires an orientation buddy and a department sponsor, such as a physician lead or medical director. Physicians spend time with leadership at their clinic as they nurture relationships important for mentorship, sponsorship, and peer support. They also meet care team members, including front desk associates, medical assistants, behavioral health clinicians, nutritionists, social workers, pharmacists, and other key colleagues and care team members. This introduces the physician to the clinical environment and physical space as well as acclimates the physician to workflows and feedback loops for regular interaction.
When physicians ultimately begin patient care, they begin with an expected productivity rate of 50%, followed by an expected productivity rate of 75%, and then an expected productivity rate of 100%. This steady increase occurs over 3 to 4 weeks depending on the physician’s comfort level. They are also provided monthly reports on work relative value unit performance so that they can track and adapt practice patterns as necessary.More details on the program can be found in Appendix 1.
Takeaways From the Implementation of the New Program
Give time for new physicians to focus on acclimating to the role and environment.
The initial 2-week period of transition—without direct patient care—ensures that physicians feel comfortable in their new ecosystem. This also supports personal transitions, as many new hires are managing relocation and acclimating themselves and their families to new settings. Even residents from our training program who returned as attending physicians found this flexibility and slow reentry essential. This also gives the clinic time to orient to an additional provider, nurture them into the team culture, and develop relationships with the care team.
Cultivate spaces for shared learning, problem-solving, and peer connection.
Orientation is delivered primarily through group learning sessions with cohorts of new physicians, thus developing spaces for networking, fostering psychological safety, encouraging personal and professional rapport, emphasizing interactive learning, and reinforcing scheduling blocks at the departmental level. New hires also participate in peer shadowing to develop clinical competencies and are assigned a workplace buddy to foster a sense of belonging and create opportunities for additional knowledge sharing and cross-training.
Strengthen physician knowledge base, confidence, and comfort in the workplace before beginning direct patient care.
Without fluency in the workflows, culture, and operations of a practice, the urgency to have physicians begin clinical care can result in frustration for the physician, patients, and clinical and administrative staff. Therefore, we complete essential training prior to seeing any patients. This includes clinical workflows, referral processes, use of alternate modalities of care (eg, telehealth, eConsults), billing protocols, population health training, patient resources, office resources, and other essential daily processes and tools. This creates efficiency in administrative management, increased productivity, and better understanding of resources available for patients’ medical, social, and behavioral needs when patient care begins.
Embrace standardization, transparency, and accountability in as many processes as possible.
Standardized knowledge-sharing and checklists are mandated at every step of the orientation process, requiring sign off from the physician lead, practice manager, and new physicians upon completion. This offers all parties the opportunity to play a role in the delivery of and accountability for skills transfer and empowers new hires to press pause if they feel unsure about any domain in the training. It is also essential in guaranteeing that all physicians—regardless of which ambulatory location they practice in—receive consistent information and expectations. A sample checklist can be found in Appendix 2.
Commit to collecting and acting on feedback for continued program improvement and individual support.
As physicians complete the program, it is necessary to create structures to measure and enhance its impact, as well as evaluate how physicians are faring following the program. Each physician completes surveys at the end of the orientation program, attends a 90-day post-program check-in with the department chair, and receives follow-up trainings on advanced topics as they become more deeply embedded in the organization.
Lessons Learned
Feedback from surveys and 90-day check-ins with leadership and physicians reflect a high degree of clarity on job roles and duties, a sense of team camaraderie, easier system navigation, and a strong sense of support. We do recognize that sustaining change takes time and our study is limited by data demonstrating the impact of these efforts. We look forward to sharing more robust data from surveys and qualitative interviews with physicians, clinical leadership, and staff in the future. Our team will conduct interviews at 90-day and 180-day checkpoints with new physicians who have gone through this program, followed by a check-in after 1 year. Additionally, new physicians as well as key stakeholders, such as physician leads, practice managers, and members of the recruitment team, have started to participate in short surveys. These are designed to better understand their experiences, what worked well, what can be improved, and the overall satisfaction of the physician and other members of the extended care team.
What follows are some comments made by the initial group of physicians that went through this program and participated in follow-up interviews:
“I really feel like part of a bigger team.”
“I knew exactly what do to when I walked into the exam room on clinic Day 1.”
“It was great to make deep connections during the early process of joining.”
“Having a buddy to direct questions and ideas to is amazing and empowering.”
“Even though the orientation was long, I felt that I learned so much that I would not have otherwise.”
“Thank you for not letting me crash and burn!”
“Great culture! I love understanding our values of health equity, diversity, and inclusion.”
In the months since our endeavor began, we have learned just how essential it is to fully and effectively integrate new hires into the organization for their own satisfaction and success—and ours. Indeed, we cannot expect to achieve the Quadruple Aim without investing in the kind of transparent and intentional orientation process that defines expectations, aligns cultural values, mitigates costly and stressful operational misunderstandings, and communicates to physicians that, not only do they belong, but their sense of belonging is our priority. While we have yet to understand the impact of this program on the fourth aim of the Quadruple Aim, we are hopeful that the benefits will be far-reaching.
It is our ultimate hope that programs like this: (1) give physicians the confidence needed to create impactful patient-centered experiences; (2) enable physicians to become more cost-effective and efficient in care delivery; (3) allow physicians to understand the populations they are serving and access tools available to mitigate health disparities and other barriers; and (4) improve the collective experience of every member of the care team, practice leadership, and clinician-patient partnership.
Corresponding author: J. Nwando Olayiwola, MD, MPH, FAAFP, The Ohio State University College of Medicine, Department of Family and Community Medicine, 2231 N High St, Ste 250, Columbus, OH 43210; [email protected].
Financial disclosures: None.
Keywords: physician onboarding; Quadruple Aim; leadership; clinician satisfaction; care team satisfaction.
1. Bodenheimer T, Sinsky C. From triple to quadruple aim: care of the patient requires care of the provider. Ann Fam Med. 2014;12(6): 573-576.
2. Maurer R. Onboarding key to retaining, engaging talent. Society for Human Resource Management. April 16, 2015. Accessed January 8, 2021. https://www.shrm.org/resourcesandtools/hr-topics/talent-acquisition/pages/onboarding-key-retaining-engaging-talent.aspx
3. Boston AG. New hire onboarding standardization and automation powers productivity gains. GlobeNewswire. March 8, 2011. Accessed January 8, 2021. http://www.globenewswire.com/news-release/2011/03/08/994239/0/en/New-Hire-Onboarding-Standardization-and-Automation-Powers-Productivity-Gains.html
4. $37 billion – US and UK business count the cost of employee misunderstanding. HR.com – Maximizing Human Potential. June 18, 2008. Accessed March 10, 2021. https://www.hr.com/en/communities/staffing_and_recruitment/37-billion---us-and-uk-businesses-count-the-cost-o_fhnduq4d.html
5. Employers risk driving new hires away with poor onboarding. Society for Human Resource Management. February 23, 2018. Accessed March 10, 2021. https://www.shrm.org/resourcesandtools/hr-topics/talent-acquisition/pages/employers-new-hires-poor-onboarding.aspx
6. Center for Primary Care Innovation and Transformation. The Ohio State University College of Medicine. Accessed January 8, 2021. https://wexnermedical.osu.edu/departments/family-medicine/pcit
7. Olayiwola, J.N. and Magaña, C. Clinical transformation in technology: a fresh change management approach for primary care. Harvard Health Policy Review. February 2, 2019. Accessed March 10, 2021. http://www.hhpronline.org/articles/2019/2/2/clinical-transformation-in-technology-a-fresh-change-management-approach-for-primary-care
From The Ohio State University College of Medicine Department of Family and Community Medicine, Columbus, OH (Candy Magaña, Jná Báez, Christine Junk, Drs. Ahmad, Conroy, and Olayiwola); The Ohio State University College of Medicine Center for Primary Care Innovation and Transformation (Candy Magaña, Jná Báez, and Dr. Olayiwola); and The Ohio State University Wexner Medical Center (Christine Harsh, Erica Esposito).
Much has been discussed about the growing crisis of professional dissatisfaction among physicians, with increasing efforts being made to incorporate physician wellness into health system strategies that move from the Triple to the Quadruple Aim.1 For many years, our health care system has been focused on improving the health of populations, optimizing the patient experience, and reducing the cost of care (Triple Aim). The inclusion of the fourth aim, improving the experience of the teams that deliver care, has become paramount in achieving the other aims.
An area often overlooked in this focus on wellness, however, is the importance of the earliest days of employment to shape and predict long-term career contentment. This is a missed opportunity, as data suggest that organizations with standardized onboarding programs boast a 62% increased productivity rate and a 50% greater retention rate among new hires.2,3 Moreover, a study by the International Institute for Management Development found that businesses lose an estimated $37 billion annually because employees do not fully understand their jobs.4 The report ties losses to “actions taken by employees who have misunderstood or misinterpreted company policies, business processes, job function, or a combination of the three.” Additionally, onboarding programs that focus strictly on technical or functional orientation tasks miss important opportunities for culture integration during the onboarding process.5 It is therefore imperative to look to effective models of employee onboarding to develop systems that position physicians and practices for success.
Challenges With Traditional Physician Onboarding
In recent years, the Department of Family and Community Medicine at The Ohio State University College of Medicine has experienced rapid organizational change. Like many primary care systems nationwide responding to disruption in health care and changing demands on the clinical workforce, the department has hired new leadership, revised strategic priorities, and witnessed an influx of faculty and staff. It has also planned an expansion of ambulatory services that will more than double the clinical workforce over the next 3 years. While an exciting time, there has been a growing need to align strategy, culture, and human capital during these changes.
As we entered this phase of transformation, we recognized that our highly individualized, ad hoc orientation system presented shortcomings. During the act of revamping our physician recruitment process, stakeholder workgroup members specifically noted that improvement efforts were needed regarding new physician orientation, as no consistent structures were previously in place. New physician orientation had been a major gap for years, resulting in dissatisfaction in the first few months of physician practice, early physician turnover, and staff frustration. For physicians, we continued to learn about their frustration and unanswered questions regarding expectations, norms, structures, and processes.
Many new hires were left with a kind of “trial by fire” entry into their roles. On the first day of clinic, a new physician would most likely need to simultaneously see patients, learn the nuances of the electronic health record (EHR), figure out where the break room was located, and quickly learn population health issues for the patients they were serving. Opportunities to meet key clinic site leadership would be at random, and new physicians might not have the opportunity to meet leadership or staff until months into their tenure; this did not allow for a sense of belonging or understanding of the many resources available to them. We learned that the quality of these ad hoc orientations also varied based on the experience and priorities of each practice’s clinic and administrative leaders, who themselves felt ill-equipped to provide a consistent, robust, and confidence-building experience. In addition, practice site management was rarely given advance time to prepare for the arrival of new physicians, which resulted in physicians perceiving practices to be unwelcoming and disorganized. Their first days were often spent with patients in clinic with no structured orientation and without understanding workflows or having systems practice knowledge.
Institutionally, the interview process satisfied some transfer of knowledge, but we were unclear of what was being consistently shared and understood in the multiple ambulatory locations where our physicians enter practice. More importantly, we knew we were missing a critical opportunity to use orientation to imbue other values of diversity and inclusion, health equity, and operational excellence into the workforce. Based on anecdotal insights from employees and our own review of successful onboarding approaches from other industries, we also knew a more structured welcoming process would predict greater long-term career satisfaction for physicians and create a foundation for providing optimal care for patients when clinical encounters began.
Reengineering Physician Onboarding
In 2019, our department developed a multipronged approach to physician onboarding, which is already paying dividends in easing acculturation and fostering team cohesion. The department tapped its Center for Primary Care Innovation and Transformation (PCIT) to direct this effort, based on its expertise in practice transformation, clinical transformation and adaptations, and workflow efficiency through process and quality improvement. The PCIT team provides support to the department and the entire health system focused on technology and innovation, health equity, and health care efficiency.6 They applied many of the tools used in the Clinical Transformation in Technology approach to lead this initiative.7
The PCIT team began identifying key stakeholders (department, clinical and ambulatory leadership, clinicians and clinical staff, community partners, human resources, and resident physicians), and then engaging those individuals in dialogue surrounding orientation needs. During scheduled in-person and virtual work sessions, stakeholders were asked to provide input on pain points for new physicians and clinic leadership and were then empowered to create an onboarding program. Applying health care quality improvement techniques, we leveraged workflow mapping, current and future state planning, and goal setting, led by the skilled process improvement and clinical transformation specialists. We coordinated a multidisciplinary process improvement team that included clinic administrators, medical directors, human resources, administrative staff, ambulatory and resident leadership, clinical leadership, and recruitment liaisons. This diverse group of leadership and staff was brought together to address these critical identified gaps and weaknesses in new physician onboarding.
Through a series of learning sessions, the workgroup provided input that was used to form an itemized physician onboarding schedule, which was then leveraged to develop Plan-Do-Study-Act (PDSA) cycles, collecting feedback in real time. Some issues that seem small can cause major distress for new physicians. For example, in our inaugural orientation implementation, a physician provided feedback that they wanted to obtain information on setting up their work email on their personal devices and was having considerable trouble figuring out how to do so. This particular topic was not initially included in the first iteration of the Department’s orientation program. We rapidly sought out different ways to embed that into the onboarding experience. The first PDSA involved integrating the university information technology team (IT) into the process but was not successful because it required extra work for the new physician and reliance on the IT schedule. The next attempt was to have IT train a department staff member, but again, this still required that the physician find time to connect with that staff member. Finally, we decided to obtain a useful tip sheet that clearly outlined the process and could be included in orientation materials. This gave the new physicians control over how and when they would work on this issue. Based on these learnings, this was incorporated as a standing agenda item and resource for incoming physicians.
Essential Elements of Effective Onboarding
The new physician onboarding program consists of 5 key elements: (1) 2-week acclimation period; (2) peer learning and connection; (3) training before beginning patient care; (4) standardization, transparency, and accountability in all processes; (5) ongoing feedback for continued program improvement with individual support (Figure).
The program begins with a 2-week period of intentional investment in individual success, during which time no patients are scheduled. In week 1, we work with new hires to set expectations for performance, understand departmental norms, and introduce culture. Physicians meet formally and informally with department and institutional leadership, as well as attend team meetings and trainings that include a range of administrative and compliance requirements, such as quality standards and expectations, compliance, billing and coding specific to family medicine, EHR management, and institutionally mandated orientations. We are also adding implicit bias and antiracism training during this period, which are essential to creating a culture of unity and belonging.
During week 2, we focus on clinic-level orientation, assigning new hires an orientation buddy and a department sponsor, such as a physician lead or medical director. Physicians spend time with leadership at their clinic as they nurture relationships important for mentorship, sponsorship, and peer support. They also meet care team members, including front desk associates, medical assistants, behavioral health clinicians, nutritionists, social workers, pharmacists, and other key colleagues and care team members. This introduces the physician to the clinical environment and physical space as well as acclimates the physician to workflows and feedback loops for regular interaction.
When physicians ultimately begin patient care, they begin with an expected productivity rate of 50%, followed by an expected productivity rate of 75%, and then an expected productivity rate of 100%. This steady increase occurs over 3 to 4 weeks depending on the physician’s comfort level. They are also provided monthly reports on work relative value unit performance so that they can track and adapt practice patterns as necessary.More details on the program can be found in Appendix 1.
Takeaways From the Implementation of the New Program
Give time for new physicians to focus on acclimating to the role and environment.
The initial 2-week period of transition—without direct patient care—ensures that physicians feel comfortable in their new ecosystem. This also supports personal transitions, as many new hires are managing relocation and acclimating themselves and their families to new settings. Even residents from our training program who returned as attending physicians found this flexibility and slow reentry essential. This also gives the clinic time to orient to an additional provider, nurture them into the team culture, and develop relationships with the care team.
Cultivate spaces for shared learning, problem-solving, and peer connection.
Orientation is delivered primarily through group learning sessions with cohorts of new physicians, thus developing spaces for networking, fostering psychological safety, encouraging personal and professional rapport, emphasizing interactive learning, and reinforcing scheduling blocks at the departmental level. New hires also participate in peer shadowing to develop clinical competencies and are assigned a workplace buddy to foster a sense of belonging and create opportunities for additional knowledge sharing and cross-training.
Strengthen physician knowledge base, confidence, and comfort in the workplace before beginning direct patient care.
Without fluency in the workflows, culture, and operations of a practice, the urgency to have physicians begin clinical care can result in frustration for the physician, patients, and clinical and administrative staff. Therefore, we complete essential training prior to seeing any patients. This includes clinical workflows, referral processes, use of alternate modalities of care (eg, telehealth, eConsults), billing protocols, population health training, patient resources, office resources, and other essential daily processes and tools. This creates efficiency in administrative management, increased productivity, and better understanding of resources available for patients’ medical, social, and behavioral needs when patient care begins.
Embrace standardization, transparency, and accountability in as many processes as possible.
Standardized knowledge-sharing and checklists are mandated at every step of the orientation process, requiring sign off from the physician lead, practice manager, and new physicians upon completion. This offers all parties the opportunity to play a role in the delivery of and accountability for skills transfer and empowers new hires to press pause if they feel unsure about any domain in the training. It is also essential in guaranteeing that all physicians—regardless of which ambulatory location they practice in—receive consistent information and expectations. A sample checklist can be found in Appendix 2.
Commit to collecting and acting on feedback for continued program improvement and individual support.
As physicians complete the program, it is necessary to create structures to measure and enhance its impact, as well as evaluate how physicians are faring following the program. Each physician completes surveys at the end of the orientation program, attends a 90-day post-program check-in with the department chair, and receives follow-up trainings on advanced topics as they become more deeply embedded in the organization.
Lessons Learned
Feedback from surveys and 90-day check-ins with leadership and physicians reflect a high degree of clarity on job roles and duties, a sense of team camaraderie, easier system navigation, and a strong sense of support. We do recognize that sustaining change takes time and our study is limited by data demonstrating the impact of these efforts. We look forward to sharing more robust data from surveys and qualitative interviews with physicians, clinical leadership, and staff in the future. Our team will conduct interviews at 90-day and 180-day checkpoints with new physicians who have gone through this program, followed by a check-in after 1 year. Additionally, new physicians as well as key stakeholders, such as physician leads, practice managers, and members of the recruitment team, have started to participate in short surveys. These are designed to better understand their experiences, what worked well, what can be improved, and the overall satisfaction of the physician and other members of the extended care team.
What follows are some comments made by the initial group of physicians that went through this program and participated in follow-up interviews:
“I really feel like part of a bigger team.”
“I knew exactly what do to when I walked into the exam room on clinic Day 1.”
“It was great to make deep connections during the early process of joining.”
“Having a buddy to direct questions and ideas to is amazing and empowering.”
“Even though the orientation was long, I felt that I learned so much that I would not have otherwise.”
“Thank you for not letting me crash and burn!”
“Great culture! I love understanding our values of health equity, diversity, and inclusion.”
In the months since our endeavor began, we have learned just how essential it is to fully and effectively integrate new hires into the organization for their own satisfaction and success—and ours. Indeed, we cannot expect to achieve the Quadruple Aim without investing in the kind of transparent and intentional orientation process that defines expectations, aligns cultural values, mitigates costly and stressful operational misunderstandings, and communicates to physicians that, not only do they belong, but their sense of belonging is our priority. While we have yet to understand the impact of this program on the fourth aim of the Quadruple Aim, we are hopeful that the benefits will be far-reaching.
It is our ultimate hope that programs like this: (1) give physicians the confidence needed to create impactful patient-centered experiences; (2) enable physicians to become more cost-effective and efficient in care delivery; (3) allow physicians to understand the populations they are serving and access tools available to mitigate health disparities and other barriers; and (4) improve the collective experience of every member of the care team, practice leadership, and clinician-patient partnership.
Corresponding author: J. Nwando Olayiwola, MD, MPH, FAAFP, The Ohio State University College of Medicine, Department of Family and Community Medicine, 2231 N High St, Ste 250, Columbus, OH 43210; [email protected].
Financial disclosures: None.
Keywords: physician onboarding; Quadruple Aim; leadership; clinician satisfaction; care team satisfaction.
From The Ohio State University College of Medicine Department of Family and Community Medicine, Columbus, OH (Candy Magaña, Jná Báez, Christine Junk, Drs. Ahmad, Conroy, and Olayiwola); The Ohio State University College of Medicine Center for Primary Care Innovation and Transformation (Candy Magaña, Jná Báez, and Dr. Olayiwola); and The Ohio State University Wexner Medical Center (Christine Harsh, Erica Esposito).
Much has been discussed about the growing crisis of professional dissatisfaction among physicians, with increasing efforts being made to incorporate physician wellness into health system strategies that move from the Triple to the Quadruple Aim.1 For many years, our health care system has been focused on improving the health of populations, optimizing the patient experience, and reducing the cost of care (Triple Aim). The inclusion of the fourth aim, improving the experience of the teams that deliver care, has become paramount in achieving the other aims.
An area often overlooked in this focus on wellness, however, is the importance of the earliest days of employment to shape and predict long-term career contentment. This is a missed opportunity, as data suggest that organizations with standardized onboarding programs boast a 62% increased productivity rate and a 50% greater retention rate among new hires.2,3 Moreover, a study by the International Institute for Management Development found that businesses lose an estimated $37 billion annually because employees do not fully understand their jobs.4 The report ties losses to “actions taken by employees who have misunderstood or misinterpreted company policies, business processes, job function, or a combination of the three.” Additionally, onboarding programs that focus strictly on technical or functional orientation tasks miss important opportunities for culture integration during the onboarding process.5 It is therefore imperative to look to effective models of employee onboarding to develop systems that position physicians and practices for success.
Challenges With Traditional Physician Onboarding
In recent years, the Department of Family and Community Medicine at The Ohio State University College of Medicine has experienced rapid organizational change. Like many primary care systems nationwide responding to disruption in health care and changing demands on the clinical workforce, the department has hired new leadership, revised strategic priorities, and witnessed an influx of faculty and staff. It has also planned an expansion of ambulatory services that will more than double the clinical workforce over the next 3 years. While an exciting time, there has been a growing need to align strategy, culture, and human capital during these changes.
As we entered this phase of transformation, we recognized that our highly individualized, ad hoc orientation system presented shortcomings. During the act of revamping our physician recruitment process, stakeholder workgroup members specifically noted that improvement efforts were needed regarding new physician orientation, as no consistent structures were previously in place. New physician orientation had been a major gap for years, resulting in dissatisfaction in the first few months of physician practice, early physician turnover, and staff frustration. For physicians, we continued to learn about their frustration and unanswered questions regarding expectations, norms, structures, and processes.
Many new hires were left with a kind of “trial by fire” entry into their roles. On the first day of clinic, a new physician would most likely need to simultaneously see patients, learn the nuances of the electronic health record (EHR), figure out where the break room was located, and quickly learn population health issues for the patients they were serving. Opportunities to meet key clinic site leadership would be at random, and new physicians might not have the opportunity to meet leadership or staff until months into their tenure; this did not allow for a sense of belonging or understanding of the many resources available to them. We learned that the quality of these ad hoc orientations also varied based on the experience and priorities of each practice’s clinic and administrative leaders, who themselves felt ill-equipped to provide a consistent, robust, and confidence-building experience. In addition, practice site management was rarely given advance time to prepare for the arrival of new physicians, which resulted in physicians perceiving practices to be unwelcoming and disorganized. Their first days were often spent with patients in clinic with no structured orientation and without understanding workflows or having systems practice knowledge.
Institutionally, the interview process satisfied some transfer of knowledge, but we were unclear of what was being consistently shared and understood in the multiple ambulatory locations where our physicians enter practice. More importantly, we knew we were missing a critical opportunity to use orientation to imbue other values of diversity and inclusion, health equity, and operational excellence into the workforce. Based on anecdotal insights from employees and our own review of successful onboarding approaches from other industries, we also knew a more structured welcoming process would predict greater long-term career satisfaction for physicians and create a foundation for providing optimal care for patients when clinical encounters began.
Reengineering Physician Onboarding
In 2019, our department developed a multipronged approach to physician onboarding, which is already paying dividends in easing acculturation and fostering team cohesion. The department tapped its Center for Primary Care Innovation and Transformation (PCIT) to direct this effort, based on its expertise in practice transformation, clinical transformation and adaptations, and workflow efficiency through process and quality improvement. The PCIT team provides support to the department and the entire health system focused on technology and innovation, health equity, and health care efficiency.6 They applied many of the tools used in the Clinical Transformation in Technology approach to lead this initiative.7
The PCIT team began identifying key stakeholders (department, clinical and ambulatory leadership, clinicians and clinical staff, community partners, human resources, and resident physicians), and then engaging those individuals in dialogue surrounding orientation needs. During scheduled in-person and virtual work sessions, stakeholders were asked to provide input on pain points for new physicians and clinic leadership and were then empowered to create an onboarding program. Applying health care quality improvement techniques, we leveraged workflow mapping, current and future state planning, and goal setting, led by the skilled process improvement and clinical transformation specialists. We coordinated a multidisciplinary process improvement team that included clinic administrators, medical directors, human resources, administrative staff, ambulatory and resident leadership, clinical leadership, and recruitment liaisons. This diverse group of leadership and staff was brought together to address these critical identified gaps and weaknesses in new physician onboarding.
Through a series of learning sessions, the workgroup provided input that was used to form an itemized physician onboarding schedule, which was then leveraged to develop Plan-Do-Study-Act (PDSA) cycles, collecting feedback in real time. Some issues that seem small can cause major distress for new physicians. For example, in our inaugural orientation implementation, a physician provided feedback that they wanted to obtain information on setting up their work email on their personal devices and was having considerable trouble figuring out how to do so. This particular topic was not initially included in the first iteration of the Department’s orientation program. We rapidly sought out different ways to embed that into the onboarding experience. The first PDSA involved integrating the university information technology team (IT) into the process but was not successful because it required extra work for the new physician and reliance on the IT schedule. The next attempt was to have IT train a department staff member, but again, this still required that the physician find time to connect with that staff member. Finally, we decided to obtain a useful tip sheet that clearly outlined the process and could be included in orientation materials. This gave the new physicians control over how and when they would work on this issue. Based on these learnings, this was incorporated as a standing agenda item and resource for incoming physicians.
Essential Elements of Effective Onboarding
The new physician onboarding program consists of 5 key elements: (1) 2-week acclimation period; (2) peer learning and connection; (3) training before beginning patient care; (4) standardization, transparency, and accountability in all processes; (5) ongoing feedback for continued program improvement with individual support (Figure).
The program begins with a 2-week period of intentional investment in individual success, during which time no patients are scheduled. In week 1, we work with new hires to set expectations for performance, understand departmental norms, and introduce culture. Physicians meet formally and informally with department and institutional leadership, as well as attend team meetings and trainings that include a range of administrative and compliance requirements, such as quality standards and expectations, compliance, billing and coding specific to family medicine, EHR management, and institutionally mandated orientations. We are also adding implicit bias and antiracism training during this period, which are essential to creating a culture of unity and belonging.
During week 2, we focus on clinic-level orientation, assigning new hires an orientation buddy and a department sponsor, such as a physician lead or medical director. Physicians spend time with leadership at their clinic as they nurture relationships important for mentorship, sponsorship, and peer support. They also meet care team members, including front desk associates, medical assistants, behavioral health clinicians, nutritionists, social workers, pharmacists, and other key colleagues and care team members. This introduces the physician to the clinical environment and physical space as well as acclimates the physician to workflows and feedback loops for regular interaction.
When physicians ultimately begin patient care, they begin with an expected productivity rate of 50%, followed by an expected productivity rate of 75%, and then an expected productivity rate of 100%. This steady increase occurs over 3 to 4 weeks depending on the physician’s comfort level. They are also provided monthly reports on work relative value unit performance so that they can track and adapt practice patterns as necessary.More details on the program can be found in Appendix 1.
Takeaways From the Implementation of the New Program
Give time for new physicians to focus on acclimating to the role and environment.
The initial 2-week period of transition—without direct patient care—ensures that physicians feel comfortable in their new ecosystem. This also supports personal transitions, as many new hires are managing relocation and acclimating themselves and their families to new settings. Even residents from our training program who returned as attending physicians found this flexibility and slow reentry essential. This also gives the clinic time to orient to an additional provider, nurture them into the team culture, and develop relationships with the care team.
Cultivate spaces for shared learning, problem-solving, and peer connection.
Orientation is delivered primarily through group learning sessions with cohorts of new physicians, thus developing spaces for networking, fostering psychological safety, encouraging personal and professional rapport, emphasizing interactive learning, and reinforcing scheduling blocks at the departmental level. New hires also participate in peer shadowing to develop clinical competencies and are assigned a workplace buddy to foster a sense of belonging and create opportunities for additional knowledge sharing and cross-training.
Strengthen physician knowledge base, confidence, and comfort in the workplace before beginning direct patient care.
Without fluency in the workflows, culture, and operations of a practice, the urgency to have physicians begin clinical care can result in frustration for the physician, patients, and clinical and administrative staff. Therefore, we complete essential training prior to seeing any patients. This includes clinical workflows, referral processes, use of alternate modalities of care (eg, telehealth, eConsults), billing protocols, population health training, patient resources, office resources, and other essential daily processes and tools. This creates efficiency in administrative management, increased productivity, and better understanding of resources available for patients’ medical, social, and behavioral needs when patient care begins.
Embrace standardization, transparency, and accountability in as many processes as possible.
Standardized knowledge-sharing and checklists are mandated at every step of the orientation process, requiring sign off from the physician lead, practice manager, and new physicians upon completion. This offers all parties the opportunity to play a role in the delivery of and accountability for skills transfer and empowers new hires to press pause if they feel unsure about any domain in the training. It is also essential in guaranteeing that all physicians—regardless of which ambulatory location they practice in—receive consistent information and expectations. A sample checklist can be found in Appendix 2.
Commit to collecting and acting on feedback for continued program improvement and individual support.
As physicians complete the program, it is necessary to create structures to measure and enhance its impact, as well as evaluate how physicians are faring following the program. Each physician completes surveys at the end of the orientation program, attends a 90-day post-program check-in with the department chair, and receives follow-up trainings on advanced topics as they become more deeply embedded in the organization.
Lessons Learned
Feedback from surveys and 90-day check-ins with leadership and physicians reflect a high degree of clarity on job roles and duties, a sense of team camaraderie, easier system navigation, and a strong sense of support. We do recognize that sustaining change takes time and our study is limited by data demonstrating the impact of these efforts. We look forward to sharing more robust data from surveys and qualitative interviews with physicians, clinical leadership, and staff in the future. Our team will conduct interviews at 90-day and 180-day checkpoints with new physicians who have gone through this program, followed by a check-in after 1 year. Additionally, new physicians as well as key stakeholders, such as physician leads, practice managers, and members of the recruitment team, have started to participate in short surveys. These are designed to better understand their experiences, what worked well, what can be improved, and the overall satisfaction of the physician and other members of the extended care team.
What follows are some comments made by the initial group of physicians that went through this program and participated in follow-up interviews:
“I really feel like part of a bigger team.”
“I knew exactly what do to when I walked into the exam room on clinic Day 1.”
“It was great to make deep connections during the early process of joining.”
“Having a buddy to direct questions and ideas to is amazing and empowering.”
“Even though the orientation was long, I felt that I learned so much that I would not have otherwise.”
“Thank you for not letting me crash and burn!”
“Great culture! I love understanding our values of health equity, diversity, and inclusion.”
In the months since our endeavor began, we have learned just how essential it is to fully and effectively integrate new hires into the organization for their own satisfaction and success—and ours. Indeed, we cannot expect to achieve the Quadruple Aim without investing in the kind of transparent and intentional orientation process that defines expectations, aligns cultural values, mitigates costly and stressful operational misunderstandings, and communicates to physicians that, not only do they belong, but their sense of belonging is our priority. While we have yet to understand the impact of this program on the fourth aim of the Quadruple Aim, we are hopeful that the benefits will be far-reaching.
It is our ultimate hope that programs like this: (1) give physicians the confidence needed to create impactful patient-centered experiences; (2) enable physicians to become more cost-effective and efficient in care delivery; (3) allow physicians to understand the populations they are serving and access tools available to mitigate health disparities and other barriers; and (4) improve the collective experience of every member of the care team, practice leadership, and clinician-patient partnership.
Corresponding author: J. Nwando Olayiwola, MD, MPH, FAAFP, The Ohio State University College of Medicine, Department of Family and Community Medicine, 2231 N High St, Ste 250, Columbus, OH 43210; [email protected].
Financial disclosures: None.
Keywords: physician onboarding; Quadruple Aim; leadership; clinician satisfaction; care team satisfaction.
1. Bodenheimer T, Sinsky C. From triple to quadruple aim: care of the patient requires care of the provider. Ann Fam Med. 2014;12(6): 573-576.
2. Maurer R. Onboarding key to retaining, engaging talent. Society for Human Resource Management. April 16, 2015. Accessed January 8, 2021. https://www.shrm.org/resourcesandtools/hr-topics/talent-acquisition/pages/onboarding-key-retaining-engaging-talent.aspx
3. Boston AG. New hire onboarding standardization and automation powers productivity gains. GlobeNewswire. March 8, 2011. Accessed January 8, 2021. http://www.globenewswire.com/news-release/2011/03/08/994239/0/en/New-Hire-Onboarding-Standardization-and-Automation-Powers-Productivity-Gains.html
4. $37 billion – US and UK business count the cost of employee misunderstanding. HR.com – Maximizing Human Potential. June 18, 2008. Accessed March 10, 2021. https://www.hr.com/en/communities/staffing_and_recruitment/37-billion---us-and-uk-businesses-count-the-cost-o_fhnduq4d.html
5. Employers risk driving new hires away with poor onboarding. Society for Human Resource Management. February 23, 2018. Accessed March 10, 2021. https://www.shrm.org/resourcesandtools/hr-topics/talent-acquisition/pages/employers-new-hires-poor-onboarding.aspx
6. Center for Primary Care Innovation and Transformation. The Ohio State University College of Medicine. Accessed January 8, 2021. https://wexnermedical.osu.edu/departments/family-medicine/pcit
7. Olayiwola, J.N. and Magaña, C. Clinical transformation in technology: a fresh change management approach for primary care. Harvard Health Policy Review. February 2, 2019. Accessed March 10, 2021. http://www.hhpronline.org/articles/2019/2/2/clinical-transformation-in-technology-a-fresh-change-management-approach-for-primary-care
1. Bodenheimer T, Sinsky C. From triple to quadruple aim: care of the patient requires care of the provider. Ann Fam Med. 2014;12(6): 573-576.
2. Maurer R. Onboarding key to retaining, engaging talent. Society for Human Resource Management. April 16, 2015. Accessed January 8, 2021. https://www.shrm.org/resourcesandtools/hr-topics/talent-acquisition/pages/onboarding-key-retaining-engaging-talent.aspx
3. Boston AG. New hire onboarding standardization and automation powers productivity gains. GlobeNewswire. March 8, 2011. Accessed January 8, 2021. http://www.globenewswire.com/news-release/2011/03/08/994239/0/en/New-Hire-Onboarding-Standardization-and-Automation-Powers-Productivity-Gains.html
4. $37 billion – US and UK business count the cost of employee misunderstanding. HR.com – Maximizing Human Potential. June 18, 2008. Accessed March 10, 2021. https://www.hr.com/en/communities/staffing_and_recruitment/37-billion---us-and-uk-businesses-count-the-cost-o_fhnduq4d.html
5. Employers risk driving new hires away with poor onboarding. Society for Human Resource Management. February 23, 2018. Accessed March 10, 2021. https://www.shrm.org/resourcesandtools/hr-topics/talent-acquisition/pages/employers-new-hires-poor-onboarding.aspx
6. Center for Primary Care Innovation and Transformation. The Ohio State University College of Medicine. Accessed January 8, 2021. https://wexnermedical.osu.edu/departments/family-medicine/pcit
7. Olayiwola, J.N. and Magaña, C. Clinical transformation in technology: a fresh change management approach for primary care. Harvard Health Policy Review. February 2, 2019. Accessed March 10, 2021. http://www.hhpronline.org/articles/2019/2/2/clinical-transformation-in-technology-a-fresh-change-management-approach-for-primary-care
An Analysis of the Involvement and Attitudes of Resident Physicians in Reporting Errors in Patient Care
From Adelante Healthcare, Mesa, AZ (Dr. Chin), University Hospitals of Cleveland, Cleveland, OH (Drs. Delozier, Bascug, Levine, Bejanishvili, and Wynbrandt and Janet C. Peachey, Rachel M. Cerminara, and Sharon M. Darkovich), and Houston Methodist Hospitals, Houston, TX (Dr. Bhakta).
Abstract
Background: Resident physicians play an active role in the reporting of errors that occur in patient care. Previous studies indicate that residents significantly underreport errors in patient care.
Methods: Fifty-four of 80 eligible residents enrolled at University Hospitals–Regional Hospitals (UH-RH) during the 2018-2019 academic year completed a survey assessing their knowledge and experience in completing Patient Advocacy and Shared Stories (PASS) reports, which serve as incident reports in the UH health system in reporting errors in patient care. A series of interventions aimed at educating residents about the PASS report system were then conducted. The 54 residents who completed the first survey received it again 4 months later.
Results: Residents demonstrated greater understanding of when filing PASS reports was appropriate after the intervention, as significantly more residents reported having been involved in a situation where they should have filed a PASS report but did not (P = 0.036).
Conclusion: In this study, residents often did not report errors in patient care because they simply did not know the process for doing so. In addition, many residents often felt that the reporting of patient errors could be used as a form of retaliation.
Keywords: resident physicians; quality improvement; high-value care; medical errors; patient safety.
Resident physicians play a critical role in patient care. Residents undergo extensive supervised training in order to one day be able to practice medicine in an unsupervised setting, with the goal of providing the highest quality of care possible. One study reported that primary care provided by residents in a training program is of similar or higher quality than that provided by attending physicians.1
Besides providing high-quality care, it is important that residents play an active role in the reporting of errors that occur regarding patient care as well as in identifying events that may compromise patient safety and quality.2 In fact, increased reporting of patient errors has been shown to decrease liability-related costs for hospitals.3 Unfortunately, physicians, and residents in particular, have historically been poor reporters of errors in patient care.4 This is especially true when comparing physicians to other health professionals, such as nurses, in error reporting.5
Several studies have examined the involvement of residents in reporting errors in patient care. One recent study showed that a graduate medical education financial incentive program significantly increased the number of patient safety events reported by residents and fellows.6 This study, along with several others, supports the concept of using incentives to help improve the reporting of errors in patient care for physicians in training.7-10 Another study used Quality Improvement Knowledge Assessment Tool (QIKAT) scores to assess quality improvement (QI) knowledge. The study demonstrated that self-assessment scores of QI skills using QIKAT scores improved following a targeted intervention.11 Because further information on the involvement and attitudes of residents in reporting errors in patient care is needed, University Hospitals of Cleveland (UH) designed and implemented a QI study during the 2018-2019 academic year. This prospective study used anonymous surveys to objectively examine the involvement and attitudes of residents in reporting errors in patient care.
Methods
The UH health system uses Patient Advocacy and Shared Stories (PASS) reports as incident reports to not only disclose errors in patient care but also to identify any events that may compromise patient safety and quality. Based on preliminary review, nurses, ancillary staff, and administrators file the majority of PASS reports.
The study group consisted of residents at University Hospitals–Regional Hospitals (UH-RH), which is comprised of 2 hospitals: University Hospitals–Richmond Medical Center (UH-RMC) and University Hospitals –Bedford Medical Center (UH-BMC). UH-RMC and UH-BMC are 2 medium-sized university-affiliated community hospitals located in the Cleveland metropolitan area in Northeast Ohio. Both serve as clinical training sites for Case Western Reserve University School of Medicine and Lake Erie College of Osteopathic Medicine, the latter of which helped fund this study. The study was submitted to the Institutional Review Board (IRB) of University Hospitals of Cleveland and granted “not human subjects research” status as a QI study.
Surveys
UH-RH offers residency programs in dermatology, emergency medicine, family medicine, internal medicine, orthopedic surgery, and physical medicine and rehabilitation, along with a 1-year transitional/preliminary year. A total of 80 residents enrolled at UH-RH during the 2018-2019 academic year. All 80 residents at UH-RH received an email in December 2018 asking them to complete an anonymous survey regarding the PASS report system. The survey was administered using the REDCap software system and consisted of 15 multiple-choice questions. As an incentive for completing the survey, residents were offered a $10 Amazon gift card. The gift cards were funded through a research grant from Lake Erie College of Osteopathic Medicine. Residents were given 1 week to complete the survey. At the end of the week, 54 of 80 residents completed the first survey.
Following the first survey, efforts were undertaken by the study authors, in conjunction with the quality improvement department at UH-RH, to educate residents about the PASS report system. These interventions included giving a lecture on the PASS report system during resident didactic sessions, sending an email to all residents about the PASS report system, and providing residents an opportunity to complete an optional online training course regarding the PASS report system. As an incentive for completing the online training course, residents were offered a $10 Amazon gift card. As before, the gift cards were funded through a research grant from Lake Erie College of Osteopathic Medicine.
A second survey was administered in April 2019, 4 months after the first survey. To determine whether the intervention made an impact on the involvement and attitudes of residents in the reporting errors in patient care, only residents who completed the first survey were sent the second survey. The second survey consisted of the same questions as the first survey and was also administered using the REDCap software system. As an incentive for completing the survey, residents were offered another $10 Amazon gift card, again were funded through a research grant from Lake Erie College of Osteopathic Medicine. Residents were given 1 week to complete the survey.
Analysis
Chi-square analyses were utilized to examine differences between preintervention and postintervention responses across categories. All analyses were conducted using R statistical software, version 3.6.1 (R Foundation for Statistical Computing).
Results
A total of 54 of 80 eligible residents responded to the first survey (Table). Twenty-nine of 54 eligible residents responded to the second survey. Postintervention, significantly more residents indicated being involved in a situation where they should have filed a PASS report but did not (58.6% vs 53.7%; P = 0.036). Improvement was seen in PASS knowledge postintervention, where fewer residents reported not knowing how to file a PASS report (31.5% vs 55.2%; P = 0.059). No other improvements were significant, nor were there significant differences in responses between any other categories pre- and postintervention.
Discussion
Errors in patient care are a common occurrence in the hospital setting. Reporting errors when they happen is important for hospitals to gain data and better care for patients, but studies show that patient errors are usually underreported. This is concerning, as data on errors and other aspects of patient care are needed to inform quality improvement programs.
This study measured residents’ attitudes and knowledge regarding the filing of a PASS report. It also aimed to increase both the frequency of and knowledge about filing a PASS report with interventions. The results from each survey indicated a statistically significant increase in knowledge of when to file a PASS report. In the first survey, 53.7% of residents responded they they were involved in an instance where they should have filed a PASS report but did not. In the second survey, 58.5% of residents reported being involved in an instance where they should have filed a PASS report but did not. This difference was statistically significant (P = 0.036), sugesting that the intervention was successful at increasing residents’ knowledge regarding PASS reports and the appropriate times to file a PASS report.
The survey results also showed a trend toward increasing aggregate knowledge level of how to file PASS reports on the first survey and second surveys (from 31.5% vs 55.2%. This demonstrates an increase in knowledge of how to file a PASS report among residents at our hospital after the intervention. It should be noted that the intervention that was performed in this study was simple, easy to perform, and can be completed at any hospital system that uses a similar system for reporting patient errors.
Another important trend indicating the effectiveness of the intervention was a 15% increase in knowledge of what the PASS report acronym stands for, along with a 13.1% aggregate increase in the number of residents who filed a PASS report. This indicated that residents may have wanted to file a PASS report previously but simply did not know how to until the intervention. In addition, there was also a decrease in the aggregate percentages of residents who had never filed a PASS report and an increase in how many PASS reports were filed.
While PASS reports are a great way for hospitals to gain data and insight into problems at their sites, there was also a negative view of PASS reports. For example, a large percentage of residents indicated that filing a PASS report would not make any difference and that PASS reports are often used as a form of retaliation, either against themselves as the submitter or the person(s) mentioned in the PASS report. More specifically, more than 50% of residents felt that PASS reports were sometimes or often used as a form of retaliation against others. While many residents correctly identified in the survey that PASS reports are not equivalent to a “write-up,” it is concerning that they still feel there is a strong potential for retaliation when filing a PASS report. This finding is unfortunate but matches the results of a multicenter study that found that 44.6% of residents felt uncomfortable reporting patient errors, possibly secondary to fear of retaliation, along with issues with the reporting system.12
It is interesting to note that a minority of residents indicated that they feel that PASS reports are filed as often as they should be (25.9% on first survey and 24.1% on second survey). This is concerning, as the data gathered through PASS reports is used to improve patient care. However, the percentage reported in our study, although low, is higher than that reported in a similar study involving patients with Medicare insurance, which showed that only 14% of patient safety events were reported.13 These results demonstrate that further interventions are necessary in order to ensure that a PASS report is filed each time a patient safety event occurs.
Another finding of note is that the majority of residents also feel that the process of filing a PASS report is too time consuming. The majority of residents who have completed a PASS report stated that it took them between 10 and 20 minutes to complete a PASS report, but those same individuals also feel that it should take < 10 minutes to complete a PASS report. This is an important issue for hospital systems to address. Reducing the time it takes to file a PASS report may facilitate an increase in the amount of PASS reports filed.
We administered our surveys using email outreach to residents asking them to complete an anonymous online survey regarding the PASS report system using the REDCap software system. Researchers have various ways of administering surveys, ranging from paper surveys, emails, and even mobile apps. One study showed that online surveys tend to have higher response rates compared to non-online surveys, such as paper surveys and telephone surveys, which is likely due to the ease of use of online surveys.14 At the same time, unsolicited email surveys have been shown to have a negative influence on response rates. Mobile apps are a new way of administering surveys. However, research has not found any significant difference in the time required to complete the survey using mobile apps compared to other forms of administering surveys. In addition, surveys using mobile apps did not have increased response rates compared to other forms of administering surveys.15
To increase the response rate of our surveys, we offered gift cards to the study population for completing the survey. Studies have shown that surveys that offer incentives tend to have higher response rates than surveys that do not.16 Also, in addition to serving as a method for gathering data from our study population, we used our surveys as an intervention to increase awareness of PASS reporting, as reported in other studies. For example, another study used the HABITS questionnaire to not only gather information about children’s diet, but also to promote behavioral change towards healthy eating habits.17
This study had several limitations. First, the study was conducted using an anonymous online survey, which means we could not clarify questions that residents found confusing or needed further explanation. For example, 17 residents indicated in the first survey that they knew how to PASS report, but 19 residents indicated in the same survey that they have filed a PASS report in the past.
A second limitation of the study was that fewer residents completed the second survey (29 of 54 eligible residents) compared to the first survey (54 of 80 eligible residents). This may have impacted the results of the analysis, as certain findings were not statistically significant, despite trends in the data.
A third limitation of the study is that not all of the residents that completed the first and second surveys completed the entire intervention. For example, some residents did not attend the didactic lecture discussing PASS reports, and as such may not have received the appropriate training prior to completing the second survey.
The findings from this study can be used by the residency programs at UH-RH and by residency programs across the country to improve the involvement and attitudes of residents in reporting errors in patient care. Hospital staff need to be encouraged and educated on how to better report patient errors and the importance of reporting these errors. It would benefit hospital systems to provide continued and targeted training to familiarize physicians with the process of reporting patient errors, and take steps to reduce the time it takes to report patient errors. By increasing the reporting of errors, hospitals will be able to improve patient care through initiatives aimed at preventing errors.
Conclusion
Residents play an important role in providing high-quality care for patients. Part of providing high-quality care is the reporting of errors in patient care when they occur. Physicians, and in particular, residents, have historically underreported errors in patient care. Part of this underreporting results from residents not knowing or understanding the process of filing a report and feeling that the reports could be used as a form of retaliation. For hospital systems to continue to improve patient care, it is important for residents to not only know how to report errors in patient care but to feel comfortable doing so.
Corresponding author: Andrew J. Chin, DO, MS, MPH, Department of Internal Medicine, Adelante Healthcare, 1705 W Main St, Mesa, AZ 85201; [email protected].
Financial disclosures: None.
Funding: This study was funded by a research grant provided by Lake Eric College of Osteopathic Medicine to Andrew J. Chin and Anish Bhakta.
1. Zallman L, Ma J, Xiao L, Lasser KE. Quality of US primary care delivered by resident and staff physicians. J Gen Intern Med. 2010;25(11):1193-1197.
2. Bagain JP. The future of graduate medical education: a systems-based approach to ensure patient safety. Acad Med. 2015;90(9):1199-1202.
3. Kachalia A, Kaufman SR, Boothman R, et al. Liability claims and costs before and after implementation of a medical disclosure program. Ann Intern Med. 2010;153(4):213-221.
4. Kaldjian LC, Jones EW, Wu BJ, et al. Reporting medical errors to improve patient safety: a survey of physicians in teaching hospitals. Arch Intern Med. 2008;168(1):40-46.
5. Rowin EJ, Lucier D, Pauker SG, et al. Does error and adverse event reporting by physicians and nurses differ? Jt Comm J Qual Patient Saf. 2008;34(9):537-545.
6. Turner DA, Bae J, Cheely G, et al. Improving resident and fellow engagement in patient safety through a graduate medical education incentive program. J Grad Med Educ. 2018;10(6):671-675.
7. Macht R, Balen A, McAneny D, Hess D. A multifaceted intervention to increase surgery resident engagement in reporting adverse events. J Surg Educ. 2015;72(6):e117-e122.
8. Scott DR, Weimer M, English C, et al. A novel approach to increase residents’ involvement in reporting adverse events. Acad Med. 2011;86(6):742-746.
9. Stewart DA, Junn J, Adams MA, et al. House staff participation in patient safety reporting: identification of predominant barriers and implementation of a pilot program. South Med J. 2016;109(7):395-400.
10. Vidyarthi AR, Green AL, Rosenbluth G, Baron RB. Engaging residents and fellows to improve institution-wide quality: the first six years of a novel financial incentive program. Acad Med. 2014;89(3):460-468.
11. Fok MC, Wong RY. Impact of a competency based curriculum on quality improvement among internal medicine residents. BMC Med Educ. 2014;14:252.
12. Wijesekera TP, Sanders L, Windish DM. Education and reporting of diagnostic errors among physicians in internal medicine training programs. JAMA Intern Med. 2018;178(11):1548-1549.
13. Levinson DR. Hospital incident reporting systems do not capture most patient harm. Washington, D.C.: U.S. Department of Health and Human Services Office of the Inspector General. January 2012. Report No. OEI-06-09-00091.
14. Evans JR, Mathur A. The value of online surveys. Internet Research. 2005;15(2):192-219.
15. Marcano Belisario JS, Jamsek J, Huckvale K, et al. Comparison of self‐administered survey questionnaire responses collected using mobile apps versus other methods. Cochrane Database of Syst Rev. 2015;7:MR000042.
16. Manfreda KL, Batagelj Z, Vehovar V. Design of web survey questionnaires: three basic experiments. J Comput Mediat Commun. 2002;7(3):JCMC731.
17. Wright ND, Groisman‐Perelstein AE, Wylie‐Rosett J, et al. A lifestyle assessment and intervention tool for pediatric weight management: the HABITS questionnaire. J Hum Nutr Diet. 2011;24(1):96-100.
From Adelante Healthcare, Mesa, AZ (Dr. Chin), University Hospitals of Cleveland, Cleveland, OH (Drs. Delozier, Bascug, Levine, Bejanishvili, and Wynbrandt and Janet C. Peachey, Rachel M. Cerminara, and Sharon M. Darkovich), and Houston Methodist Hospitals, Houston, TX (Dr. Bhakta).
Abstract
Background: Resident physicians play an active role in the reporting of errors that occur in patient care. Previous studies indicate that residents significantly underreport errors in patient care.
Methods: Fifty-four of 80 eligible residents enrolled at University Hospitals–Regional Hospitals (UH-RH) during the 2018-2019 academic year completed a survey assessing their knowledge and experience in completing Patient Advocacy and Shared Stories (PASS) reports, which serve as incident reports in the UH health system in reporting errors in patient care. A series of interventions aimed at educating residents about the PASS report system were then conducted. The 54 residents who completed the first survey received it again 4 months later.
Results: Residents demonstrated greater understanding of when filing PASS reports was appropriate after the intervention, as significantly more residents reported having been involved in a situation where they should have filed a PASS report but did not (P = 0.036).
Conclusion: In this study, residents often did not report errors in patient care because they simply did not know the process for doing so. In addition, many residents often felt that the reporting of patient errors could be used as a form of retaliation.
Keywords: resident physicians; quality improvement; high-value care; medical errors; patient safety.
Resident physicians play a critical role in patient care. Residents undergo extensive supervised training in order to one day be able to practice medicine in an unsupervised setting, with the goal of providing the highest quality of care possible. One study reported that primary care provided by residents in a training program is of similar or higher quality than that provided by attending physicians.1
Besides providing high-quality care, it is important that residents play an active role in the reporting of errors that occur regarding patient care as well as in identifying events that may compromise patient safety and quality.2 In fact, increased reporting of patient errors has been shown to decrease liability-related costs for hospitals.3 Unfortunately, physicians, and residents in particular, have historically been poor reporters of errors in patient care.4 This is especially true when comparing physicians to other health professionals, such as nurses, in error reporting.5
Several studies have examined the involvement of residents in reporting errors in patient care. One recent study showed that a graduate medical education financial incentive program significantly increased the number of patient safety events reported by residents and fellows.6 This study, along with several others, supports the concept of using incentives to help improve the reporting of errors in patient care for physicians in training.7-10 Another study used Quality Improvement Knowledge Assessment Tool (QIKAT) scores to assess quality improvement (QI) knowledge. The study demonstrated that self-assessment scores of QI skills using QIKAT scores improved following a targeted intervention.11 Because further information on the involvement and attitudes of residents in reporting errors in patient care is needed, University Hospitals of Cleveland (UH) designed and implemented a QI study during the 2018-2019 academic year. This prospective study used anonymous surveys to objectively examine the involvement and attitudes of residents in reporting errors in patient care.
Methods
The UH health system uses Patient Advocacy and Shared Stories (PASS) reports as incident reports to not only disclose errors in patient care but also to identify any events that may compromise patient safety and quality. Based on preliminary review, nurses, ancillary staff, and administrators file the majority of PASS reports.
The study group consisted of residents at University Hospitals–Regional Hospitals (UH-RH), which is comprised of 2 hospitals: University Hospitals–Richmond Medical Center (UH-RMC) and University Hospitals –Bedford Medical Center (UH-BMC). UH-RMC and UH-BMC are 2 medium-sized university-affiliated community hospitals located in the Cleveland metropolitan area in Northeast Ohio. Both serve as clinical training sites for Case Western Reserve University School of Medicine and Lake Erie College of Osteopathic Medicine, the latter of which helped fund this study. The study was submitted to the Institutional Review Board (IRB) of University Hospitals of Cleveland and granted “not human subjects research” status as a QI study.
Surveys
UH-RH offers residency programs in dermatology, emergency medicine, family medicine, internal medicine, orthopedic surgery, and physical medicine and rehabilitation, along with a 1-year transitional/preliminary year. A total of 80 residents enrolled at UH-RH during the 2018-2019 academic year. All 80 residents at UH-RH received an email in December 2018 asking them to complete an anonymous survey regarding the PASS report system. The survey was administered using the REDCap software system and consisted of 15 multiple-choice questions. As an incentive for completing the survey, residents were offered a $10 Amazon gift card. The gift cards were funded through a research grant from Lake Erie College of Osteopathic Medicine. Residents were given 1 week to complete the survey. At the end of the week, 54 of 80 residents completed the first survey.
Following the first survey, efforts were undertaken by the study authors, in conjunction with the quality improvement department at UH-RH, to educate residents about the PASS report system. These interventions included giving a lecture on the PASS report system during resident didactic sessions, sending an email to all residents about the PASS report system, and providing residents an opportunity to complete an optional online training course regarding the PASS report system. As an incentive for completing the online training course, residents were offered a $10 Amazon gift card. As before, the gift cards were funded through a research grant from Lake Erie College of Osteopathic Medicine.
A second survey was administered in April 2019, 4 months after the first survey. To determine whether the intervention made an impact on the involvement and attitudes of residents in the reporting errors in patient care, only residents who completed the first survey were sent the second survey. The second survey consisted of the same questions as the first survey and was also administered using the REDCap software system. As an incentive for completing the survey, residents were offered another $10 Amazon gift card, again were funded through a research grant from Lake Erie College of Osteopathic Medicine. Residents were given 1 week to complete the survey.
Analysis
Chi-square analyses were utilized to examine differences between preintervention and postintervention responses across categories. All analyses were conducted using R statistical software, version 3.6.1 (R Foundation for Statistical Computing).
Results
A total of 54 of 80 eligible residents responded to the first survey (Table). Twenty-nine of 54 eligible residents responded to the second survey. Postintervention, significantly more residents indicated being involved in a situation where they should have filed a PASS report but did not (58.6% vs 53.7%; P = 0.036). Improvement was seen in PASS knowledge postintervention, where fewer residents reported not knowing how to file a PASS report (31.5% vs 55.2%; P = 0.059). No other improvements were significant, nor were there significant differences in responses between any other categories pre- and postintervention.
Discussion
Errors in patient care are a common occurrence in the hospital setting. Reporting errors when they happen is important for hospitals to gain data and better care for patients, but studies show that patient errors are usually underreported. This is concerning, as data on errors and other aspects of patient care are needed to inform quality improvement programs.
This study measured residents’ attitudes and knowledge regarding the filing of a PASS report. It also aimed to increase both the frequency of and knowledge about filing a PASS report with interventions. The results from each survey indicated a statistically significant increase in knowledge of when to file a PASS report. In the first survey, 53.7% of residents responded they they were involved in an instance where they should have filed a PASS report but did not. In the second survey, 58.5% of residents reported being involved in an instance where they should have filed a PASS report but did not. This difference was statistically significant (P = 0.036), sugesting that the intervention was successful at increasing residents’ knowledge regarding PASS reports and the appropriate times to file a PASS report.
The survey results also showed a trend toward increasing aggregate knowledge level of how to file PASS reports on the first survey and second surveys (from 31.5% vs 55.2%. This demonstrates an increase in knowledge of how to file a PASS report among residents at our hospital after the intervention. It should be noted that the intervention that was performed in this study was simple, easy to perform, and can be completed at any hospital system that uses a similar system for reporting patient errors.
Another important trend indicating the effectiveness of the intervention was a 15% increase in knowledge of what the PASS report acronym stands for, along with a 13.1% aggregate increase in the number of residents who filed a PASS report. This indicated that residents may have wanted to file a PASS report previously but simply did not know how to until the intervention. In addition, there was also a decrease in the aggregate percentages of residents who had never filed a PASS report and an increase in how many PASS reports were filed.
While PASS reports are a great way for hospitals to gain data and insight into problems at their sites, there was also a negative view of PASS reports. For example, a large percentage of residents indicated that filing a PASS report would not make any difference and that PASS reports are often used as a form of retaliation, either against themselves as the submitter or the person(s) mentioned in the PASS report. More specifically, more than 50% of residents felt that PASS reports were sometimes or often used as a form of retaliation against others. While many residents correctly identified in the survey that PASS reports are not equivalent to a “write-up,” it is concerning that they still feel there is a strong potential for retaliation when filing a PASS report. This finding is unfortunate but matches the results of a multicenter study that found that 44.6% of residents felt uncomfortable reporting patient errors, possibly secondary to fear of retaliation, along with issues with the reporting system.12
It is interesting to note that a minority of residents indicated that they feel that PASS reports are filed as often as they should be (25.9% on first survey and 24.1% on second survey). This is concerning, as the data gathered through PASS reports is used to improve patient care. However, the percentage reported in our study, although low, is higher than that reported in a similar study involving patients with Medicare insurance, which showed that only 14% of patient safety events were reported.13 These results demonstrate that further interventions are necessary in order to ensure that a PASS report is filed each time a patient safety event occurs.
Another finding of note is that the majority of residents also feel that the process of filing a PASS report is too time consuming. The majority of residents who have completed a PASS report stated that it took them between 10 and 20 minutes to complete a PASS report, but those same individuals also feel that it should take < 10 minutes to complete a PASS report. This is an important issue for hospital systems to address. Reducing the time it takes to file a PASS report may facilitate an increase in the amount of PASS reports filed.
We administered our surveys using email outreach to residents asking them to complete an anonymous online survey regarding the PASS report system using the REDCap software system. Researchers have various ways of administering surveys, ranging from paper surveys, emails, and even mobile apps. One study showed that online surveys tend to have higher response rates compared to non-online surveys, such as paper surveys and telephone surveys, which is likely due to the ease of use of online surveys.14 At the same time, unsolicited email surveys have been shown to have a negative influence on response rates. Mobile apps are a new way of administering surveys. However, research has not found any significant difference in the time required to complete the survey using mobile apps compared to other forms of administering surveys. In addition, surveys using mobile apps did not have increased response rates compared to other forms of administering surveys.15
To increase the response rate of our surveys, we offered gift cards to the study population for completing the survey. Studies have shown that surveys that offer incentives tend to have higher response rates than surveys that do not.16 Also, in addition to serving as a method for gathering data from our study population, we used our surveys as an intervention to increase awareness of PASS reporting, as reported in other studies. For example, another study used the HABITS questionnaire to not only gather information about children’s diet, but also to promote behavioral change towards healthy eating habits.17
This study had several limitations. First, the study was conducted using an anonymous online survey, which means we could not clarify questions that residents found confusing or needed further explanation. For example, 17 residents indicated in the first survey that they knew how to PASS report, but 19 residents indicated in the same survey that they have filed a PASS report in the past.
A second limitation of the study was that fewer residents completed the second survey (29 of 54 eligible residents) compared to the first survey (54 of 80 eligible residents). This may have impacted the results of the analysis, as certain findings were not statistically significant, despite trends in the data.
A third limitation of the study is that not all of the residents that completed the first and second surveys completed the entire intervention. For example, some residents did not attend the didactic lecture discussing PASS reports, and as such may not have received the appropriate training prior to completing the second survey.
The findings from this study can be used by the residency programs at UH-RH and by residency programs across the country to improve the involvement and attitudes of residents in reporting errors in patient care. Hospital staff need to be encouraged and educated on how to better report patient errors and the importance of reporting these errors. It would benefit hospital systems to provide continued and targeted training to familiarize physicians with the process of reporting patient errors, and take steps to reduce the time it takes to report patient errors. By increasing the reporting of errors, hospitals will be able to improve patient care through initiatives aimed at preventing errors.
Conclusion
Residents play an important role in providing high-quality care for patients. Part of providing high-quality care is the reporting of errors in patient care when they occur. Physicians, and in particular, residents, have historically underreported errors in patient care. Part of this underreporting results from residents not knowing or understanding the process of filing a report and feeling that the reports could be used as a form of retaliation. For hospital systems to continue to improve patient care, it is important for residents to not only know how to report errors in patient care but to feel comfortable doing so.
Corresponding author: Andrew J. Chin, DO, MS, MPH, Department of Internal Medicine, Adelante Healthcare, 1705 W Main St, Mesa, AZ 85201; [email protected].
Financial disclosures: None.
Funding: This study was funded by a research grant provided by Lake Eric College of Osteopathic Medicine to Andrew J. Chin and Anish Bhakta.
From Adelante Healthcare, Mesa, AZ (Dr. Chin), University Hospitals of Cleveland, Cleveland, OH (Drs. Delozier, Bascug, Levine, Bejanishvili, and Wynbrandt and Janet C. Peachey, Rachel M. Cerminara, and Sharon M. Darkovich), and Houston Methodist Hospitals, Houston, TX (Dr. Bhakta).
Abstract
Background: Resident physicians play an active role in the reporting of errors that occur in patient care. Previous studies indicate that residents significantly underreport errors in patient care.
Methods: Fifty-four of 80 eligible residents enrolled at University Hospitals–Regional Hospitals (UH-RH) during the 2018-2019 academic year completed a survey assessing their knowledge and experience in completing Patient Advocacy and Shared Stories (PASS) reports, which serve as incident reports in the UH health system in reporting errors in patient care. A series of interventions aimed at educating residents about the PASS report system were then conducted. The 54 residents who completed the first survey received it again 4 months later.
Results: Residents demonstrated greater understanding of when filing PASS reports was appropriate after the intervention, as significantly more residents reported having been involved in a situation where they should have filed a PASS report but did not (P = 0.036).
Conclusion: In this study, residents often did not report errors in patient care because they simply did not know the process for doing so. In addition, many residents often felt that the reporting of patient errors could be used as a form of retaliation.
Keywords: resident physicians; quality improvement; high-value care; medical errors; patient safety.
Resident physicians play a critical role in patient care. Residents undergo extensive supervised training in order to one day be able to practice medicine in an unsupervised setting, with the goal of providing the highest quality of care possible. One study reported that primary care provided by residents in a training program is of similar or higher quality than that provided by attending physicians.1
Besides providing high-quality care, it is important that residents play an active role in the reporting of errors that occur regarding patient care as well as in identifying events that may compromise patient safety and quality.2 In fact, increased reporting of patient errors has been shown to decrease liability-related costs for hospitals.3 Unfortunately, physicians, and residents in particular, have historically been poor reporters of errors in patient care.4 This is especially true when comparing physicians to other health professionals, such as nurses, in error reporting.5
Several studies have examined the involvement of residents in reporting errors in patient care. One recent study showed that a graduate medical education financial incentive program significantly increased the number of patient safety events reported by residents and fellows.6 This study, along with several others, supports the concept of using incentives to help improve the reporting of errors in patient care for physicians in training.7-10 Another study used Quality Improvement Knowledge Assessment Tool (QIKAT) scores to assess quality improvement (QI) knowledge. The study demonstrated that self-assessment scores of QI skills using QIKAT scores improved following a targeted intervention.11 Because further information on the involvement and attitudes of residents in reporting errors in patient care is needed, University Hospitals of Cleveland (UH) designed and implemented a QI study during the 2018-2019 academic year. This prospective study used anonymous surveys to objectively examine the involvement and attitudes of residents in reporting errors in patient care.
Methods
The UH health system uses Patient Advocacy and Shared Stories (PASS) reports as incident reports to not only disclose errors in patient care but also to identify any events that may compromise patient safety and quality. Based on preliminary review, nurses, ancillary staff, and administrators file the majority of PASS reports.
The study group consisted of residents at University Hospitals–Regional Hospitals (UH-RH), which is comprised of 2 hospitals: University Hospitals–Richmond Medical Center (UH-RMC) and University Hospitals –Bedford Medical Center (UH-BMC). UH-RMC and UH-BMC are 2 medium-sized university-affiliated community hospitals located in the Cleveland metropolitan area in Northeast Ohio. Both serve as clinical training sites for Case Western Reserve University School of Medicine and Lake Erie College of Osteopathic Medicine, the latter of which helped fund this study. The study was submitted to the Institutional Review Board (IRB) of University Hospitals of Cleveland and granted “not human subjects research” status as a QI study.
Surveys
UH-RH offers residency programs in dermatology, emergency medicine, family medicine, internal medicine, orthopedic surgery, and physical medicine and rehabilitation, along with a 1-year transitional/preliminary year. A total of 80 residents enrolled at UH-RH during the 2018-2019 academic year. All 80 residents at UH-RH received an email in December 2018 asking them to complete an anonymous survey regarding the PASS report system. The survey was administered using the REDCap software system and consisted of 15 multiple-choice questions. As an incentive for completing the survey, residents were offered a $10 Amazon gift card. The gift cards were funded through a research grant from Lake Erie College of Osteopathic Medicine. Residents were given 1 week to complete the survey. At the end of the week, 54 of 80 residents completed the first survey.
Following the first survey, efforts were undertaken by the study authors, in conjunction with the quality improvement department at UH-RH, to educate residents about the PASS report system. These interventions included giving a lecture on the PASS report system during resident didactic sessions, sending an email to all residents about the PASS report system, and providing residents an opportunity to complete an optional online training course regarding the PASS report system. As an incentive for completing the online training course, residents were offered a $10 Amazon gift card. As before, the gift cards were funded through a research grant from Lake Erie College of Osteopathic Medicine.
A second survey was administered in April 2019, 4 months after the first survey. To determine whether the intervention made an impact on the involvement and attitudes of residents in the reporting errors in patient care, only residents who completed the first survey were sent the second survey. The second survey consisted of the same questions as the first survey and was also administered using the REDCap software system. As an incentive for completing the survey, residents were offered another $10 Amazon gift card, again were funded through a research grant from Lake Erie College of Osteopathic Medicine. Residents were given 1 week to complete the survey.
Analysis
Chi-square analyses were utilized to examine differences between preintervention and postintervention responses across categories. All analyses were conducted using R statistical software, version 3.6.1 (R Foundation for Statistical Computing).
Results
A total of 54 of 80 eligible residents responded to the first survey (Table). Twenty-nine of 54 eligible residents responded to the second survey. Postintervention, significantly more residents indicated being involved in a situation where they should have filed a PASS report but did not (58.6% vs 53.7%; P = 0.036). Improvement was seen in PASS knowledge postintervention, where fewer residents reported not knowing how to file a PASS report (31.5% vs 55.2%; P = 0.059). No other improvements were significant, nor were there significant differences in responses between any other categories pre- and postintervention.
Discussion
Errors in patient care are a common occurrence in the hospital setting. Reporting errors when they happen is important for hospitals to gain data and better care for patients, but studies show that patient errors are usually underreported. This is concerning, as data on errors and other aspects of patient care are needed to inform quality improvement programs.
This study measured residents’ attitudes and knowledge regarding the filing of a PASS report. It also aimed to increase both the frequency of and knowledge about filing a PASS report with interventions. The results from each survey indicated a statistically significant increase in knowledge of when to file a PASS report. In the first survey, 53.7% of residents responded they they were involved in an instance where they should have filed a PASS report but did not. In the second survey, 58.5% of residents reported being involved in an instance where they should have filed a PASS report but did not. This difference was statistically significant (P = 0.036), sugesting that the intervention was successful at increasing residents’ knowledge regarding PASS reports and the appropriate times to file a PASS report.
The survey results also showed a trend toward increasing aggregate knowledge level of how to file PASS reports on the first survey and second surveys (from 31.5% vs 55.2%. This demonstrates an increase in knowledge of how to file a PASS report among residents at our hospital after the intervention. It should be noted that the intervention that was performed in this study was simple, easy to perform, and can be completed at any hospital system that uses a similar system for reporting patient errors.
Another important trend indicating the effectiveness of the intervention was a 15% increase in knowledge of what the PASS report acronym stands for, along with a 13.1% aggregate increase in the number of residents who filed a PASS report. This indicated that residents may have wanted to file a PASS report previously but simply did not know how to until the intervention. In addition, there was also a decrease in the aggregate percentages of residents who had never filed a PASS report and an increase in how many PASS reports were filed.
While PASS reports are a great way for hospitals to gain data and insight into problems at their sites, there was also a negative view of PASS reports. For example, a large percentage of residents indicated that filing a PASS report would not make any difference and that PASS reports are often used as a form of retaliation, either against themselves as the submitter or the person(s) mentioned in the PASS report. More specifically, more than 50% of residents felt that PASS reports were sometimes or often used as a form of retaliation against others. While many residents correctly identified in the survey that PASS reports are not equivalent to a “write-up,” it is concerning that they still feel there is a strong potential for retaliation when filing a PASS report. This finding is unfortunate but matches the results of a multicenter study that found that 44.6% of residents felt uncomfortable reporting patient errors, possibly secondary to fear of retaliation, along with issues with the reporting system.12
It is interesting to note that a minority of residents indicated that they feel that PASS reports are filed as often as they should be (25.9% on first survey and 24.1% on second survey). This is concerning, as the data gathered through PASS reports is used to improve patient care. However, the percentage reported in our study, although low, is higher than that reported in a similar study involving patients with Medicare insurance, which showed that only 14% of patient safety events were reported.13 These results demonstrate that further interventions are necessary in order to ensure that a PASS report is filed each time a patient safety event occurs.
Another finding of note is that the majority of residents also feel that the process of filing a PASS report is too time consuming. The majority of residents who have completed a PASS report stated that it took them between 10 and 20 minutes to complete a PASS report, but those same individuals also feel that it should take < 10 minutes to complete a PASS report. This is an important issue for hospital systems to address. Reducing the time it takes to file a PASS report may facilitate an increase in the amount of PASS reports filed.
We administered our surveys using email outreach to residents asking them to complete an anonymous online survey regarding the PASS report system using the REDCap software system. Researchers have various ways of administering surveys, ranging from paper surveys, emails, and even mobile apps. One study showed that online surveys tend to have higher response rates compared to non-online surveys, such as paper surveys and telephone surveys, which is likely due to the ease of use of online surveys.14 At the same time, unsolicited email surveys have been shown to have a negative influence on response rates. Mobile apps are a new way of administering surveys. However, research has not found any significant difference in the time required to complete the survey using mobile apps compared to other forms of administering surveys. In addition, surveys using mobile apps did not have increased response rates compared to other forms of administering surveys.15
To increase the response rate of our surveys, we offered gift cards to the study population for completing the survey. Studies have shown that surveys that offer incentives tend to have higher response rates than surveys that do not.16 Also, in addition to serving as a method for gathering data from our study population, we used our surveys as an intervention to increase awareness of PASS reporting, as reported in other studies. For example, another study used the HABITS questionnaire to not only gather information about children’s diet, but also to promote behavioral change towards healthy eating habits.17
This study had several limitations. First, the study was conducted using an anonymous online survey, which means we could not clarify questions that residents found confusing or needed further explanation. For example, 17 residents indicated in the first survey that they knew how to PASS report, but 19 residents indicated in the same survey that they have filed a PASS report in the past.
A second limitation of the study was that fewer residents completed the second survey (29 of 54 eligible residents) compared to the first survey (54 of 80 eligible residents). This may have impacted the results of the analysis, as certain findings were not statistically significant, despite trends in the data.
A third limitation of the study is that not all of the residents that completed the first and second surveys completed the entire intervention. For example, some residents did not attend the didactic lecture discussing PASS reports, and as such may not have received the appropriate training prior to completing the second survey.
The findings from this study can be used by the residency programs at UH-RH and by residency programs across the country to improve the involvement and attitudes of residents in reporting errors in patient care. Hospital staff need to be encouraged and educated on how to better report patient errors and the importance of reporting these errors. It would benefit hospital systems to provide continued and targeted training to familiarize physicians with the process of reporting patient errors, and take steps to reduce the time it takes to report patient errors. By increasing the reporting of errors, hospitals will be able to improve patient care through initiatives aimed at preventing errors.
Conclusion
Residents play an important role in providing high-quality care for patients. Part of providing high-quality care is the reporting of errors in patient care when they occur. Physicians, and in particular, residents, have historically underreported errors in patient care. Part of this underreporting results from residents not knowing or understanding the process of filing a report and feeling that the reports could be used as a form of retaliation. For hospital systems to continue to improve patient care, it is important for residents to not only know how to report errors in patient care but to feel comfortable doing so.
Corresponding author: Andrew J. Chin, DO, MS, MPH, Department of Internal Medicine, Adelante Healthcare, 1705 W Main St, Mesa, AZ 85201; [email protected].
Financial disclosures: None.
Funding: This study was funded by a research grant provided by Lake Eric College of Osteopathic Medicine to Andrew J. Chin and Anish Bhakta.
1. Zallman L, Ma J, Xiao L, Lasser KE. Quality of US primary care delivered by resident and staff physicians. J Gen Intern Med. 2010;25(11):1193-1197.
2. Bagain JP. The future of graduate medical education: a systems-based approach to ensure patient safety. Acad Med. 2015;90(9):1199-1202.
3. Kachalia A, Kaufman SR, Boothman R, et al. Liability claims and costs before and after implementation of a medical disclosure program. Ann Intern Med. 2010;153(4):213-221.
4. Kaldjian LC, Jones EW, Wu BJ, et al. Reporting medical errors to improve patient safety: a survey of physicians in teaching hospitals. Arch Intern Med. 2008;168(1):40-46.
5. Rowin EJ, Lucier D, Pauker SG, et al. Does error and adverse event reporting by physicians and nurses differ? Jt Comm J Qual Patient Saf. 2008;34(9):537-545.
6. Turner DA, Bae J, Cheely G, et al. Improving resident and fellow engagement in patient safety through a graduate medical education incentive program. J Grad Med Educ. 2018;10(6):671-675.
7. Macht R, Balen A, McAneny D, Hess D. A multifaceted intervention to increase surgery resident engagement in reporting adverse events. J Surg Educ. 2015;72(6):e117-e122.
8. Scott DR, Weimer M, English C, et al. A novel approach to increase residents’ involvement in reporting adverse events. Acad Med. 2011;86(6):742-746.
9. Stewart DA, Junn J, Adams MA, et al. House staff participation in patient safety reporting: identification of predominant barriers and implementation of a pilot program. South Med J. 2016;109(7):395-400.
10. Vidyarthi AR, Green AL, Rosenbluth G, Baron RB. Engaging residents and fellows to improve institution-wide quality: the first six years of a novel financial incentive program. Acad Med. 2014;89(3):460-468.
11. Fok MC, Wong RY. Impact of a competency based curriculum on quality improvement among internal medicine residents. BMC Med Educ. 2014;14:252.
12. Wijesekera TP, Sanders L, Windish DM. Education and reporting of diagnostic errors among physicians in internal medicine training programs. JAMA Intern Med. 2018;178(11):1548-1549.
13. Levinson DR. Hospital incident reporting systems do not capture most patient harm. Washington, D.C.: U.S. Department of Health and Human Services Office of the Inspector General. January 2012. Report No. OEI-06-09-00091.
14. Evans JR, Mathur A. The value of online surveys. Internet Research. 2005;15(2):192-219.
15. Marcano Belisario JS, Jamsek J, Huckvale K, et al. Comparison of self‐administered survey questionnaire responses collected using mobile apps versus other methods. Cochrane Database of Syst Rev. 2015;7:MR000042.
16. Manfreda KL, Batagelj Z, Vehovar V. Design of web survey questionnaires: three basic experiments. J Comput Mediat Commun. 2002;7(3):JCMC731.
17. Wright ND, Groisman‐Perelstein AE, Wylie‐Rosett J, et al. A lifestyle assessment and intervention tool for pediatric weight management: the HABITS questionnaire. J Hum Nutr Diet. 2011;24(1):96-100.
1. Zallman L, Ma J, Xiao L, Lasser KE. Quality of US primary care delivered by resident and staff physicians. J Gen Intern Med. 2010;25(11):1193-1197.
2. Bagain JP. The future of graduate medical education: a systems-based approach to ensure patient safety. Acad Med. 2015;90(9):1199-1202.
3. Kachalia A, Kaufman SR, Boothman R, et al. Liability claims and costs before and after implementation of a medical disclosure program. Ann Intern Med. 2010;153(4):213-221.
4. Kaldjian LC, Jones EW, Wu BJ, et al. Reporting medical errors to improve patient safety: a survey of physicians in teaching hospitals. Arch Intern Med. 2008;168(1):40-46.
5. Rowin EJ, Lucier D, Pauker SG, et al. Does error and adverse event reporting by physicians and nurses differ? Jt Comm J Qual Patient Saf. 2008;34(9):537-545.
6. Turner DA, Bae J, Cheely G, et al. Improving resident and fellow engagement in patient safety through a graduate medical education incentive program. J Grad Med Educ. 2018;10(6):671-675.
7. Macht R, Balen A, McAneny D, Hess D. A multifaceted intervention to increase surgery resident engagement in reporting adverse events. J Surg Educ. 2015;72(6):e117-e122.
8. Scott DR, Weimer M, English C, et al. A novel approach to increase residents’ involvement in reporting adverse events. Acad Med. 2011;86(6):742-746.
9. Stewart DA, Junn J, Adams MA, et al. House staff participation in patient safety reporting: identification of predominant barriers and implementation of a pilot program. South Med J. 2016;109(7):395-400.
10. Vidyarthi AR, Green AL, Rosenbluth G, Baron RB. Engaging residents and fellows to improve institution-wide quality: the first six years of a novel financial incentive program. Acad Med. 2014;89(3):460-468.
11. Fok MC, Wong RY. Impact of a competency based curriculum on quality improvement among internal medicine residents. BMC Med Educ. 2014;14:252.
12. Wijesekera TP, Sanders L, Windish DM. Education and reporting of diagnostic errors among physicians in internal medicine training programs. JAMA Intern Med. 2018;178(11):1548-1549.
13. Levinson DR. Hospital incident reporting systems do not capture most patient harm. Washington, D.C.: U.S. Department of Health and Human Services Office of the Inspector General. January 2012. Report No. OEI-06-09-00091.
14. Evans JR, Mathur A. The value of online surveys. Internet Research. 2005;15(2):192-219.
15. Marcano Belisario JS, Jamsek J, Huckvale K, et al. Comparison of self‐administered survey questionnaire responses collected using mobile apps versus other methods. Cochrane Database of Syst Rev. 2015;7:MR000042.
16. Manfreda KL, Batagelj Z, Vehovar V. Design of web survey questionnaires: three basic experiments. J Comput Mediat Commun. 2002;7(3):JCMC731.
17. Wright ND, Groisman‐Perelstein AE, Wylie‐Rosett J, et al. A lifestyle assessment and intervention tool for pediatric weight management: the HABITS questionnaire. J Hum Nutr Diet. 2011;24(1):96-100.
Implementing the AMI READMITS Risk Assessment Score to Increase Referrals Among Patients With Type I Myocardial Infarction
From The Johns Hopkins Hospital, Baltimore, MD (Dr. Muganlinskaya and Dr. Skojec, retired); The George Washington University, Washington, DC (Dr. Posey); and Johns Hopkins University, Baltimore, MD (Dr. Resar).
Abstract
Objective: Assessing the risk characteristics of patients with acute myocardial infarction (MI) can help providers make appropriate referral decisions. This quality improvement project sought to improve timely, appropriate referrals among patients with type I MI by adding a risk assessment, the AMI READMITS score, to the existing referral protocol.
Methods: Patients’ chart data were analyzed to assess changes in referrals and timely follow-up appointments from pre-intervention to intervention. A survey assessed providers’ satisfaction with the new referral protocol.
Results: Among 57 patients (n = 29 preintervention; n = 28 intervention), documented referrals increased significantly from 66% to 89% (χ2 = 4.571, df = 1, P = 0.033); and timely appointments increased by 10%, which was not significant (χ2 = 3.550, df = 2, P = 0.169). Most providers agreed that the new protocol was easy to use, useful in making referral decisions, and improved the referral process. All agreed the risk score should be incorporated into electronic clinical notes. Provider opinions related to implementing the risk score in clinical practice were mixed. Qualitative feedback suggests this was due to limited validation of the AMI READMITS score in reducing readmissions.
Conclusions: Our risk-based referral protocol helped to increase appropriate referrals among patients with type I MI. Provider adoption may be enhanced by incorporating the protocol into electronic clinical notes. Research to further validate the accuracy of the AMI READMITS score in predicting readmissions may support adoption of the protocol in clinical practice.
Keywords: quality improvement; type I myocardial infarction; referral process; readmission risk; risk assessment; chart review.
Early follow-up after discharge is an important strategy to reduce the risk of unplanned hospital readmissions among patients with various conditions.1-3 While patient confounding factors, such as chronic health problems, environment, socioeconomic status, and literacy, make it difficult to avoid all unplanned readmissions, early follow-up may help providers identify and appropriately manage some health-related issues, and as such is a pivotal element of a readmission prevention strategy.4 There is evidence that patients with non-ST elevation myocardial infarction (NSTEMI) who have an outpatient appointment with a physician within 7 days after discharge have a lower risk of 30-day readmission.5
Our hospital’s postmyocardial infarction clinic was created to prevent unplanned readmissions within 30 days after discharge among patients with type I myocardial infarction (MI). Since inception, the number of referrals has been much lower than expected. In 2018, the total number of patients discharged from the hospital with type I MI and any troponin I level above 0.40 ng/mL was 313. Most of these patients were discharged from the hospital’s cardiac units; however, only 91 referrals were made. To increase referrals, the cardiology nurse practitioners (NPs) developed a post-MI referral protocol (Figure 1). However, this protocol was not consistently used and referrals to the clinic remained low.
Evidence-based risk assessment tools have the potential to increase effective patient management. For example, cardiology providers at the hospital utilize various scores, such as CHA2DS2-VASc6 and the Society of Thoracic Surgery risk score,7 to plan patient management. Among the scores used to predict unplanned readmissions for MI patients, the most promising is the AMI READMITS score.8 Unlike other nonspecific prediction models, the AMI READMITS score was developed based on variables extracted from the electronic health records (EHRs) of patients who were hospitalized for MI and readmitted within 30 days after discharge. Recognizing the potential to increase referrals by integrating an MI-specific risk assessment, this quality improvement study modified the existing referral protocol to include the patients’ AMI READMITS score and recommendations for follow-up.
Currently, there are no clear recommendations on how soon after discharge patients with MI should undergo follow-up. As research data vary, we selected 7 days follow-up for patients from high risk groups based on the “See you in 7” initiative for patients with heart failure (HF) and MI,9,10 as well as evidence that patients with NSTEMI have a lower risk of 30-day readmission if they have follow-up within 7 days after discharge5; and we selected 14 days follow-up for patients from low-risk groups based on evidence that postdischarge follow-up within 14 days reduces risk of 30-day readmission in patients with acute myocardial infarction (AMI) and/or acutely decompensated HF.11
Methods
This project was designed to answer the following question: For adult patients with type I MI, does implementation of a readmission risk assessment referral protocol increase the percentage of referrals and appointments scheduled within a recommended time? Anticipated outcomes included: (1) increased referrals to a cardiologist or the post-MI clinic; (2) increased scheduled follow-up appointments within 7 to 14 days; (3) provider satisfaction with the usability and usefulness of the new protocol; and (4) consistent provider adoption of the new risk assessment referral protocol.
To evaluate the degree to which these outcomes were achieved, we reviewed patient charts for 2 months prior and 2 months during implementation of the new referral protocol. As shown in Figure 2, the new protocol added the following process steps to the existing protocol: calculation of the AMI READMITS score, recommendations for follow-up based on patients’ risk score, and guidance to refer patients to the post-MI clinic if patients did not have an appointment with a cardiologist within 7 to 14 days after discharge. Patients’ risk assessment scores were obtained from forms completed by clinicians during the intervention. Clinician’s perceptions related to the usability and usefulness of the new protocol and feedback related to its long-term adoption were assessed using a descriptive survey.
The institutional review board classified this project as a quality improvement project. To avoid potential loss of patient privacy, no identifiable data were collected, a unique identifier unrelated to patients’ records was generated for each patient, and data were saved on a password-protected cardiology office computer.
Population
The project population included all adult patients (≥ 18 years old) with type I MI who were admitted or transferred to the hospital, had a percutaneous coronary intervention (PCI), or were managed without PCI and discharged from the hospital’s cardiac care unit (CCU) and progressive cardiac care unit (PCCU). The criteria for type I MI included the “detection of a rise and/or fall of cardiac troponin with at least 1 value above the 99th percentile and with at least 1 of the following: symptoms of acute myocardial ischemia; new ischemic electrocardiographic (ECG) changes; development of new pathological Q waves; imaging evidence of new loss of viable myocardium or new regional wall motion abnormality in a pattern consistent with an ischemic etiology; identification of a coronary thrombus by angiography including intracoronary imaging or by autopsy.”12 The study excluded patients with type I MI who were referred for coronary bypass surgery.
Intervention
The revised risk assessment protocol was implemented within the CCU and PCCU. The lead investigator met with each provider to discuss the role of the post-MI clinic, current referral rates, the purpose of the project, and the new referral process to be completed during the project for each patient discharged with type I MI. Cardiology NPs, fellows, and residents were asked to use the risk-assessment form to calculate patients’ risk for readmission, and refer patients to the post-MI clinic if an appointment with a cardiologist was not available within 7 to 14 days after discharge. Every week during the intervention phase, the investigator sent reminder emails to ensure form completion. Providers were asked to calculate and write the score, the discharge and referral dates, where referrals were made (a cardiologist or the post-MI clinic), date of appointment, and reason for not scheduling an appointment or not referring on the risk assessment form, and to drop the completed forms in specific labeled boxes located at the CCU and PCCU work stations. The investigator collected the completed forms weekly. When the number of discharged patients did not match the number of completed forms, the investigator followed up with discharging providers to understand why.
Data and Data Collection
Data to determine whether the use of the new protocol increased discharge referrals among patients with type I MI within the recommended timeframes were collected by electronic chart review. Data included discharging unit, patients’ age, gender, admission and discharge date, diagnosis, referral to a cardiologist and the post-MI clinic, and appointment date. Clinical data needed to calculate the AMI READMITS score was also collected: PCI within 24 hours, serum creatinine, systolic blood pressure (SBP), brain natriuretic peptide (BNP), and diabetes status.
Data to assess provider satisfaction with the usability and usefulness of the new protocol were gathered through an online survey. The survey included 1 question related to the providers’ role, 1 question asking whether they used the risk assessment for each patient, and 5 Likert-items assessing the ease of usage. An additional open-ended question asked providers to share feedback related to integrating the AMI READMITS risk assessment score to the post-MI referral protocol long term.
To evaluate how consistently providers utilized the new referral protocol when discharging patients with type I MI, the number of completed forms was compared with the number of those patients who were discharged.
Statistical Analysis
Descriptive statistics were used to summarize patient demographics and to calculate the frequency of referrals before and during the intervention. Chi-square statistics were calculated to determine whether the change in percentage of referrals and timely referrals was significant. Descriptive statistics were used to determine the level of provider satisfaction related to each survey item. A content analysis method was used to synthesize themes from the open-ended question asking clinicians to share their feedback related to the new protocol.
Results
Fifty-seven patients met the study inclusion criteria: 29 patients during the preintervention phase and 28 patients during the intervention phase. There were 35 male (61.4%) and 22 female (38.6%) patients. Twenty-five patients (43.9%) were from age groups 41 through 60 years and 61 through 80 years, respectively, representing the majority of included patients. Seven patients (12.3%) were from the 81 years and older age group. There were no patients in the age group 18 through 40 years. Based on the AMI READMITS score calculation, 57.9% (n = 33) patients were from a low-risk group (includes extremely low and low risk for readmission) and 42.1% (n = 24) were from a high-risk group (includes moderate, high, and extremely high risk for readmission).
Provider adoption of the new protocol during the intervention was high. Referral forms were completed for 82% (n = 23) of the 28 patients during the intervention. Analysis findings showed a statistically significant increase in documented referrals after implementing the new referral protocol. During the preintervention phase, 66% (n = 19) of patients with type I MI were referred to see a cardiologist or an NP at a post-MI clinic and there was no documented referral for 34% (n = 10) of patients. During the intervention phase, 89% (n = 25) of patients were referred and there was no documented referral for 11% (n = 3) of patients. Chi-square results indicated that the increase in referrals was significant (χ2 = 4.571, df = 1, P = 0.033).
Data analysis examined whether patient referrals fell within the recommended timeframe of 7 days for the high-risk group (included moderate-to-extremely high risk) and 14 days for the low-risk group (included low-to-extremely low risk). During the preintervention phase, 31% (n = 9) of patient referrals were scheduled as recommended; 28% (n = 8) of patient referrals were scheduled but delayed; and there was no referral date documented for 41% (n = 12) of patients. During the intervention phase, referrals scheduled as recommended increased to 53% (n = 15); 25% (n = 7) of referrals were scheduled but delayed; and there was no referral date documented for 21.4% (n = 6) of patients. The change in appointments scheduled as recommended was not significant (χ2 = 3.550, df = 2, P = 0.169).
Surveys were emailed to 25 cardiology fellows and 3 cardiology NPs who participated in this study. Eighteen of the 28 clinicians (15 cardiology fellows and 3 cardiology NPs) responded for a response rate of 64%. One of several residents who rotated through the CCU and PCCU during the intervention also completed the survey, for a total of 19 participants. When asked if the protocol was easy to use, 79% agreed or strongly agreed. Eighteen of the 19 participants (95%) agreed or strongly agreed that the protocol was useful in making referral decisions. Sixty-eight percent agreed or strongly agreed that the AMI READMITS risk assessment score improves referral process. All participants agreed or strongly agreed that there should be an option to incorporate the AMI READMITS risk assessment score into electronic clinical notes. When asked whether the AMI READMITS risk score should be implemented in clinical practice, responses were mixed (Figure 3). A common theme among the 4 participants who responded with comments was the need for additional data to validate the usefulness of the AMI READMITS to reduce readmissions. In addition, 1 participant commented that “manual calculation [of the risk score] is not ideal.”
Discussion
This project demonstrated that implementing an evidence-based referral protocol integrating the AMI-READMITS score can increase timely postdischarge referrals among patients with type I MI. The percentage of appropriately scheduled appointments increased during the intervention phase; however, a relatively high number of appointments were scheduled outside of the recommended timeframe, similar to preintervention. Thus, while the new protocol increased referrals and provider documentation of these referrals, it appears that challenges in scheduling timely referral appointments remained. This project did not examine the reasons for delayed appointments.
The survey findings indicated that providers were generally satisfied with the usability and usefulness of the new risk assessment protocol. A large majority agreed or strongly agreed that it was easy to use and useful in making referral decisions, and most agreed or strongly agreed that it improves the referral process. Mixed opinions regarding implementing the AMI READMITS score in clinical practice, combined with qualitative findings, suggest that a lack of external validation of the AMI READMITS presents a barrier to its long-term adoption. All providers who participated in the survey agreed or strongly agreed that the risk assessment should be incorporated into electronic clinical notes. We have begun the process of working with the EHR vendor to automate the AMI risk-assessment within the referral work-flow, which will provide an opportunity for a follow-up quality improvement study.
This quality improvement project has several limitations. First, it implemented a small change in 2 inpatient units at 1 hospital using a simple pre- posttest design. Therefore, the findings are not generalizable to other settings. Prior to the intervention, some referrals may have been made without documentation. While the authors were able to trace undocumented referrals for patients who were referred to the post-MI clinic or to a cardiologist affiliated with the hospital, some patients may have been referred to cardiologists who were not affiliated with the hospital. Another limitation was that the self-created provider survey used was not tested in other clinical settings; thus, it cannot be determined whether the sensitivity and specificity of the survey questions are high. In addition, the clinical providers who participated in the study knew the study team, which may have influenced their behavior during the study period. Furthermore, the identified improvement in clinicians’ referral practices may not be sustainable due to the complexity and effort required to manually calculate the risk score. This limitation could be eliminated by integrating the risk score calculation into the EHR.
Conclusion
Early follow-up after discharge plays an important role in supporting patients’ self-management of some risk factors (ie, diet, weight, and smoking) and identifying gaps in postdischarge care which may lead to readmission. This project provides evidence that integrating the AMI READMITS risk assessment score into the referral process can help to guide discharge decision-making and increase timely, appropriate referrals for patients with MI. Integration of a specific risk assessment, such as the AMI READMITS, within the post-MI referral protocol may help clinicians make more efficient, educated referral decisions. Future studies should explore more specifically how and why the new protocol impacts clinicians’ decision-making and behavior related to post-MI referrals. In addition, future studies should investigate challenges associated with scheduling postdischarge appointments. It will be important to investigate how integration of the new protocol within the EHR may increase efficiency, consistency, and provider satisfaction with the new referral process. Additional research investigating the effects of the AMI READMITS score on readmissions reduction will be important to promote long-term adoption of the improved referral protocol in clinical practice.
Acknowledgments: The authors thank Shelly Conaway, ANP-BC, MSN, Angela Street, ANP-BC, MSN, Andrew Geis, ACNP-BC, MSN, Richard P. Jones II, MD, Eunice Young, MD, Joy Rothwell, MSN, RN-BC, Allison Olazo, MBA, MSN, RN-BC, Elizabeth Heck, RN-BC, and Matthew Trojanowski, MHA, MS, RRT, CSSBB for their support of this study.
Corresponding author: Nailya Muganlinskaya, DNP, MPH, ACNP-BC, MSN, The Johns Hopkins Hospital, 1800 Orleans St, Baltimore, MD 21287; [email protected].
Financial disclosures: None.
1. Why it is important to improve care transitions? Society of Hospital Medicine. Accessed June 15, 2020. https://www.hospitalmedicine.org/clinical-topics/care-transitions/
2. Tong L, Arnold T, Yang J, et al. The association between outpatient follow-up visits and all-cause non-elective 30-day readmissions: a retrospective observational cohort study. PloS One. 2018;13(7):e0200691.
3. Jackson C, Shahsahebi M, Wedlake T, DuBard CA. Timeliness of outpatient follow-up: an evidence-based approach for planning after hospital discharge. Ann Fam Med. 2015;13(2):115-22.
4. Health Research & Educational Trust. Preventable Readmissions Change Package. American Hospital Association. Updated December 2015. Accessed June 10, 2020. https://www.aha.org/sites/default/files/hiin/HRETHEN_ChangePackage_Readmissions.pd
5. Tung Y-C, Chang G-M, Chang H-Y, Yu T-H. Relationship between early physician follow-up and 30-day readmission after acute myocardial infarction and heart failure. Plos One. 2017;12(1):e0170061.
6. Kaplan RM, Koehler J, Zieger PD, et al. Stroke risk as a function of atrial fibrillation duration and CHA2DS2-VASc score. Circulation. 2019;140(20):1639-46.
7. Balan P, Zhao Y, Johnson S, et al. The Society of Thoracic Surgery Risk Score as a predictor of 30-day mortality in transcatheter vs surgical aortic valve replacement: a single-center experience and its implications for the development of a TAVR risk-prediction model. J Invasive Cardiol. 2017;29(3):109-14.
8. Smith LN, Makam AN, Darden D, et al. Acute myocardial infarction readmission risk prediction models: A systematic review of model performance. Circ Cardiovasc Qual Outcomes9.9. 2018;11(1):e003885.
9. Baker H, Oliver-McNeil S, Deng L, Hummel SL. See you in 7: regional hospital collaboration and outcomes in Medicare heart failure patients. JACC Heart Fail. 2015;3(10):765-73.
10. Batten A, Jaeger C, Griffen D, et al. See you in 7: improving acute myocardial infarction follow-up care. BMJ Open Qual. 2018;7(2):e000296.
11. Lee DW, Armistead L, Coleman H, et al. Abstract 15387: Post-discharge follow-up within 14 days reduces 30-day hospital readmission rates in patients with acute myocardial infarction and/or acutely decompensated heart failure. Circulation. 2018;134 (1):A 15387.
12. Thygesen K, Alpert JS, Jaffe AS, et al. Fourth universal definition of myocardial infarction. Circulation. 2018;138 (20):e:618-51.
From The Johns Hopkins Hospital, Baltimore, MD (Dr. Muganlinskaya and Dr. Skojec, retired); The George Washington University, Washington, DC (Dr. Posey); and Johns Hopkins University, Baltimore, MD (Dr. Resar).
Abstract
Objective: Assessing the risk characteristics of patients with acute myocardial infarction (MI) can help providers make appropriate referral decisions. This quality improvement project sought to improve timely, appropriate referrals among patients with type I MI by adding a risk assessment, the AMI READMITS score, to the existing referral protocol.
Methods: Patients’ chart data were analyzed to assess changes in referrals and timely follow-up appointments from pre-intervention to intervention. A survey assessed providers’ satisfaction with the new referral protocol.
Results: Among 57 patients (n = 29 preintervention; n = 28 intervention), documented referrals increased significantly from 66% to 89% (χ2 = 4.571, df = 1, P = 0.033); and timely appointments increased by 10%, which was not significant (χ2 = 3.550, df = 2, P = 0.169). Most providers agreed that the new protocol was easy to use, useful in making referral decisions, and improved the referral process. All agreed the risk score should be incorporated into electronic clinical notes. Provider opinions related to implementing the risk score in clinical practice were mixed. Qualitative feedback suggests this was due to limited validation of the AMI READMITS score in reducing readmissions.
Conclusions: Our risk-based referral protocol helped to increase appropriate referrals among patients with type I MI. Provider adoption may be enhanced by incorporating the protocol into electronic clinical notes. Research to further validate the accuracy of the AMI READMITS score in predicting readmissions may support adoption of the protocol in clinical practice.
Keywords: quality improvement; type I myocardial infarction; referral process; readmission risk; risk assessment; chart review.
Early follow-up after discharge is an important strategy to reduce the risk of unplanned hospital readmissions among patients with various conditions.1-3 While patient confounding factors, such as chronic health problems, environment, socioeconomic status, and literacy, make it difficult to avoid all unplanned readmissions, early follow-up may help providers identify and appropriately manage some health-related issues, and as such is a pivotal element of a readmission prevention strategy.4 There is evidence that patients with non-ST elevation myocardial infarction (NSTEMI) who have an outpatient appointment with a physician within 7 days after discharge have a lower risk of 30-day readmission.5
Our hospital’s postmyocardial infarction clinic was created to prevent unplanned readmissions within 30 days after discharge among patients with type I myocardial infarction (MI). Since inception, the number of referrals has been much lower than expected. In 2018, the total number of patients discharged from the hospital with type I MI and any troponin I level above 0.40 ng/mL was 313. Most of these patients were discharged from the hospital’s cardiac units; however, only 91 referrals were made. To increase referrals, the cardiology nurse practitioners (NPs) developed a post-MI referral protocol (Figure 1). However, this protocol was not consistently used and referrals to the clinic remained low.
Evidence-based risk assessment tools have the potential to increase effective patient management. For example, cardiology providers at the hospital utilize various scores, such as CHA2DS2-VASc6 and the Society of Thoracic Surgery risk score,7 to plan patient management. Among the scores used to predict unplanned readmissions for MI patients, the most promising is the AMI READMITS score.8 Unlike other nonspecific prediction models, the AMI READMITS score was developed based on variables extracted from the electronic health records (EHRs) of patients who were hospitalized for MI and readmitted within 30 days after discharge. Recognizing the potential to increase referrals by integrating an MI-specific risk assessment, this quality improvement study modified the existing referral protocol to include the patients’ AMI READMITS score and recommendations for follow-up.
Currently, there are no clear recommendations on how soon after discharge patients with MI should undergo follow-up. As research data vary, we selected 7 days follow-up for patients from high risk groups based on the “See you in 7” initiative for patients with heart failure (HF) and MI,9,10 as well as evidence that patients with NSTEMI have a lower risk of 30-day readmission if they have follow-up within 7 days after discharge5; and we selected 14 days follow-up for patients from low-risk groups based on evidence that postdischarge follow-up within 14 days reduces risk of 30-day readmission in patients with acute myocardial infarction (AMI) and/or acutely decompensated HF.11
Methods
This project was designed to answer the following question: For adult patients with type I MI, does implementation of a readmission risk assessment referral protocol increase the percentage of referrals and appointments scheduled within a recommended time? Anticipated outcomes included: (1) increased referrals to a cardiologist or the post-MI clinic; (2) increased scheduled follow-up appointments within 7 to 14 days; (3) provider satisfaction with the usability and usefulness of the new protocol; and (4) consistent provider adoption of the new risk assessment referral protocol.
To evaluate the degree to which these outcomes were achieved, we reviewed patient charts for 2 months prior and 2 months during implementation of the new referral protocol. As shown in Figure 2, the new protocol added the following process steps to the existing protocol: calculation of the AMI READMITS score, recommendations for follow-up based on patients’ risk score, and guidance to refer patients to the post-MI clinic if patients did not have an appointment with a cardiologist within 7 to 14 days after discharge. Patients’ risk assessment scores were obtained from forms completed by clinicians during the intervention. Clinician’s perceptions related to the usability and usefulness of the new protocol and feedback related to its long-term adoption were assessed using a descriptive survey.
The institutional review board classified this project as a quality improvement project. To avoid potential loss of patient privacy, no identifiable data were collected, a unique identifier unrelated to patients’ records was generated for each patient, and data were saved on a password-protected cardiology office computer.
Population
The project population included all adult patients (≥ 18 years old) with type I MI who were admitted or transferred to the hospital, had a percutaneous coronary intervention (PCI), or were managed without PCI and discharged from the hospital’s cardiac care unit (CCU) and progressive cardiac care unit (PCCU). The criteria for type I MI included the “detection of a rise and/or fall of cardiac troponin with at least 1 value above the 99th percentile and with at least 1 of the following: symptoms of acute myocardial ischemia; new ischemic electrocardiographic (ECG) changes; development of new pathological Q waves; imaging evidence of new loss of viable myocardium or new regional wall motion abnormality in a pattern consistent with an ischemic etiology; identification of a coronary thrombus by angiography including intracoronary imaging or by autopsy.”12 The study excluded patients with type I MI who were referred for coronary bypass surgery.
Intervention
The revised risk assessment protocol was implemented within the CCU and PCCU. The lead investigator met with each provider to discuss the role of the post-MI clinic, current referral rates, the purpose of the project, and the new referral process to be completed during the project for each patient discharged with type I MI. Cardiology NPs, fellows, and residents were asked to use the risk-assessment form to calculate patients’ risk for readmission, and refer patients to the post-MI clinic if an appointment with a cardiologist was not available within 7 to 14 days after discharge. Every week during the intervention phase, the investigator sent reminder emails to ensure form completion. Providers were asked to calculate and write the score, the discharge and referral dates, where referrals were made (a cardiologist or the post-MI clinic), date of appointment, and reason for not scheduling an appointment or not referring on the risk assessment form, and to drop the completed forms in specific labeled boxes located at the CCU and PCCU work stations. The investigator collected the completed forms weekly. When the number of discharged patients did not match the number of completed forms, the investigator followed up with discharging providers to understand why.
Data and Data Collection
Data to determine whether the use of the new protocol increased discharge referrals among patients with type I MI within the recommended timeframes were collected by electronic chart review. Data included discharging unit, patients’ age, gender, admission and discharge date, diagnosis, referral to a cardiologist and the post-MI clinic, and appointment date. Clinical data needed to calculate the AMI READMITS score was also collected: PCI within 24 hours, serum creatinine, systolic blood pressure (SBP), brain natriuretic peptide (BNP), and diabetes status.
Data to assess provider satisfaction with the usability and usefulness of the new protocol were gathered through an online survey. The survey included 1 question related to the providers’ role, 1 question asking whether they used the risk assessment for each patient, and 5 Likert-items assessing the ease of usage. An additional open-ended question asked providers to share feedback related to integrating the AMI READMITS risk assessment score to the post-MI referral protocol long term.
To evaluate how consistently providers utilized the new referral protocol when discharging patients with type I MI, the number of completed forms was compared with the number of those patients who were discharged.
Statistical Analysis
Descriptive statistics were used to summarize patient demographics and to calculate the frequency of referrals before and during the intervention. Chi-square statistics were calculated to determine whether the change in percentage of referrals and timely referrals was significant. Descriptive statistics were used to determine the level of provider satisfaction related to each survey item. A content analysis method was used to synthesize themes from the open-ended question asking clinicians to share their feedback related to the new protocol.
Results
Fifty-seven patients met the study inclusion criteria: 29 patients during the preintervention phase and 28 patients during the intervention phase. There were 35 male (61.4%) and 22 female (38.6%) patients. Twenty-five patients (43.9%) were from age groups 41 through 60 years and 61 through 80 years, respectively, representing the majority of included patients. Seven patients (12.3%) were from the 81 years and older age group. There were no patients in the age group 18 through 40 years. Based on the AMI READMITS score calculation, 57.9% (n = 33) patients were from a low-risk group (includes extremely low and low risk for readmission) and 42.1% (n = 24) were from a high-risk group (includes moderate, high, and extremely high risk for readmission).
Provider adoption of the new protocol during the intervention was high. Referral forms were completed for 82% (n = 23) of the 28 patients during the intervention. Analysis findings showed a statistically significant increase in documented referrals after implementing the new referral protocol. During the preintervention phase, 66% (n = 19) of patients with type I MI were referred to see a cardiologist or an NP at a post-MI clinic and there was no documented referral for 34% (n = 10) of patients. During the intervention phase, 89% (n = 25) of patients were referred and there was no documented referral for 11% (n = 3) of patients. Chi-square results indicated that the increase in referrals was significant (χ2 = 4.571, df = 1, P = 0.033).
Data analysis examined whether patient referrals fell within the recommended timeframe of 7 days for the high-risk group (included moderate-to-extremely high risk) and 14 days for the low-risk group (included low-to-extremely low risk). During the preintervention phase, 31% (n = 9) of patient referrals were scheduled as recommended; 28% (n = 8) of patient referrals were scheduled but delayed; and there was no referral date documented for 41% (n = 12) of patients. During the intervention phase, referrals scheduled as recommended increased to 53% (n = 15); 25% (n = 7) of referrals were scheduled but delayed; and there was no referral date documented for 21.4% (n = 6) of patients. The change in appointments scheduled as recommended was not significant (χ2 = 3.550, df = 2, P = 0.169).
Surveys were emailed to 25 cardiology fellows and 3 cardiology NPs who participated in this study. Eighteen of the 28 clinicians (15 cardiology fellows and 3 cardiology NPs) responded for a response rate of 64%. One of several residents who rotated through the CCU and PCCU during the intervention also completed the survey, for a total of 19 participants. When asked if the protocol was easy to use, 79% agreed or strongly agreed. Eighteen of the 19 participants (95%) agreed or strongly agreed that the protocol was useful in making referral decisions. Sixty-eight percent agreed or strongly agreed that the AMI READMITS risk assessment score improves referral process. All participants agreed or strongly agreed that there should be an option to incorporate the AMI READMITS risk assessment score into electronic clinical notes. When asked whether the AMI READMITS risk score should be implemented in clinical practice, responses were mixed (Figure 3). A common theme among the 4 participants who responded with comments was the need for additional data to validate the usefulness of the AMI READMITS to reduce readmissions. In addition, 1 participant commented that “manual calculation [of the risk score] is not ideal.”
Discussion
This project demonstrated that implementing an evidence-based referral protocol integrating the AMI-READMITS score can increase timely postdischarge referrals among patients with type I MI. The percentage of appropriately scheduled appointments increased during the intervention phase; however, a relatively high number of appointments were scheduled outside of the recommended timeframe, similar to preintervention. Thus, while the new protocol increased referrals and provider documentation of these referrals, it appears that challenges in scheduling timely referral appointments remained. This project did not examine the reasons for delayed appointments.
The survey findings indicated that providers were generally satisfied with the usability and usefulness of the new risk assessment protocol. A large majority agreed or strongly agreed that it was easy to use and useful in making referral decisions, and most agreed or strongly agreed that it improves the referral process. Mixed opinions regarding implementing the AMI READMITS score in clinical practice, combined with qualitative findings, suggest that a lack of external validation of the AMI READMITS presents a barrier to its long-term adoption. All providers who participated in the survey agreed or strongly agreed that the risk assessment should be incorporated into electronic clinical notes. We have begun the process of working with the EHR vendor to automate the AMI risk-assessment within the referral work-flow, which will provide an opportunity for a follow-up quality improvement study.
This quality improvement project has several limitations. First, it implemented a small change in 2 inpatient units at 1 hospital using a simple pre- posttest design. Therefore, the findings are not generalizable to other settings. Prior to the intervention, some referrals may have been made without documentation. While the authors were able to trace undocumented referrals for patients who were referred to the post-MI clinic or to a cardiologist affiliated with the hospital, some patients may have been referred to cardiologists who were not affiliated with the hospital. Another limitation was that the self-created provider survey used was not tested in other clinical settings; thus, it cannot be determined whether the sensitivity and specificity of the survey questions are high. In addition, the clinical providers who participated in the study knew the study team, which may have influenced their behavior during the study period. Furthermore, the identified improvement in clinicians’ referral practices may not be sustainable due to the complexity and effort required to manually calculate the risk score. This limitation could be eliminated by integrating the risk score calculation into the EHR.
Conclusion
Early follow-up after discharge plays an important role in supporting patients’ self-management of some risk factors (ie, diet, weight, and smoking) and identifying gaps in postdischarge care which may lead to readmission. This project provides evidence that integrating the AMI READMITS risk assessment score into the referral process can help to guide discharge decision-making and increase timely, appropriate referrals for patients with MI. Integration of a specific risk assessment, such as the AMI READMITS, within the post-MI referral protocol may help clinicians make more efficient, educated referral decisions. Future studies should explore more specifically how and why the new protocol impacts clinicians’ decision-making and behavior related to post-MI referrals. In addition, future studies should investigate challenges associated with scheduling postdischarge appointments. It will be important to investigate how integration of the new protocol within the EHR may increase efficiency, consistency, and provider satisfaction with the new referral process. Additional research investigating the effects of the AMI READMITS score on readmissions reduction will be important to promote long-term adoption of the improved referral protocol in clinical practice.
Acknowledgments: The authors thank Shelly Conaway, ANP-BC, MSN, Angela Street, ANP-BC, MSN, Andrew Geis, ACNP-BC, MSN, Richard P. Jones II, MD, Eunice Young, MD, Joy Rothwell, MSN, RN-BC, Allison Olazo, MBA, MSN, RN-BC, Elizabeth Heck, RN-BC, and Matthew Trojanowski, MHA, MS, RRT, CSSBB for their support of this study.
Corresponding author: Nailya Muganlinskaya, DNP, MPH, ACNP-BC, MSN, The Johns Hopkins Hospital, 1800 Orleans St, Baltimore, MD 21287; [email protected].
Financial disclosures: None.
From The Johns Hopkins Hospital, Baltimore, MD (Dr. Muganlinskaya and Dr. Skojec, retired); The George Washington University, Washington, DC (Dr. Posey); and Johns Hopkins University, Baltimore, MD (Dr. Resar).
Abstract
Objective: Assessing the risk characteristics of patients with acute myocardial infarction (MI) can help providers make appropriate referral decisions. This quality improvement project sought to improve timely, appropriate referrals among patients with type I MI by adding a risk assessment, the AMI READMITS score, to the existing referral protocol.
Methods: Patients’ chart data were analyzed to assess changes in referrals and timely follow-up appointments from pre-intervention to intervention. A survey assessed providers’ satisfaction with the new referral protocol.
Results: Among 57 patients (n = 29 preintervention; n = 28 intervention), documented referrals increased significantly from 66% to 89% (χ2 = 4.571, df = 1, P = 0.033); and timely appointments increased by 10%, which was not significant (χ2 = 3.550, df = 2, P = 0.169). Most providers agreed that the new protocol was easy to use, useful in making referral decisions, and improved the referral process. All agreed the risk score should be incorporated into electronic clinical notes. Provider opinions related to implementing the risk score in clinical practice were mixed. Qualitative feedback suggests this was due to limited validation of the AMI READMITS score in reducing readmissions.
Conclusions: Our risk-based referral protocol helped to increase appropriate referrals among patients with type I MI. Provider adoption may be enhanced by incorporating the protocol into electronic clinical notes. Research to further validate the accuracy of the AMI READMITS score in predicting readmissions may support adoption of the protocol in clinical practice.
Keywords: quality improvement; type I myocardial infarction; referral process; readmission risk; risk assessment; chart review.
Early follow-up after discharge is an important strategy to reduce the risk of unplanned hospital readmissions among patients with various conditions.1-3 While patient confounding factors, such as chronic health problems, environment, socioeconomic status, and literacy, make it difficult to avoid all unplanned readmissions, early follow-up may help providers identify and appropriately manage some health-related issues, and as such is a pivotal element of a readmission prevention strategy.4 There is evidence that patients with non-ST elevation myocardial infarction (NSTEMI) who have an outpatient appointment with a physician within 7 days after discharge have a lower risk of 30-day readmission.5
Our hospital’s postmyocardial infarction clinic was created to prevent unplanned readmissions within 30 days after discharge among patients with type I myocardial infarction (MI). Since inception, the number of referrals has been much lower than expected. In 2018, the total number of patients discharged from the hospital with type I MI and any troponin I level above 0.40 ng/mL was 313. Most of these patients were discharged from the hospital’s cardiac units; however, only 91 referrals were made. To increase referrals, the cardiology nurse practitioners (NPs) developed a post-MI referral protocol (Figure 1). However, this protocol was not consistently used and referrals to the clinic remained low.
Evidence-based risk assessment tools have the potential to increase effective patient management. For example, cardiology providers at the hospital utilize various scores, such as CHA2DS2-VASc6 and the Society of Thoracic Surgery risk score,7 to plan patient management. Among the scores used to predict unplanned readmissions for MI patients, the most promising is the AMI READMITS score.8 Unlike other nonspecific prediction models, the AMI READMITS score was developed based on variables extracted from the electronic health records (EHRs) of patients who were hospitalized for MI and readmitted within 30 days after discharge. Recognizing the potential to increase referrals by integrating an MI-specific risk assessment, this quality improvement study modified the existing referral protocol to include the patients’ AMI READMITS score and recommendations for follow-up.
Currently, there are no clear recommendations on how soon after discharge patients with MI should undergo follow-up. As research data vary, we selected 7 days follow-up for patients from high risk groups based on the “See you in 7” initiative for patients with heart failure (HF) and MI,9,10 as well as evidence that patients with NSTEMI have a lower risk of 30-day readmission if they have follow-up within 7 days after discharge5; and we selected 14 days follow-up for patients from low-risk groups based on evidence that postdischarge follow-up within 14 days reduces risk of 30-day readmission in patients with acute myocardial infarction (AMI) and/or acutely decompensated HF.11
Methods
This project was designed to answer the following question: For adult patients with type I MI, does implementation of a readmission risk assessment referral protocol increase the percentage of referrals and appointments scheduled within a recommended time? Anticipated outcomes included: (1) increased referrals to a cardiologist or the post-MI clinic; (2) increased scheduled follow-up appointments within 7 to 14 days; (3) provider satisfaction with the usability and usefulness of the new protocol; and (4) consistent provider adoption of the new risk assessment referral protocol.
To evaluate the degree to which these outcomes were achieved, we reviewed patient charts for 2 months prior and 2 months during implementation of the new referral protocol. As shown in Figure 2, the new protocol added the following process steps to the existing protocol: calculation of the AMI READMITS score, recommendations for follow-up based on patients’ risk score, and guidance to refer patients to the post-MI clinic if patients did not have an appointment with a cardiologist within 7 to 14 days after discharge. Patients’ risk assessment scores were obtained from forms completed by clinicians during the intervention. Clinician’s perceptions related to the usability and usefulness of the new protocol and feedback related to its long-term adoption were assessed using a descriptive survey.
The institutional review board classified this project as a quality improvement project. To avoid potential loss of patient privacy, no identifiable data were collected, a unique identifier unrelated to patients’ records was generated for each patient, and data were saved on a password-protected cardiology office computer.
Population
The project population included all adult patients (≥ 18 years old) with type I MI who were admitted or transferred to the hospital, had a percutaneous coronary intervention (PCI), or were managed without PCI and discharged from the hospital’s cardiac care unit (CCU) and progressive cardiac care unit (PCCU). The criteria for type I MI included the “detection of a rise and/or fall of cardiac troponin with at least 1 value above the 99th percentile and with at least 1 of the following: symptoms of acute myocardial ischemia; new ischemic electrocardiographic (ECG) changes; development of new pathological Q waves; imaging evidence of new loss of viable myocardium or new regional wall motion abnormality in a pattern consistent with an ischemic etiology; identification of a coronary thrombus by angiography including intracoronary imaging or by autopsy.”12 The study excluded patients with type I MI who were referred for coronary bypass surgery.
Intervention
The revised risk assessment protocol was implemented within the CCU and PCCU. The lead investigator met with each provider to discuss the role of the post-MI clinic, current referral rates, the purpose of the project, and the new referral process to be completed during the project for each patient discharged with type I MI. Cardiology NPs, fellows, and residents were asked to use the risk-assessment form to calculate patients’ risk for readmission, and refer patients to the post-MI clinic if an appointment with a cardiologist was not available within 7 to 14 days after discharge. Every week during the intervention phase, the investigator sent reminder emails to ensure form completion. Providers were asked to calculate and write the score, the discharge and referral dates, where referrals were made (a cardiologist or the post-MI clinic), date of appointment, and reason for not scheduling an appointment or not referring on the risk assessment form, and to drop the completed forms in specific labeled boxes located at the CCU and PCCU work stations. The investigator collected the completed forms weekly. When the number of discharged patients did not match the number of completed forms, the investigator followed up with discharging providers to understand why.
Data and Data Collection
Data to determine whether the use of the new protocol increased discharge referrals among patients with type I MI within the recommended timeframes were collected by electronic chart review. Data included discharging unit, patients’ age, gender, admission and discharge date, diagnosis, referral to a cardiologist and the post-MI clinic, and appointment date. Clinical data needed to calculate the AMI READMITS score was also collected: PCI within 24 hours, serum creatinine, systolic blood pressure (SBP), brain natriuretic peptide (BNP), and diabetes status.
Data to assess provider satisfaction with the usability and usefulness of the new protocol were gathered through an online survey. The survey included 1 question related to the providers’ role, 1 question asking whether they used the risk assessment for each patient, and 5 Likert-items assessing the ease of usage. An additional open-ended question asked providers to share feedback related to integrating the AMI READMITS risk assessment score to the post-MI referral protocol long term.
To evaluate how consistently providers utilized the new referral protocol when discharging patients with type I MI, the number of completed forms was compared with the number of those patients who were discharged.
Statistical Analysis
Descriptive statistics were used to summarize patient demographics and to calculate the frequency of referrals before and during the intervention. Chi-square statistics were calculated to determine whether the change in percentage of referrals and timely referrals was significant. Descriptive statistics were used to determine the level of provider satisfaction related to each survey item. A content analysis method was used to synthesize themes from the open-ended question asking clinicians to share their feedback related to the new protocol.
Results
Fifty-seven patients met the study inclusion criteria: 29 patients during the preintervention phase and 28 patients during the intervention phase. There were 35 male (61.4%) and 22 female (38.6%) patients. Twenty-five patients (43.9%) were from age groups 41 through 60 years and 61 through 80 years, respectively, representing the majority of included patients. Seven patients (12.3%) were from the 81 years and older age group. There were no patients in the age group 18 through 40 years. Based on the AMI READMITS score calculation, 57.9% (n = 33) patients were from a low-risk group (includes extremely low and low risk for readmission) and 42.1% (n = 24) were from a high-risk group (includes moderate, high, and extremely high risk for readmission).
Provider adoption of the new protocol during the intervention was high. Referral forms were completed for 82% (n = 23) of the 28 patients during the intervention. Analysis findings showed a statistically significant increase in documented referrals after implementing the new referral protocol. During the preintervention phase, 66% (n = 19) of patients with type I MI were referred to see a cardiologist or an NP at a post-MI clinic and there was no documented referral for 34% (n = 10) of patients. During the intervention phase, 89% (n = 25) of patients were referred and there was no documented referral for 11% (n = 3) of patients. Chi-square results indicated that the increase in referrals was significant (χ2 = 4.571, df = 1, P = 0.033).
Data analysis examined whether patient referrals fell within the recommended timeframe of 7 days for the high-risk group (included moderate-to-extremely high risk) and 14 days for the low-risk group (included low-to-extremely low risk). During the preintervention phase, 31% (n = 9) of patient referrals were scheduled as recommended; 28% (n = 8) of patient referrals were scheduled but delayed; and there was no referral date documented for 41% (n = 12) of patients. During the intervention phase, referrals scheduled as recommended increased to 53% (n = 15); 25% (n = 7) of referrals were scheduled but delayed; and there was no referral date documented for 21.4% (n = 6) of patients. The change in appointments scheduled as recommended was not significant (χ2 = 3.550, df = 2, P = 0.169).
Surveys were emailed to 25 cardiology fellows and 3 cardiology NPs who participated in this study. Eighteen of the 28 clinicians (15 cardiology fellows and 3 cardiology NPs) responded for a response rate of 64%. One of several residents who rotated through the CCU and PCCU during the intervention also completed the survey, for a total of 19 participants. When asked if the protocol was easy to use, 79% agreed or strongly agreed. Eighteen of the 19 participants (95%) agreed or strongly agreed that the protocol was useful in making referral decisions. Sixty-eight percent agreed or strongly agreed that the AMI READMITS risk assessment score improves referral process. All participants agreed or strongly agreed that there should be an option to incorporate the AMI READMITS risk assessment score into electronic clinical notes. When asked whether the AMI READMITS risk score should be implemented in clinical practice, responses were mixed (Figure 3). A common theme among the 4 participants who responded with comments was the need for additional data to validate the usefulness of the AMI READMITS to reduce readmissions. In addition, 1 participant commented that “manual calculation [of the risk score] is not ideal.”
Discussion
This project demonstrated that implementing an evidence-based referral protocol integrating the AMI-READMITS score can increase timely postdischarge referrals among patients with type I MI. The percentage of appropriately scheduled appointments increased during the intervention phase; however, a relatively high number of appointments were scheduled outside of the recommended timeframe, similar to preintervention. Thus, while the new protocol increased referrals and provider documentation of these referrals, it appears that challenges in scheduling timely referral appointments remained. This project did not examine the reasons for delayed appointments.
The survey findings indicated that providers were generally satisfied with the usability and usefulness of the new risk assessment protocol. A large majority agreed or strongly agreed that it was easy to use and useful in making referral decisions, and most agreed or strongly agreed that it improves the referral process. Mixed opinions regarding implementing the AMI READMITS score in clinical practice, combined with qualitative findings, suggest that a lack of external validation of the AMI READMITS presents a barrier to its long-term adoption. All providers who participated in the survey agreed or strongly agreed that the risk assessment should be incorporated into electronic clinical notes. We have begun the process of working with the EHR vendor to automate the AMI risk-assessment within the referral work-flow, which will provide an opportunity for a follow-up quality improvement study.
This quality improvement project has several limitations. First, it implemented a small change in 2 inpatient units at 1 hospital using a simple pre- posttest design. Therefore, the findings are not generalizable to other settings. Prior to the intervention, some referrals may have been made without documentation. While the authors were able to trace undocumented referrals for patients who were referred to the post-MI clinic or to a cardiologist affiliated with the hospital, some patients may have been referred to cardiologists who were not affiliated with the hospital. Another limitation was that the self-created provider survey used was not tested in other clinical settings; thus, it cannot be determined whether the sensitivity and specificity of the survey questions are high. In addition, the clinical providers who participated in the study knew the study team, which may have influenced their behavior during the study period. Furthermore, the identified improvement in clinicians’ referral practices may not be sustainable due to the complexity and effort required to manually calculate the risk score. This limitation could be eliminated by integrating the risk score calculation into the EHR.
Conclusion
Early follow-up after discharge plays an important role in supporting patients’ self-management of some risk factors (ie, diet, weight, and smoking) and identifying gaps in postdischarge care which may lead to readmission. This project provides evidence that integrating the AMI READMITS risk assessment score into the referral process can help to guide discharge decision-making and increase timely, appropriate referrals for patients with MI. Integration of a specific risk assessment, such as the AMI READMITS, within the post-MI referral protocol may help clinicians make more efficient, educated referral decisions. Future studies should explore more specifically how and why the new protocol impacts clinicians’ decision-making and behavior related to post-MI referrals. In addition, future studies should investigate challenges associated with scheduling postdischarge appointments. It will be important to investigate how integration of the new protocol within the EHR may increase efficiency, consistency, and provider satisfaction with the new referral process. Additional research investigating the effects of the AMI READMITS score on readmissions reduction will be important to promote long-term adoption of the improved referral protocol in clinical practice.
Acknowledgments: The authors thank Shelly Conaway, ANP-BC, MSN, Angela Street, ANP-BC, MSN, Andrew Geis, ACNP-BC, MSN, Richard P. Jones II, MD, Eunice Young, MD, Joy Rothwell, MSN, RN-BC, Allison Olazo, MBA, MSN, RN-BC, Elizabeth Heck, RN-BC, and Matthew Trojanowski, MHA, MS, RRT, CSSBB for their support of this study.
Corresponding author: Nailya Muganlinskaya, DNP, MPH, ACNP-BC, MSN, The Johns Hopkins Hospital, 1800 Orleans St, Baltimore, MD 21287; [email protected].
Financial disclosures: None.
1. Why it is important to improve care transitions? Society of Hospital Medicine. Accessed June 15, 2020. https://www.hospitalmedicine.org/clinical-topics/care-transitions/
2. Tong L, Arnold T, Yang J, et al. The association between outpatient follow-up visits and all-cause non-elective 30-day readmissions: a retrospective observational cohort study. PloS One. 2018;13(7):e0200691.
3. Jackson C, Shahsahebi M, Wedlake T, DuBard CA. Timeliness of outpatient follow-up: an evidence-based approach for planning after hospital discharge. Ann Fam Med. 2015;13(2):115-22.
4. Health Research & Educational Trust. Preventable Readmissions Change Package. American Hospital Association. Updated December 2015. Accessed June 10, 2020. https://www.aha.org/sites/default/files/hiin/HRETHEN_ChangePackage_Readmissions.pd
5. Tung Y-C, Chang G-M, Chang H-Y, Yu T-H. Relationship between early physician follow-up and 30-day readmission after acute myocardial infarction and heart failure. Plos One. 2017;12(1):e0170061.
6. Kaplan RM, Koehler J, Zieger PD, et al. Stroke risk as a function of atrial fibrillation duration and CHA2DS2-VASc score. Circulation. 2019;140(20):1639-46.
7. Balan P, Zhao Y, Johnson S, et al. The Society of Thoracic Surgery Risk Score as a predictor of 30-day mortality in transcatheter vs surgical aortic valve replacement: a single-center experience and its implications for the development of a TAVR risk-prediction model. J Invasive Cardiol. 2017;29(3):109-14.
8. Smith LN, Makam AN, Darden D, et al. Acute myocardial infarction readmission risk prediction models: A systematic review of model performance. Circ Cardiovasc Qual Outcomes9.9. 2018;11(1):e003885.
9. Baker H, Oliver-McNeil S, Deng L, Hummel SL. See you in 7: regional hospital collaboration and outcomes in Medicare heart failure patients. JACC Heart Fail. 2015;3(10):765-73.
10. Batten A, Jaeger C, Griffen D, et al. See you in 7: improving acute myocardial infarction follow-up care. BMJ Open Qual. 2018;7(2):e000296.
11. Lee DW, Armistead L, Coleman H, et al. Abstract 15387: Post-discharge follow-up within 14 days reduces 30-day hospital readmission rates in patients with acute myocardial infarction and/or acutely decompensated heart failure. Circulation. 2018;134 (1):A 15387.
12. Thygesen K, Alpert JS, Jaffe AS, et al. Fourth universal definition of myocardial infarction. Circulation. 2018;138 (20):e:618-51.
1. Why it is important to improve care transitions? Society of Hospital Medicine. Accessed June 15, 2020. https://www.hospitalmedicine.org/clinical-topics/care-transitions/
2. Tong L, Arnold T, Yang J, et al. The association between outpatient follow-up visits and all-cause non-elective 30-day readmissions: a retrospective observational cohort study. PloS One. 2018;13(7):e0200691.
3. Jackson C, Shahsahebi M, Wedlake T, DuBard CA. Timeliness of outpatient follow-up: an evidence-based approach for planning after hospital discharge. Ann Fam Med. 2015;13(2):115-22.
4. Health Research & Educational Trust. Preventable Readmissions Change Package. American Hospital Association. Updated December 2015. Accessed June 10, 2020. https://www.aha.org/sites/default/files/hiin/HRETHEN_ChangePackage_Readmissions.pd
5. Tung Y-C, Chang G-M, Chang H-Y, Yu T-H. Relationship between early physician follow-up and 30-day readmission after acute myocardial infarction and heart failure. Plos One. 2017;12(1):e0170061.
6. Kaplan RM, Koehler J, Zieger PD, et al. Stroke risk as a function of atrial fibrillation duration and CHA2DS2-VASc score. Circulation. 2019;140(20):1639-46.
7. Balan P, Zhao Y, Johnson S, et al. The Society of Thoracic Surgery Risk Score as a predictor of 30-day mortality in transcatheter vs surgical aortic valve replacement: a single-center experience and its implications for the development of a TAVR risk-prediction model. J Invasive Cardiol. 2017;29(3):109-14.
8. Smith LN, Makam AN, Darden D, et al. Acute myocardial infarction readmission risk prediction models: A systematic review of model performance. Circ Cardiovasc Qual Outcomes9.9. 2018;11(1):e003885.
9. Baker H, Oliver-McNeil S, Deng L, Hummel SL. See you in 7: regional hospital collaboration and outcomes in Medicare heart failure patients. JACC Heart Fail. 2015;3(10):765-73.
10. Batten A, Jaeger C, Griffen D, et al. See you in 7: improving acute myocardial infarction follow-up care. BMJ Open Qual. 2018;7(2):e000296.
11. Lee DW, Armistead L, Coleman H, et al. Abstract 15387: Post-discharge follow-up within 14 days reduces 30-day hospital readmission rates in patients with acute myocardial infarction and/or acutely decompensated heart failure. Circulation. 2018;134 (1):A 15387.
12. Thygesen K, Alpert JS, Jaffe AS, et al. Fourth universal definition of myocardial infarction. Circulation. 2018;138 (20):e:618-51.
Senate confirms Murthy as Surgeon General
Seven Republicans – Bill Cassidy (La.), Susan Collins (Maine), Roger Marshall (Kan.), Susan Murkowski (Alaska), Rob Portman (Ohio), Mitt Romney (Utah), and Dan Sullivan (Alaska) – joined all the Democrats and independents in the 57-43 vote approving Dr. Murthy’s nomination.
Dr. Murthy, 43, previously served as the 19th Surgeon General, from December 2014 to April 2017, when he was asked to step down by President Donald J. Trump.
Surgeons General serve 4-year terms.
During his first tenure, Dr. Murthy issued the first-ever Surgeon General’s report on the crisis of addiction and issued a call to action to doctors to help battle the opioid crisis.
When Dr. Murthy was nominated by President-elect Joseph R. Biden Jr. in December, he was acting as cochair of the incoming administration’s COVID-19 transition advisory board.
Early in 2020, before the COVID-19 pandemic hit, Dr. Murthy published a timely book: “Together: The Healing Power of Human Connection in a Sometimes Lonely World”.
He earned his bachelor’s degree from Harvard and his MD and MBA degrees from Yale. He completed his internal medicine residency at Brigham and Women’s Hospital in Boston, where he also served as a hospitalist, and later joined Harvard Medical School as a faculty member in internal medicine.
He is married to Alice Chen, MD. The couple have two children.
A version of this article first appeared on WebMD.com.
Seven Republicans – Bill Cassidy (La.), Susan Collins (Maine), Roger Marshall (Kan.), Susan Murkowski (Alaska), Rob Portman (Ohio), Mitt Romney (Utah), and Dan Sullivan (Alaska) – joined all the Democrats and independents in the 57-43 vote approving Dr. Murthy’s nomination.
Dr. Murthy, 43, previously served as the 19th Surgeon General, from December 2014 to April 2017, when he was asked to step down by President Donald J. Trump.
Surgeons General serve 4-year terms.
During his first tenure, Dr. Murthy issued the first-ever Surgeon General’s report on the crisis of addiction and issued a call to action to doctors to help battle the opioid crisis.
When Dr. Murthy was nominated by President-elect Joseph R. Biden Jr. in December, he was acting as cochair of the incoming administration’s COVID-19 transition advisory board.
Early in 2020, before the COVID-19 pandemic hit, Dr. Murthy published a timely book: “Together: The Healing Power of Human Connection in a Sometimes Lonely World”.
He earned his bachelor’s degree from Harvard and his MD and MBA degrees from Yale. He completed his internal medicine residency at Brigham and Women’s Hospital in Boston, where he also served as a hospitalist, and later joined Harvard Medical School as a faculty member in internal medicine.
He is married to Alice Chen, MD. The couple have two children.
A version of this article first appeared on WebMD.com.
Seven Republicans – Bill Cassidy (La.), Susan Collins (Maine), Roger Marshall (Kan.), Susan Murkowski (Alaska), Rob Portman (Ohio), Mitt Romney (Utah), and Dan Sullivan (Alaska) – joined all the Democrats and independents in the 57-43 vote approving Dr. Murthy’s nomination.
Dr. Murthy, 43, previously served as the 19th Surgeon General, from December 2014 to April 2017, when he was asked to step down by President Donald J. Trump.
Surgeons General serve 4-year terms.
During his first tenure, Dr. Murthy issued the first-ever Surgeon General’s report on the crisis of addiction and issued a call to action to doctors to help battle the opioid crisis.
When Dr. Murthy was nominated by President-elect Joseph R. Biden Jr. in December, he was acting as cochair of the incoming administration’s COVID-19 transition advisory board.
Early in 2020, before the COVID-19 pandemic hit, Dr. Murthy published a timely book: “Together: The Healing Power of Human Connection in a Sometimes Lonely World”.
He earned his bachelor’s degree from Harvard and his MD and MBA degrees from Yale. He completed his internal medicine residency at Brigham and Women’s Hospital in Boston, where he also served as a hospitalist, and later joined Harvard Medical School as a faculty member in internal medicine.
He is married to Alice Chen, MD. The couple have two children.
A version of this article first appeared on WebMD.com.
Change is hard: Lessons from an EHR conversion
During this “go-live,” 5 hospitals and approximately 300 ambulatory service and physician practice locations made the transition, consolidating over 100 disparate electronic systems and dozens of interfaces into one world-class medical record.
If you’ve ever been part of such an event, you know it is anything but simple. On the contrary, it requires an enormous financial investment along with years of planning, hours of meetings, and months of training. No matter how much preparation goes into it, there are sure to be bumps along the way. It is a traumatic and stressful time for all involved, but the end result is well worth the effort. Still, there are lessons to be learned and wisdom to be gleaned, and this month we’d like to share a few that we found most important. We believe that many of these are useful lessons even to those who will never live through a go-live.
Safety always comes first
Patient safety is a term so often used that it has a tendency to be taken for granted. Health systems build processes and procedures to ensure safety – some even win awards and recognition for their efforts. But the best (and safest) health care institutions build patient safety into their cultures. More than just being taught to use checklists or buzzwords, the staff at these institutions are encouraged to put the welfare of patients first, making all other activities secondary to this pursuit. We had the opportunity to witness the benefits of such a culture during this go-live and were incredibly impressed with the results.
To be successful in an EHR transition of any magnitude, an organization needs to hold patient safety as a core value and provide its employees with the tools to execute on that value. This enables staff to prepare adequately and to identify risks and opportunities before the conversion takes place. Once go-live occurs, staff also must feel empowered to speak up when they identify problem areas that might jeopardize patients’ care. They also must be given a clear escalation path to ensure their voices can be heard. Most importantly, everyone must understand that the electronic health record itself is just one piece of a major operational change.
As workflows are modified to adapt to the new technology, unsafe processes should be called out and fixed quickly. While the EHR may offer the latest in decision support and system integration, no advancement in technology can make up for bad outcomes, nor justify processes that lead to patient harm.
Training is no substitute for good support
It takes a long time to train thousands of employees, especially when that training must occur during the era of social distancing in the midst of a pandemic. Still, even in the best of times, education should be married to hands-on experience in order to have a real impact. Unfortunately, this is extremely challenging.
Trainees forget much of what they’ve learned in the weeks or months between education and go-live, so they must be given immediately accessible support to bridge the gap. This is known as “at-the-elbow” (ATE) support, and as the name implies, it consists of individuals who are familiar with the new system and are always available to end users, answering their questions and helping them navigate. Since health care never sleeps, this support needs to be offered 24/7, and it should also be flexible and plentiful.
There are many areas that will require more support than anticipated to accommodate the number of clinical and other staff who will use the system, so support staff must be nimble and available for redeployment. In addition, ensuring high-quality support is essential. As many ATE experts are hired contractors, their knowledge base and communications skills can vary widely. Accountability is key, and end users should feel empowered to identify gaps in coverage and deficits in knowledge base in the ATE.
As employees become more familiar with the new system, the need for ATE will wane, but there will still be questions that arise for many weeks to months, and new EHR users will also be added all the time. A good after–go-live support system should remain available so clinical and clerical employees can get just-in-time assistance whenever they need it.
Users should be given clear expectations
Clinicians going through an EHR conversion may be frustrated to discover that the data transferred from their old system into the new one is not quite what they expected. While structured elements such as allergies and immunizations may transfer, unstructured patient histories may not come over at all.
There may be gaps in data, or the opposite may even be true: an overabundance of useless information may transfer over, leaving doctors with dozens of meaningless data points to sift through and eliminate to clean up the chart. This can be extremely time-consuming and discouraging and may jeopardize the success of the go-live.
Providers deserve clear expectations prior to conversion. They should be told what will and will not transfer and be informed that there will be extra work required for documentation at the outset. They may also want the option to preemptively reduce patient volumes to accommodate the additional effort involved in preparing charts. No matter what, this will be a heavy lift, and physicians should understand the implications long before go-live to prepare accordingly.
Old habits die hard
One of the most common complaints we’ve heard following EHR conversions is that “things just worked better in the old system.” We always respond with a question: “Were things better, or just different?” The truth may lie somewhere in the middle, but there is no question that muscle memory develops over many years, and change is difficult no matter how much better the new system is. Still, appropriate expectations, access to just-in-time support, and a continual focus on safety will ensure that the long-term benefits of a patient-centered and integrated electronic record will far outweigh the initial challenges of go-live.
Dr. Notte is a family physician and chief medical officer of Abington (Pa.) Hospital–Jefferson Health. Dr. Skolnik is professor of family and community medicine at Sidney Kimmel Medical College, Philadelphia, and associate director of the family medicine residency program at Abington Hospital–Jefferson Health. They have no conflicts related to the content of this piece.
During this “go-live,” 5 hospitals and approximately 300 ambulatory service and physician practice locations made the transition, consolidating over 100 disparate electronic systems and dozens of interfaces into one world-class medical record.
If you’ve ever been part of such an event, you know it is anything but simple. On the contrary, it requires an enormous financial investment along with years of planning, hours of meetings, and months of training. No matter how much preparation goes into it, there are sure to be bumps along the way. It is a traumatic and stressful time for all involved, but the end result is well worth the effort. Still, there are lessons to be learned and wisdom to be gleaned, and this month we’d like to share a few that we found most important. We believe that many of these are useful lessons even to those who will never live through a go-live.
Safety always comes first
Patient safety is a term so often used that it has a tendency to be taken for granted. Health systems build processes and procedures to ensure safety – some even win awards and recognition for their efforts. But the best (and safest) health care institutions build patient safety into their cultures. More than just being taught to use checklists or buzzwords, the staff at these institutions are encouraged to put the welfare of patients first, making all other activities secondary to this pursuit. We had the opportunity to witness the benefits of such a culture during this go-live and were incredibly impressed with the results.
To be successful in an EHR transition of any magnitude, an organization needs to hold patient safety as a core value and provide its employees with the tools to execute on that value. This enables staff to prepare adequately and to identify risks and opportunities before the conversion takes place. Once go-live occurs, staff also must feel empowered to speak up when they identify problem areas that might jeopardize patients’ care. They also must be given a clear escalation path to ensure their voices can be heard. Most importantly, everyone must understand that the electronic health record itself is just one piece of a major operational change.
As workflows are modified to adapt to the new technology, unsafe processes should be called out and fixed quickly. While the EHR may offer the latest in decision support and system integration, no advancement in technology can make up for bad outcomes, nor justify processes that lead to patient harm.
Training is no substitute for good support
It takes a long time to train thousands of employees, especially when that training must occur during the era of social distancing in the midst of a pandemic. Still, even in the best of times, education should be married to hands-on experience in order to have a real impact. Unfortunately, this is extremely challenging.
Trainees forget much of what they’ve learned in the weeks or months between education and go-live, so they must be given immediately accessible support to bridge the gap. This is known as “at-the-elbow” (ATE) support, and as the name implies, it consists of individuals who are familiar with the new system and are always available to end users, answering their questions and helping them navigate. Since health care never sleeps, this support needs to be offered 24/7, and it should also be flexible and plentiful.
There are many areas that will require more support than anticipated to accommodate the number of clinical and other staff who will use the system, so support staff must be nimble and available for redeployment. In addition, ensuring high-quality support is essential. As many ATE experts are hired contractors, their knowledge base and communications skills can vary widely. Accountability is key, and end users should feel empowered to identify gaps in coverage and deficits in knowledge base in the ATE.
As employees become more familiar with the new system, the need for ATE will wane, but there will still be questions that arise for many weeks to months, and new EHR users will also be added all the time. A good after–go-live support system should remain available so clinical and clerical employees can get just-in-time assistance whenever they need it.
Users should be given clear expectations
Clinicians going through an EHR conversion may be frustrated to discover that the data transferred from their old system into the new one is not quite what they expected. While structured elements such as allergies and immunizations may transfer, unstructured patient histories may not come over at all.
There may be gaps in data, or the opposite may even be true: an overabundance of useless information may transfer over, leaving doctors with dozens of meaningless data points to sift through and eliminate to clean up the chart. This can be extremely time-consuming and discouraging and may jeopardize the success of the go-live.
Providers deserve clear expectations prior to conversion. They should be told what will and will not transfer and be informed that there will be extra work required for documentation at the outset. They may also want the option to preemptively reduce patient volumes to accommodate the additional effort involved in preparing charts. No matter what, this will be a heavy lift, and physicians should understand the implications long before go-live to prepare accordingly.
Old habits die hard
One of the most common complaints we’ve heard following EHR conversions is that “things just worked better in the old system.” We always respond with a question: “Were things better, or just different?” The truth may lie somewhere in the middle, but there is no question that muscle memory develops over many years, and change is difficult no matter how much better the new system is. Still, appropriate expectations, access to just-in-time support, and a continual focus on safety will ensure that the long-term benefits of a patient-centered and integrated electronic record will far outweigh the initial challenges of go-live.
Dr. Notte is a family physician and chief medical officer of Abington (Pa.) Hospital–Jefferson Health. Dr. Skolnik is professor of family and community medicine at Sidney Kimmel Medical College, Philadelphia, and associate director of the family medicine residency program at Abington Hospital–Jefferson Health. They have no conflicts related to the content of this piece.
During this “go-live,” 5 hospitals and approximately 300 ambulatory service and physician practice locations made the transition, consolidating over 100 disparate electronic systems and dozens of interfaces into one world-class medical record.
If you’ve ever been part of such an event, you know it is anything but simple. On the contrary, it requires an enormous financial investment along with years of planning, hours of meetings, and months of training. No matter how much preparation goes into it, there are sure to be bumps along the way. It is a traumatic and stressful time for all involved, but the end result is well worth the effort. Still, there are lessons to be learned and wisdom to be gleaned, and this month we’d like to share a few that we found most important. We believe that many of these are useful lessons even to those who will never live through a go-live.
Safety always comes first
Patient safety is a term so often used that it has a tendency to be taken for granted. Health systems build processes and procedures to ensure safety – some even win awards and recognition for their efforts. But the best (and safest) health care institutions build patient safety into their cultures. More than just being taught to use checklists or buzzwords, the staff at these institutions are encouraged to put the welfare of patients first, making all other activities secondary to this pursuit. We had the opportunity to witness the benefits of such a culture during this go-live and were incredibly impressed with the results.
To be successful in an EHR transition of any magnitude, an organization needs to hold patient safety as a core value and provide its employees with the tools to execute on that value. This enables staff to prepare adequately and to identify risks and opportunities before the conversion takes place. Once go-live occurs, staff also must feel empowered to speak up when they identify problem areas that might jeopardize patients’ care. They also must be given a clear escalation path to ensure their voices can be heard. Most importantly, everyone must understand that the electronic health record itself is just one piece of a major operational change.
As workflows are modified to adapt to the new technology, unsafe processes should be called out and fixed quickly. While the EHR may offer the latest in decision support and system integration, no advancement in technology can make up for bad outcomes, nor justify processes that lead to patient harm.
Training is no substitute for good support
It takes a long time to train thousands of employees, especially when that training must occur during the era of social distancing in the midst of a pandemic. Still, even in the best of times, education should be married to hands-on experience in order to have a real impact. Unfortunately, this is extremely challenging.
Trainees forget much of what they’ve learned in the weeks or months between education and go-live, so they must be given immediately accessible support to bridge the gap. This is known as “at-the-elbow” (ATE) support, and as the name implies, it consists of individuals who are familiar with the new system and are always available to end users, answering their questions and helping them navigate. Since health care never sleeps, this support needs to be offered 24/7, and it should also be flexible and plentiful.
There are many areas that will require more support than anticipated to accommodate the number of clinical and other staff who will use the system, so support staff must be nimble and available for redeployment. In addition, ensuring high-quality support is essential. As many ATE experts are hired contractors, their knowledge base and communications skills can vary widely. Accountability is key, and end users should feel empowered to identify gaps in coverage and deficits in knowledge base in the ATE.
As employees become more familiar with the new system, the need for ATE will wane, but there will still be questions that arise for many weeks to months, and new EHR users will also be added all the time. A good after–go-live support system should remain available so clinical and clerical employees can get just-in-time assistance whenever they need it.
Users should be given clear expectations
Clinicians going through an EHR conversion may be frustrated to discover that the data transferred from their old system into the new one is not quite what they expected. While structured elements such as allergies and immunizations may transfer, unstructured patient histories may not come over at all.
There may be gaps in data, or the opposite may even be true: an overabundance of useless information may transfer over, leaving doctors with dozens of meaningless data points to sift through and eliminate to clean up the chart. This can be extremely time-consuming and discouraging and may jeopardize the success of the go-live.
Providers deserve clear expectations prior to conversion. They should be told what will and will not transfer and be informed that there will be extra work required for documentation at the outset. They may also want the option to preemptively reduce patient volumes to accommodate the additional effort involved in preparing charts. No matter what, this will be a heavy lift, and physicians should understand the implications long before go-live to prepare accordingly.
Old habits die hard
One of the most common complaints we’ve heard following EHR conversions is that “things just worked better in the old system.” We always respond with a question: “Were things better, or just different?” The truth may lie somewhere in the middle, but there is no question that muscle memory develops over many years, and change is difficult no matter how much better the new system is. Still, appropriate expectations, access to just-in-time support, and a continual focus on safety will ensure that the long-term benefits of a patient-centered and integrated electronic record will far outweigh the initial challenges of go-live.
Dr. Notte is a family physician and chief medical officer of Abington (Pa.) Hospital–Jefferson Health. Dr. Skolnik is professor of family and community medicine at Sidney Kimmel Medical College, Philadelphia, and associate director of the family medicine residency program at Abington Hospital–Jefferson Health. They have no conflicts related to the content of this piece.
2021 match sets records: Who matched and who didn’t?
A total of 38,106 positions were offered, up 850 spots (2.3%) from 2020. Of those, 35,194 were first-year (PGY-1) positions, which was 928 more than the previous year (2.7%). A record 5,915 programs were part of the Match, 88 more than 2020.
“The application and recruitment cycle was upended as a result of the pandemic, yet the results of the Match continue to demonstrate strong and consistent outcomes for participants,” Donna L. Lamb, DHSc, MBA, BSN, NRMP president and CEO, said in a new release.
The report comes amid a year of Zoom interview fatigue, canceled testing, and virus fears and work-arounds, which the NMRP has never had to wrestle with since it was established in 1952.
Despite challenges, fill rates increased across the board. Of the 38,106 total positions offered, 36,179 were filled, representing a 2.6% increase over 2020. Of the 35,194 first-year positions available, 33,535 were filled, representing a 2.9% increase.
Those rates drove the percentage of all positions filled to 94.9% (up from 94.6%) and the percentage of PGY-1 positions filled to 94.8% (also up from 94.6%). There were 1,927 unfilled positions, a decline of 71 (3.6%) from 2020.
Primary care results strong
Of the first-year positions offered, 17,649 (49.6%) were in family medicine, internal medicine, and pediatrics. That’s an increase of 514 positions (3%) over 2020.
Of first-year positions offered in 2021, 16,860 (95.5%) were filled. U.S. seniors took 11,013 (65.3%) of those slots; that represents a slight decline (0.3%) from 2020. Family medicine saw a gain of 63 U.S. MD seniors who matched, and internal medicine saw a gain of 93 U.S. DO seniors who matched.
Some specialties filled all positions
PGY-1 specialties with 30 positions or more that filled all available positions include dermatology, medicine – emergency medicine, medicine – pediatrics, neurologic surgery, otolaryngology, integrated plastic surgery, and vascular surgery.*
PGY-1 specialties with 30 positions or more that filled more than 90% with U.S. seniors include dermatology (100%), medicine – emergency medicine (93.6%), medicine – pediatrics (93.5%), otolaryngology (93.2%), orthopedic surgery (92.8%), and integrated plastic surgery (90.4%).*
PGY-1 specialties with at least 30 positions that filled less than 50% with U.S. seniors include pathology (41.4 %) and surgery–preliminary (28%).
The number of U.S. citizen international medical graduates who submitted rank-ordered lists was 5,295, an increase of 128 (2.5%) over 2020 and the highest in 6 years; 3,152 of them matched to first-year positions, down two PGY-1 matched applicants over last year.
Full data are available on the NRMP’s website.
Correction, 3/22/21: An earlier version of this article misstated the affected specialties.
A version of this article first appeared on Medscape.com.
A total of 38,106 positions were offered, up 850 spots (2.3%) from 2020. Of those, 35,194 were first-year (PGY-1) positions, which was 928 more than the previous year (2.7%). A record 5,915 programs were part of the Match, 88 more than 2020.
“The application and recruitment cycle was upended as a result of the pandemic, yet the results of the Match continue to demonstrate strong and consistent outcomes for participants,” Donna L. Lamb, DHSc, MBA, BSN, NRMP president and CEO, said in a new release.
The report comes amid a year of Zoom interview fatigue, canceled testing, and virus fears and work-arounds, which the NMRP has never had to wrestle with since it was established in 1952.
Despite challenges, fill rates increased across the board. Of the 38,106 total positions offered, 36,179 were filled, representing a 2.6% increase over 2020. Of the 35,194 first-year positions available, 33,535 were filled, representing a 2.9% increase.
Those rates drove the percentage of all positions filled to 94.9% (up from 94.6%) and the percentage of PGY-1 positions filled to 94.8% (also up from 94.6%). There were 1,927 unfilled positions, a decline of 71 (3.6%) from 2020.
Primary care results strong
Of the first-year positions offered, 17,649 (49.6%) were in family medicine, internal medicine, and pediatrics. That’s an increase of 514 positions (3%) over 2020.
Of first-year positions offered in 2021, 16,860 (95.5%) were filled. U.S. seniors took 11,013 (65.3%) of those slots; that represents a slight decline (0.3%) from 2020. Family medicine saw a gain of 63 U.S. MD seniors who matched, and internal medicine saw a gain of 93 U.S. DO seniors who matched.
Some specialties filled all positions
PGY-1 specialties with 30 positions or more that filled all available positions include dermatology, medicine – emergency medicine, medicine – pediatrics, neurologic surgery, otolaryngology, integrated plastic surgery, and vascular surgery.*
PGY-1 specialties with 30 positions or more that filled more than 90% with U.S. seniors include dermatology (100%), medicine – emergency medicine (93.6%), medicine – pediatrics (93.5%), otolaryngology (93.2%), orthopedic surgery (92.8%), and integrated plastic surgery (90.4%).*
PGY-1 specialties with at least 30 positions that filled less than 50% with U.S. seniors include pathology (41.4 %) and surgery–preliminary (28%).
The number of U.S. citizen international medical graduates who submitted rank-ordered lists was 5,295, an increase of 128 (2.5%) over 2020 and the highest in 6 years; 3,152 of them matched to first-year positions, down two PGY-1 matched applicants over last year.
Full data are available on the NRMP’s website.
Correction, 3/22/21: An earlier version of this article misstated the affected specialties.
A version of this article first appeared on Medscape.com.
A total of 38,106 positions were offered, up 850 spots (2.3%) from 2020. Of those, 35,194 were first-year (PGY-1) positions, which was 928 more than the previous year (2.7%). A record 5,915 programs were part of the Match, 88 more than 2020.
“The application and recruitment cycle was upended as a result of the pandemic, yet the results of the Match continue to demonstrate strong and consistent outcomes for participants,” Donna L. Lamb, DHSc, MBA, BSN, NRMP president and CEO, said in a new release.
The report comes amid a year of Zoom interview fatigue, canceled testing, and virus fears and work-arounds, which the NMRP has never had to wrestle with since it was established in 1952.
Despite challenges, fill rates increased across the board. Of the 38,106 total positions offered, 36,179 were filled, representing a 2.6% increase over 2020. Of the 35,194 first-year positions available, 33,535 were filled, representing a 2.9% increase.
Those rates drove the percentage of all positions filled to 94.9% (up from 94.6%) and the percentage of PGY-1 positions filled to 94.8% (also up from 94.6%). There were 1,927 unfilled positions, a decline of 71 (3.6%) from 2020.
Primary care results strong
Of the first-year positions offered, 17,649 (49.6%) were in family medicine, internal medicine, and pediatrics. That’s an increase of 514 positions (3%) over 2020.
Of first-year positions offered in 2021, 16,860 (95.5%) were filled. U.S. seniors took 11,013 (65.3%) of those slots; that represents a slight decline (0.3%) from 2020. Family medicine saw a gain of 63 U.S. MD seniors who matched, and internal medicine saw a gain of 93 U.S. DO seniors who matched.
Some specialties filled all positions
PGY-1 specialties with 30 positions or more that filled all available positions include dermatology, medicine – emergency medicine, medicine – pediatrics, neurologic surgery, otolaryngology, integrated plastic surgery, and vascular surgery.*
PGY-1 specialties with 30 positions or more that filled more than 90% with U.S. seniors include dermatology (100%), medicine – emergency medicine (93.6%), medicine – pediatrics (93.5%), otolaryngology (93.2%), orthopedic surgery (92.8%), and integrated plastic surgery (90.4%).*
PGY-1 specialties with at least 30 positions that filled less than 50% with U.S. seniors include pathology (41.4 %) and surgery–preliminary (28%).
The number of U.S. citizen international medical graduates who submitted rank-ordered lists was 5,295, an increase of 128 (2.5%) over 2020 and the highest in 6 years; 3,152 of them matched to first-year positions, down two PGY-1 matched applicants over last year.
Full data are available on the NRMP’s website.
Correction, 3/22/21: An earlier version of this article misstated the affected specialties.
A version of this article first appeared on Medscape.com.
Is pediatric subspecialty training financially worth it?
Pursuing fellowship training is often financially costly in terms of lifetime earnings, compared with starting a career as a general pediatrician immediately after residency, a report suggests.
Researchers found that most pediatric subspecialists – including those practicing neurology, pulmonology, and adolescent medicine – do not see a financial return from additional training because of the delays in receiving increased compensation and the repayment of educational debt.
“Most pediatric subspecialists don’t experience a relative increase in compensation after training compared to a general pediatrician, so there isn’t a financial benefit to additional training,” lead author Eva Catenaccio, MD, from the division of pediatric neurology, department of neurology, Johns Hopkins University, Baltimore, told this news organization.
The findings, published online March 8 in Pediatrics, contribute to the ongoing debate about the length of pediatric fellowship training programs. The data also provide evidence for the potential effect of a pediatric subspecialty loan repayment program.
Pediatric subspecialty training rarely pays off
However, not all practitioners in pediatric subspecialties would find themselves in the red relative to their generalist peers. Three subspecialties had a positive financial return: cardiology, critical care, and neonatology. Dr. Catenaccio explained that this may be because these subspecialties tend to be “inpatient procedure oriented, which are often more [lucrative] than outpatient cognitive–oriented subspecialties, such as pediatric infectious diseases, endocrinology, or adolescent medicine.”
Enrolling in a pediatric fellowship program resulted in lifetime financial returns that ranged from an increase of $852,129 for cardiology, relative to general pediatrics, to a loss of $1,594,366 for adolescent medicine, researchers found.
For the study, researchers calculated the financial returns of 15 pediatric subspecialties – emergency medicine, neurology, cardiology, critical care, neonatology, hematology and oncology, pulmonology, hospitalist medicine, allergy and immunology, gastroenterology, rheumatology, nephrology, adolescent medicine, infectious diseases, and endocrinology – in comparison with returns of private practice general pediatrics on the basis of 2018-2019 data on fellowship stipends, compensation, and educational debt.
They obtained most of the data from the Association of American Medical Colleges Survey of Resident/Fellow Stipends and Benefits, AAMC’s annual Medical School Faculty Salary Report, and the AAMC Medical School Graduation Questionnaire.
Richard Mink, MD, department of pediatrics, Harbor-UCLA Medical Center, Torrance, Calif., noted that it would have been helpful to have also compared the lifetime earnings of practitioners in pediatric subspecialties to academic general pediatricians and not just those in private practice.
The financial gap has worsened
To better understand which aspects of fellowship training have the greatest effect on lifetime compensation, Dr. Catenaccio and colleagues evaluated the potential effects of shortening fellowship length, eliminating school debt, and implementing a federal loan repayment plan. These changes enhanced the returns of cardiology, critical care, and neonatology – subspecialties that had already seen financial returns before these changes – and resulted in a positive financial return for emergency medicine.
The changes also narrowed the financial gap between subspecialties and general pediatrics. However, the remaining subspecialties still earned less than private practice pediatrics.
The new study is an update to a 2011 report, which reflected 2007-2008 data for 11 subspecialties. This time around, the researchers included the subspecialty of hospitalist medicine, which was approved as a board-certified subspecialty by the American Board of Pediatrics in 2014, as well as neurology, allergy and immunology, and adolescent medicine.
“I was most surprised that the additional pediatric subspecialties we included since the 2011 report followed the same general trend, with pediatric subspecialty training having a lower lifetime earning potential than general pediatrics,” Dr. Catenaccio said.
Comparing results from the two study periods showed that the financial gap between general pediatrics and subspecialty pediatrics worsened over time. For example, the financial return for pediatric endocrinology decreased an additional $500,000 between 2007 and 2018.
The researchers believe a combination of increased educational debt burden, slow growth in compensation, and changing interest rates over time have caused the financial differences between general pediatrics and subspecialty pediatrics to become more pronounced.
‘Pediatric subspecialty training is worth it!’
Despite the financial gaps, Dr. Catenaccio and colleagues say pediatric subspecialty training is still worthwhile but that policymakers should address these financial differences to help guide workforce distribution in a way that meets the needs of patients.
“I think pediatric subspecialty training is worth it,” said Dr. Catenaccio, who’s pursuing pediatric subspecialty training. “There are so many factors that go into choosing a specialty or subspecialty in medicine, including the desire to care for a particular patient population, interest in certain diseases or organ systems, lifestyle considerations, and research opportunities.”
But it’s also important for trainees to be aware of economic considerations in their decision-making.
Dr. Mink, who wrote an accompanying commentary, agrees that young clinicians should not make career decisions on the basis of metrics such as lifetime earning measures.
“I think people who go into pediatrics have decided that money is not the driving force,” said Dr. Mink. He noted that pediatricians are usually not paid well, compared with other specialists. “To me the important thing is you have to like what you’re doing.”
A 2020 study found that trainees who chose a career in pediatric pulmonology, a subspecialty, said that financial considerations were not the driving factor in their decision-making. Nevertheless, Dr. Mink also believes young clinicians should take into account their educational debt.
The further widening of the financial gap between general pediatrics and pediatric subspecialties could lead to shortages in the pediatric subspecialty workforce.
The authors and Dr. Mink have disclosed no relevant financial relationships.
A version of this article first appeared on Medscape.com.
Pursuing fellowship training is often financially costly in terms of lifetime earnings, compared with starting a career as a general pediatrician immediately after residency, a report suggests.
Researchers found that most pediatric subspecialists – including those practicing neurology, pulmonology, and adolescent medicine – do not see a financial return from additional training because of the delays in receiving increased compensation and the repayment of educational debt.
“Most pediatric subspecialists don’t experience a relative increase in compensation after training compared to a general pediatrician, so there isn’t a financial benefit to additional training,” lead author Eva Catenaccio, MD, from the division of pediatric neurology, department of neurology, Johns Hopkins University, Baltimore, told this news organization.
The findings, published online March 8 in Pediatrics, contribute to the ongoing debate about the length of pediatric fellowship training programs. The data also provide evidence for the potential effect of a pediatric subspecialty loan repayment program.
Pediatric subspecialty training rarely pays off
However, not all practitioners in pediatric subspecialties would find themselves in the red relative to their generalist peers. Three subspecialties had a positive financial return: cardiology, critical care, and neonatology. Dr. Catenaccio explained that this may be because these subspecialties tend to be “inpatient procedure oriented, which are often more [lucrative] than outpatient cognitive–oriented subspecialties, such as pediatric infectious diseases, endocrinology, or adolescent medicine.”
Enrolling in a pediatric fellowship program resulted in lifetime financial returns that ranged from an increase of $852,129 for cardiology, relative to general pediatrics, to a loss of $1,594,366 for adolescent medicine, researchers found.
For the study, researchers calculated the financial returns of 15 pediatric subspecialties – emergency medicine, neurology, cardiology, critical care, neonatology, hematology and oncology, pulmonology, hospitalist medicine, allergy and immunology, gastroenterology, rheumatology, nephrology, adolescent medicine, infectious diseases, and endocrinology – in comparison with returns of private practice general pediatrics on the basis of 2018-2019 data on fellowship stipends, compensation, and educational debt.
They obtained most of the data from the Association of American Medical Colleges Survey of Resident/Fellow Stipends and Benefits, AAMC’s annual Medical School Faculty Salary Report, and the AAMC Medical School Graduation Questionnaire.
Richard Mink, MD, department of pediatrics, Harbor-UCLA Medical Center, Torrance, Calif., noted that it would have been helpful to have also compared the lifetime earnings of practitioners in pediatric subspecialties to academic general pediatricians and not just those in private practice.
The financial gap has worsened
To better understand which aspects of fellowship training have the greatest effect on lifetime compensation, Dr. Catenaccio and colleagues evaluated the potential effects of shortening fellowship length, eliminating school debt, and implementing a federal loan repayment plan. These changes enhanced the returns of cardiology, critical care, and neonatology – subspecialties that had already seen financial returns before these changes – and resulted in a positive financial return for emergency medicine.
The changes also narrowed the financial gap between subspecialties and general pediatrics. However, the remaining subspecialties still earned less than private practice pediatrics.
The new study is an update to a 2011 report, which reflected 2007-2008 data for 11 subspecialties. This time around, the researchers included the subspecialty of hospitalist medicine, which was approved as a board-certified subspecialty by the American Board of Pediatrics in 2014, as well as neurology, allergy and immunology, and adolescent medicine.
“I was most surprised that the additional pediatric subspecialties we included since the 2011 report followed the same general trend, with pediatric subspecialty training having a lower lifetime earning potential than general pediatrics,” Dr. Catenaccio said.
Comparing results from the two study periods showed that the financial gap between general pediatrics and subspecialty pediatrics worsened over time. For example, the financial return for pediatric endocrinology decreased an additional $500,000 between 2007 and 2018.
The researchers believe a combination of increased educational debt burden, slow growth in compensation, and changing interest rates over time have caused the financial differences between general pediatrics and subspecialty pediatrics to become more pronounced.
‘Pediatric subspecialty training is worth it!’
Despite the financial gaps, Dr. Catenaccio and colleagues say pediatric subspecialty training is still worthwhile but that policymakers should address these financial differences to help guide workforce distribution in a way that meets the needs of patients.
“I think pediatric subspecialty training is worth it,” said Dr. Catenaccio, who’s pursuing pediatric subspecialty training. “There are so many factors that go into choosing a specialty or subspecialty in medicine, including the desire to care for a particular patient population, interest in certain diseases or organ systems, lifestyle considerations, and research opportunities.”
But it’s also important for trainees to be aware of economic considerations in their decision-making.
Dr. Mink, who wrote an accompanying commentary, agrees that young clinicians should not make career decisions on the basis of metrics such as lifetime earning measures.
“I think people who go into pediatrics have decided that money is not the driving force,” said Dr. Mink. He noted that pediatricians are usually not paid well, compared with other specialists. “To me the important thing is you have to like what you’re doing.”
A 2020 study found that trainees who chose a career in pediatric pulmonology, a subspecialty, said that financial considerations were not the driving factor in their decision-making. Nevertheless, Dr. Mink also believes young clinicians should take into account their educational debt.
The further widening of the financial gap between general pediatrics and pediatric subspecialties could lead to shortages in the pediatric subspecialty workforce.
The authors and Dr. Mink have disclosed no relevant financial relationships.
A version of this article first appeared on Medscape.com.
Pursuing fellowship training is often financially costly in terms of lifetime earnings, compared with starting a career as a general pediatrician immediately after residency, a report suggests.
Researchers found that most pediatric subspecialists – including those practicing neurology, pulmonology, and adolescent medicine – do not see a financial return from additional training because of the delays in receiving increased compensation and the repayment of educational debt.
“Most pediatric subspecialists don’t experience a relative increase in compensation after training compared to a general pediatrician, so there isn’t a financial benefit to additional training,” lead author Eva Catenaccio, MD, from the division of pediatric neurology, department of neurology, Johns Hopkins University, Baltimore, told this news organization.
The findings, published online March 8 in Pediatrics, contribute to the ongoing debate about the length of pediatric fellowship training programs. The data also provide evidence for the potential effect of a pediatric subspecialty loan repayment program.
Pediatric subspecialty training rarely pays off
However, not all practitioners in pediatric subspecialties would find themselves in the red relative to their generalist peers. Three subspecialties had a positive financial return: cardiology, critical care, and neonatology. Dr. Catenaccio explained that this may be because these subspecialties tend to be “inpatient procedure oriented, which are often more [lucrative] than outpatient cognitive–oriented subspecialties, such as pediatric infectious diseases, endocrinology, or adolescent medicine.”
Enrolling in a pediatric fellowship program resulted in lifetime financial returns that ranged from an increase of $852,129 for cardiology, relative to general pediatrics, to a loss of $1,594,366 for adolescent medicine, researchers found.
For the study, researchers calculated the financial returns of 15 pediatric subspecialties – emergency medicine, neurology, cardiology, critical care, neonatology, hematology and oncology, pulmonology, hospitalist medicine, allergy and immunology, gastroenterology, rheumatology, nephrology, adolescent medicine, infectious diseases, and endocrinology – in comparison with returns of private practice general pediatrics on the basis of 2018-2019 data on fellowship stipends, compensation, and educational debt.
They obtained most of the data from the Association of American Medical Colleges Survey of Resident/Fellow Stipends and Benefits, AAMC’s annual Medical School Faculty Salary Report, and the AAMC Medical School Graduation Questionnaire.
Richard Mink, MD, department of pediatrics, Harbor-UCLA Medical Center, Torrance, Calif., noted that it would have been helpful to have also compared the lifetime earnings of practitioners in pediatric subspecialties to academic general pediatricians and not just those in private practice.
The financial gap has worsened
To better understand which aspects of fellowship training have the greatest effect on lifetime compensation, Dr. Catenaccio and colleagues evaluated the potential effects of shortening fellowship length, eliminating school debt, and implementing a federal loan repayment plan. These changes enhanced the returns of cardiology, critical care, and neonatology – subspecialties that had already seen financial returns before these changes – and resulted in a positive financial return for emergency medicine.
The changes also narrowed the financial gap between subspecialties and general pediatrics. However, the remaining subspecialties still earned less than private practice pediatrics.
The new study is an update to a 2011 report, which reflected 2007-2008 data for 11 subspecialties. This time around, the researchers included the subspecialty of hospitalist medicine, which was approved as a board-certified subspecialty by the American Board of Pediatrics in 2014, as well as neurology, allergy and immunology, and adolescent medicine.
“I was most surprised that the additional pediatric subspecialties we included since the 2011 report followed the same general trend, with pediatric subspecialty training having a lower lifetime earning potential than general pediatrics,” Dr. Catenaccio said.
Comparing results from the two study periods showed that the financial gap between general pediatrics and subspecialty pediatrics worsened over time. For example, the financial return for pediatric endocrinology decreased an additional $500,000 between 2007 and 2018.
The researchers believe a combination of increased educational debt burden, slow growth in compensation, and changing interest rates over time have caused the financial differences between general pediatrics and subspecialty pediatrics to become more pronounced.
‘Pediatric subspecialty training is worth it!’
Despite the financial gaps, Dr. Catenaccio and colleagues say pediatric subspecialty training is still worthwhile but that policymakers should address these financial differences to help guide workforce distribution in a way that meets the needs of patients.
“I think pediatric subspecialty training is worth it,” said Dr. Catenaccio, who’s pursuing pediatric subspecialty training. “There are so many factors that go into choosing a specialty or subspecialty in medicine, including the desire to care for a particular patient population, interest in certain diseases or organ systems, lifestyle considerations, and research opportunities.”
But it’s also important for trainees to be aware of economic considerations in their decision-making.
Dr. Mink, who wrote an accompanying commentary, agrees that young clinicians should not make career decisions on the basis of metrics such as lifetime earning measures.
“I think people who go into pediatrics have decided that money is not the driving force,” said Dr. Mink. He noted that pediatricians are usually not paid well, compared with other specialists. “To me the important thing is you have to like what you’re doing.”
A 2020 study found that trainees who chose a career in pediatric pulmonology, a subspecialty, said that financial considerations were not the driving factor in their decision-making. Nevertheless, Dr. Mink also believes young clinicians should take into account their educational debt.
The further widening of the financial gap between general pediatrics and pediatric subspecialties could lead to shortages in the pediatric subspecialty workforce.
The authors and Dr. Mink have disclosed no relevant financial relationships.
A version of this article first appeared on Medscape.com.
Have you used ambulatory cervical ripening in your practice?
[polldaddy:10771173]
[polldaddy:10771173]
[polldaddy:10771173]
A return of holism? It never left osteopathic medicine
I enjoyed Dr. Jonas’s article, “A new model of care to return holism to family medicine” (J Fam Pract. 2020;69:493-498).
However, I wanted to point out that for more than 100 years the concept of the patient-centered medical home, and the outgrowth of that, has been part of osteopathic medical education, founded by A.T. Still, MD, in the 1800s.
Congratulations to the allopathic medicine profession for recognizing its significance.
Steven Shapiro, DO
Fenton, MI
I enjoyed Dr. Jonas’s article, “A new model of care to return holism to family medicine” (J Fam Pract. 2020;69:493-498).
However, I wanted to point out that for more than 100 years the concept of the patient-centered medical home, and the outgrowth of that, has been part of osteopathic medical education, founded by A.T. Still, MD, in the 1800s.
Congratulations to the allopathic medicine profession for recognizing its significance.
Steven Shapiro, DO
Fenton, MI
I enjoyed Dr. Jonas’s article, “A new model of care to return holism to family medicine” (J Fam Pract. 2020;69:493-498).
However, I wanted to point out that for more than 100 years the concept of the patient-centered medical home, and the outgrowth of that, has been part of osteopathic medical education, founded by A.T. Still, MD, in the 1800s.
Congratulations to the allopathic medicine profession for recognizing its significance.
Steven Shapiro, DO
Fenton, MI
Let’s apply the lessons from the AIDS crisis to the COVID-19 pandemic
In 2020, COVID-19 disrupted our medical system, and life in general. In the 1980s, the AIDS epidemic devastated communities and overwhelmed hospitals. There were lessons learned from the AIDS epidemic that can be applied to the current situation.
Patients with HIV-spectrum illness faced stigmatization and societal indifference, including rejection by family members, increased rates of suicide, fears of sexual and/or intrauterine transmission, substance abuse issues, and alterations of body image for those with wasting syndromes and disfiguring Kaposi lesions. AIDS prevention strategies such as the provision of condoms and needle exchange programs were controversial, and many caregivers exposed to contaminated fluids had to endure months of antiretroviral treatment.
Similar to the AIDS epidemic, the COVID-19 pandemic has had significant psychological implications for patients and caregivers. Patients with COVID-19 infections also face feelings of guilt over potentially exposing a family member to the virus; devastating socioeconomic issues; restrictive hospital visitation policies for family members; disease news oversaturation; and feelings of hopelessness. People with AIDS in the 1980s faced the possibility of dying alone, and there was initial skepticism about medications to treat HIV—just as some individuals are now uneasy about recently introduced coronavirus vaccines.
The similarities of both diseases allow us some foresight on how to deal with current COVID-19 issues. Looking back on the AIDS epidemic should teach us to prioritize attending to the mental health of sufferers and caregivers, creating advocacy and support groups for when a patient’s family is unavailable, instilling public confidence in treatment options, maintaining staff morale, addressing substance abuse (due to COVID-related stress), and depoliticizing prevention strategies. Addressing these issues is especially critical for minority populations.
As respected medical care leaders, we can provide and draw extra attention to the needs of patients’ family members and health care personnel during this COVID-19 pandemic. Hopefully, the distribution of vaccines will shorten some of our communal and professional distress.
Robert Frierson, MD
Steven Lippmann, MD
Louisville, KY
In 2020, COVID-19 disrupted our medical system, and life in general. In the 1980s, the AIDS epidemic devastated communities and overwhelmed hospitals. There were lessons learned from the AIDS epidemic that can be applied to the current situation.
Patients with HIV-spectrum illness faced stigmatization and societal indifference, including rejection by family members, increased rates of suicide, fears of sexual and/or intrauterine transmission, substance abuse issues, and alterations of body image for those with wasting syndromes and disfiguring Kaposi lesions. AIDS prevention strategies such as the provision of condoms and needle exchange programs were controversial, and many caregivers exposed to contaminated fluids had to endure months of antiretroviral treatment.
Similar to the AIDS epidemic, the COVID-19 pandemic has had significant psychological implications for patients and caregivers. Patients with COVID-19 infections also face feelings of guilt over potentially exposing a family member to the virus; devastating socioeconomic issues; restrictive hospital visitation policies for family members; disease news oversaturation; and feelings of hopelessness. People with AIDS in the 1980s faced the possibility of dying alone, and there was initial skepticism about medications to treat HIV—just as some individuals are now uneasy about recently introduced coronavirus vaccines.
The similarities of both diseases allow us some foresight on how to deal with current COVID-19 issues. Looking back on the AIDS epidemic should teach us to prioritize attending to the mental health of sufferers and caregivers, creating advocacy and support groups for when a patient’s family is unavailable, instilling public confidence in treatment options, maintaining staff morale, addressing substance abuse (due to COVID-related stress), and depoliticizing prevention strategies. Addressing these issues is especially critical for minority populations.
As respected medical care leaders, we can provide and draw extra attention to the needs of patients’ family members and health care personnel during this COVID-19 pandemic. Hopefully, the distribution of vaccines will shorten some of our communal and professional distress.
Robert Frierson, MD
Steven Lippmann, MD
Louisville, KY
In 2020, COVID-19 disrupted our medical system, and life in general. In the 1980s, the AIDS epidemic devastated communities and overwhelmed hospitals. There were lessons learned from the AIDS epidemic that can be applied to the current situation.
Patients with HIV-spectrum illness faced stigmatization and societal indifference, including rejection by family members, increased rates of suicide, fears of sexual and/or intrauterine transmission, substance abuse issues, and alterations of body image for those with wasting syndromes and disfiguring Kaposi lesions. AIDS prevention strategies such as the provision of condoms and needle exchange programs were controversial, and many caregivers exposed to contaminated fluids had to endure months of antiretroviral treatment.
Similar to the AIDS epidemic, the COVID-19 pandemic has had significant psychological implications for patients and caregivers. Patients with COVID-19 infections also face feelings of guilt over potentially exposing a family member to the virus; devastating socioeconomic issues; restrictive hospital visitation policies for family members; disease news oversaturation; and feelings of hopelessness. People with AIDS in the 1980s faced the possibility of dying alone, and there was initial skepticism about medications to treat HIV—just as some individuals are now uneasy about recently introduced coronavirus vaccines.
The similarities of both diseases allow us some foresight on how to deal with current COVID-19 issues. Looking back on the AIDS epidemic should teach us to prioritize attending to the mental health of sufferers and caregivers, creating advocacy and support groups for when a patient’s family is unavailable, instilling public confidence in treatment options, maintaining staff morale, addressing substance abuse (due to COVID-related stress), and depoliticizing prevention strategies. Addressing these issues is especially critical for minority populations.
As respected medical care leaders, we can provide and draw extra attention to the needs of patients’ family members and health care personnel during this COVID-19 pandemic. Hopefully, the distribution of vaccines will shorten some of our communal and professional distress.
Robert Frierson, MD
Steven Lippmann, MD
Louisville, KY