View as “PDF version”
When we did our first CPOE implementation back in 2007, we were flooded with suggestions for potential alerts that could be used for patient safety. We had many ideas for alerts ourselves. But we readily recognized the need to limit such alerts in order to avoid alert fatigue.
Perhaps the prime opportunity to use alerts to drive clinician behavior is in medication safety. Computers can work rapidly in the background to identify issues like allergies, drug-drug interactions, drug-disease contraindications, effects of renal or hepatic dysfunction, and other considerations that should be taken into account when prescribing medications. Most clinicians cannot take the time to consider all those factors when prescribing. So, computer-generated alerts can provide important information to the clinician at the time of order entry.
There are 2 types of alerts: “soft” alerts and “hard” alerts. Soft alerts are suggestions that the clinician can choose to implement or ignore. Hard alerts are ones that require the clinician to do something, such as accept the recommendation or explain why he/she is overriding the alert (or, in the extreme, would prevent the clinician’s action all together). Interruptive alerts are ones that require interruption of a clinician’s workflow to answer questions or input additional information. Hard alerts are examples of interruptive alerts. There are probably also some “soft” alerts that are interruptive (eg. the alert might make the clinician read a long message).
A recent review (Cerqueira 2021) assessed the effectiveness of interruptive medication-prescriber alerts in changing prescriber behavior and improving patient outcomes in ambulatory care settings via computerized provider order entry (CPOE) systems. The authors found that clinician behavior was influenced in the majority of studies, with most noting a positive change. They found that alerts decreased pharmaceutical costs, moved medications toward preferred medications tiers and steered treatments toward evidence-based choices. Importantly, they also decreased prescribing errors.
But they also found that clinician feedback was rarely solicited and, when it was, showed frustration with alerts creating a time delay. Notably, only one of the nine studies in their review reported feedback that was overall ‘positive’. Clinicians often commented that the alerts were inappropriate and intrusive.
Shi et al. (Shi 2021) recently studied barriers to using clinical decision support in ambulatory care. One of the 7 primary barriers they identified was the use of false and disruptive alarms. Another barrier was the requirement to redesign workflow.
Bombarding clinicians with alerts that are inappropriate and interruptive leads to the phenomenon of “alert fatigue” in which clinicians begin to ignore and override most alerts, even those that have the potential to avert unwanted outcomes.
There have been many studies linking clinician burnout to use of electronic medical records (EMR’s). While most relate to the overall time spent on the EMR (see our May 2021 What's New in the Patient Safety World column “More on Time Spent on the EMR”), frustration with clinical decision support (CDS) systems has been identified as a key factor in leading to clinician burnout. Jankovic and Chen (Jankovic 2020) reviewed articles dealing with aspects of CDS that contribute to burnout and identify key themes for improving the acceptability of CDS to clinicians, with the goal of decreasing said burnout.
In our June 2020 What's New in the Patient Safety World column “EMR and Medication Safety: Better But Not Yet There” we discussed results from the Leapfrog CPOE EHR evaluation tool shows some improvement over time but highlights the persistence of vulnerabilities and the wide variability of hospital CPOE EHR systems to identify medication errors and prevent adverse drug events (ADE’s). Classen et al. (Classen 2020) looked at results from over 2300 hospitals. The overall mean total score increased from 53.9% in 2009 to 65.6% in 2018. The mean hospital score for the “basic” CDS category increased from 69.8% in 2009 to 85.6% in 2018. The mean hospital score for the “advanced” CDS category increased from 29.6% in 2009 to 46.1% in 2018. Hospital EHR’s did better on some categories than others. For example, they did best on the drug-allergy category in each year, increasing from 92.9% in 2009 to 98.4% in 2018. The lowest performing category throughout the study was drug-diagnosis contraindications, where the mean score was only 20.4% in 2009 and 33.2% in 2018. The authors conclude “these systems have only modestly increased their safety performance during a 10-year period, leaving critical deficiencies in these systems to detect and prevent critical safety issues.”
So, why haven’t we been able to use CDS more effectively to improve medication safety and efficacy? It largely boils down to issues related to design and implementation of clinical alerts and reminders.
When the phenomenon of “alert fatigue” became apparent to us, we recognized the need to limit our alerts to those situations we considered most important for patient care. We did the following:
Those actions were very much in line with the recommendations the whitepaper “Safe Practices to Reduce CPOE Alert Fatigue through Monitoring, Analysis, and Optimization” from ECRI's Partnership for Health IT Patient Safety (see our March 2021 What's New in the Patient Safety World column “ECRI Partnership Whitepaper on Alert Fatigue”). The ECRI whitepaper (ECRI 2021) actually goes into more detail about the metrics you should be following.
Regarding metrics, one unique metric not mentioned in the ECRI whitepaper but potentially very valuable was described by Einbinder and colleagues (Einbinder 2014): the “number needed to remind” (NNR). Analogous to the number needed to treat (NNT), this is the number of patients reached by a reminder to result in one recommended action being taken. They compared this to the “reminder performance” (RP), which is simply the measure of how often a recommended action is subsequently taken when the reminder is displayed.
Alagiakrishnan et al. (Alagiakrishnan 2019) used the NNR in primary care and geriatric clinics to assess clinical decision support for potentially inappropriate medications (PIM’s) from Beers criteria. Their CDS system used alerts known as Best Practice Advisories (BPA’s) to direct providers to a navigator where orders management, clinical information and educational materials were available. The BPA’s were used to advise clinicians of PIM’s among their patients' medication lists or new orders. The reminder performance (RP) across both clinics was 17.3%, which corresponds to an NNR of 5.8. The reminder performance was 37.1% in geriatric clinics vs. 13.4% in primary care clinics. The NNR in the primary care clinic was 7.4 and NNR in the geriatric clinic was 2.7.
They also developed a metric “Number Needed to Deprescribe” (NND) or the number of alert presentations specific to a medication and patient presented to a physician user before there was a deprescribing event. The reminder performance for deprescribing events was lower at 1.2%. The number needed to deprescribe (NND) was 82, with values for the primary care clinic of 80 and the geriatrics clinic of 96. There was considerable variation in all 3 parameters by the class of medication for which the BPA alerts fired.
The ECRI Partnership whitepaper reiterates the “5 Rights” model of clinical decision support (CDS) adopted from Osheroff et al. (Osheroff 2012):
One of the metrics in the ECRI whitepaper looks at whom the alert is targeted to. That raises a point we’ve often made in the past: you need to target the alert to the individual(s) most likely to get the desired action accomplished. That may not always be the clinician ordering something on CPOE or an ePrescribing system. For example, when an antidiarrheal medication is ordered on a patient receiving antibiotics, an alert to consider C. diff infection might be better targeted to an infection control worker than to the ordering clinician. Or alerts regarding some medication issues might be better targeted to clinical pharmacists.
The ECRI whitepaper also asks one very important question: “Is an alert the appropriate tool?”. Essentially, that is askng whether there is an alternative to accomplish the same goal. We’d like to emphasize that last point. In our March 3, 2009 Patient Safety Tip of the Week “Overriding Alerts…Like Surfin’ the Web” we noted that use of standardized order sets may avoid the need to generate some alerts (though standardized order sets can create some problems of the own, particularly when they contain outdated information that is no longer appropriate).
There is another phenomenon we’ve seen over and over. That is the disparity in effectiveness between “prospective” alerts and “look-back" alerts. When we’ve tried to use alerts to avoid using a “potentially inappropriate” medication (an example might be tricyclic antidepressants in an elderly patient), we found that clinicians almost never stopped that medication once they had already previously prescribed it. On the other hand, the alert was effective at preventing new prescriptions for that medication in the elderly. Awdishu et al. (Awdishu 2016), looking at the impact of alerts on prescribing in patients with renal disease, also found that prospective alerts had a greater impact than look-back alerts (55.6% vs 10.3%).
Marcilly et al. developed evidence-based usability design principles for medication alerting systems (Marcilly 2018). They note that alerts should:
The Marcilly paper notes that alert design should take into account parameters such as the
patient’s clinical context or the clinician’s specialty. That makes sense. An alert to a primary care physician reminding them to adjust a medication dose based upon the patient’s renal function might just create unnecessary “noise” for a nephrologist entering a similar order. Similarly, regarding context, some alerts about renal dosing may not be appropriate for a patient already on dialysis.
Alerts are also more likely to be complied with if they offer “actionable” tools. For example, an alert that offers an alternative action or choice is much more likely to be complied with than one that simply suggests the original action is discouraged.
Shah et al. (Shah 2021) looked at CDS alerts regarding renal dosing in hospitalized patients. They found that alerts were nearly always presented inappropriately and were all overridden during the 1-year period studied. This was distinctly different from data they had previously seen in a legacy system in which medication-related CDS alerts associated with renal insufficiency had been found to be the most clinically beneficial. They identified several potential reasons why the current medication-related CDS alerts associated with renal insufficiency were less effective than they had been in the legacy homegrown system:
The Shah study tells us we cannot simply adopt all vendor-generated CDS alerts and reminders. Rather, we must apply the same rigorous evaluation and monitoring to those alerts and reminders that we used when we designed all our own alerts and reminders. We need to make sure they are evidence-based, are actionable, offer alternatives where appropriate, are prospective, fit with workflow, take clinical context and clinician specialty into consideration, and, most of all, are limited in volume to those that are most likely to impact patient safety and patient outcomes so we can avoid alert fatigue.
We can’t duplicate the extensive literature review done by the ECRI Partnership. We encourage you to go back to their whitepaper (ECRI 2021). Their “evidence tables” in the appendix includes a summary of the findings from 12 key studies. We hope you’ll use lessons learned in those plus the key elements in today’s column to make sure you are getting the most out of your clinical decision support tools without generating alert fatigue and contributing to clinician burnout.
See some of our other Patient Safety Tip of the Week columns dealing with unintended consequences of technology and other healthcare IT issues:
See some of our previous columns dealing with the Leapfrog CPOE EHR evaluaton tool:
Cerqueira O, Gill M, Swar B, et al. The effectiveness of interruptive prescribing alerts in ambulatory CPOE to change prescriber behaviour & improve safety. BMJ Quality & Safety 2021; Published Online First: 19 April 2021
Shi Y, Amill-Rosario A, Rudin RS, et al. Barriers to using clinical decision support in ambulatory care: Do clinics in health systems fare better? Journal of the American Medical Informatics Association 2021; ocab064 Published: 25 April 2021
Jankovic I., Chen JH. Clinical Decision Support and Implications for the Clinician Burnout Crisis. Yearbook of medical informatics 2020; 29(1): 145-154
Classen DC, Holmgren AJ, Co Z, et al. National Trends in the Safety Performance of Electronic Health Record Systems From 2009 to 2018. JAMA Netw Open 2020; 3(5): e205547
ECRI Partnership for Health IT Patient Safety. Safe Practices to Reduce CPOE Alert Fatigue through Monitoring, Analysis, and Optimization. ECRI 2021
Einbinder J, Hebel E, Wright A, Panzenhagen M, Middleton B. The number needed to remind: a measure for assessing CDS effectiveness. AMIA Annu Symp Proc 2014; 14: 506-515
Alagiakrishnan K, Ballermann M, Rolfson D, et al. Utilization of computerized clinical decision support for potentially inappropriate medications. Clin Interv Aging 2019; 14: 753–762
Osheroff JA, Teich JM, Levick D, et al. Improving outcomes with clinical decision support: an implementer’s guide. 2nd ed. Chicago (IL): Healthcare Information and Management Systems Society; 2012
Awdishu L, Coates CR, Lyddane A, et al. The impact of real-time alerting on appropriate prescribing in kidney disease: a cluster randomized controlled trial. J Am Med Inform Assoc 2016; 23(3): 609-616
Marcilly R, Ammenwerth E, Roehrer E, et al. Evidence-Based usability design principles for medication alerting systems. BMC Med Inform Decis Mak 2018; 18: 69
Shah SN, Amato MG, Garlo KG, et al. Renal medication-related clinical decision support (CDS) alerts and overrides in the inpatient setting following implementation of a commercial electronic health record: implications for designing more effective alerts. Journal of the American Medical Informatics Association 2021; ocaa222
Print “PDF version”