Background To improve patient safety, organisations must systematically measure avoidable harms. Clinical surveillance—consisting of prospective case finding and peer review—could improve identification of adverse events (AEs), preventable AEs and potential AEs. The authors sought to describe and compare findings of clinical surveillance on four clinical services in an academic hospital.
Methods Clinical surveillance was performed by a nurse observer who monitored patients for prespecified clinical events and collected standard information about each event. A multidisciplinary, peer-review committee rated causation for each event. Events were subsequently classified in terms of severity and type.
Results The authors monitored 1406 patients during their admission to four hospital services: Cardiac Surgery Intensive Care (n=226), Intensive Care (n=211), General Internal Medicine (n=453) and Obstetrics (n=516). The authors detected 245 AEs during 9300 patient days of observation (2.6 AEs per 100 patient days). 88 AEs (33%) were preventable. The proportion of patients experiencing at least one AE, preventable AE or potential AE was 13.7%, 6.1% and 5.3%, respectively. AE risk varied between services, ranging from 1.4% of Obstetrics to 11% of Internal Medicine and Intensive Care patients experiencing at least one preventable AE. The proportion of patients experiencing AEs resulting in permanent disability or death varied between services: ranging from 0.2% on Obstetrics to 4.9% on Cardiac Surgery Intensive Care. No services shared the most frequent AE type.
Conclusions Using clinical surveillance, the authors identified a high risk of AE and significant variation in AE risks and subtypes between services. These findings suggest that institutions will need to evaluate service-specific safety problems to set priorities and design improvement strategies.
- Medical errors
- adverse events
- prospective surveillance
- patient safety
- near miss
This is an open-access article distributed under the terms of the Creative Commons Attribution Non-commercial License, which permits use, distribution, and reproduction in any medium, provided the original work is properly cited, the use is non commercial and is otherwise in compliance with the license. See: http://creativecommons.org/licenses/by-nc/2.0/ and http://creativecommons.org/licenses/by-nc/2.0/legalcode.
Statistics from Altmetric.com
If you wish to reuse any or all of this article please use the link below which will take you to the Copyright Clearance Center’s RightsLink service. You will be able to get a quick price and instant permission to reuse the content in many different ways.
Improving patient safety requires the minimisation of treatment-related harm. This is typically measured as adverse events (AEs) (harms caused by medical care), preventable AEs (harm caused by errors) and potential AEs (errors with the potential for harm). Numerous studies have demonstrated a high incidence of AEs and preventable AEs in hospitalised patients.1–8 These studies have prompted significant investments to improve patient safety.
Detection of AEs and preventable AEs is a primary step to achieving a safe healthcare system. By systematically measuring these outcomes and analysing their causes, healthcare planners may design system changes to prevent them.9–14 Unfortunately, most hospitals and health systems have rudimentary approaches to identify AEs. These approaches include the use of voluntary reporting, chart reviews and scanning of administrative data.15 Each of these methods has important limitations including a failure to report, inconsistent peer review and poor specificity for patient safety problems.13 15–23
Detecting AEs is especially challenging because of the invisible nature of treatment-related harm. When a patient experiences a poor outcome, it is often assumed the outcome was caused by the underlying disease process. While this is often true, careful peer review will often determine that medical care was, at least, partially responsible. Furthermore, adverse outcomes and the circumstances leading to them are often not systematically documented. This will often lead to disagreement between peers when chart review is used to identify AEs.24 25
Clinical surveillance is a promising method of AE detection. In this method, patients and providers are directly and indirectly monitored by a trained observer prospectively.26–29 As soon as clinical events are identified, the observer records prespecified information to facilitate subsequent peer review. This avoids missing information that occurs from incomplete documentation in the patient record.
Clinical surveillance has several advantages over existing AE detection methods. It consists of active surveillance rather than relying solely on voluntary reporting; it ensures timely collection of information relevant to case classification rather than retrospectively relying on what is documented in charts, and it enhances peer review through timely review of cases when reviewers still remember the case. On the other hand, although it has been used successfully for years in tracking surgical complications and healthcare-associated infections, it has not been widely adopted for all AE types in multiple, diverse settings.
In this study, we describe our findings from implementing clinical surveillance in our healthcare facility. We describe a general process and the results obtained during its application in four diverse services at an academic hospital. The comparison of results between services was of particular interest, as it would validate our ability to learn useful information from the programme and would direct strategies for improvement.
This study took place at a 1000-bed multicampus urban-based academic hospital. The hospital provides tertiary-level patient care, including organ transplants, and is a Level 1 trauma centre. We performed our clinical surveillance activities as part of a research study, which was approved by the Ottawa Hospital Research Ethics Board.
We conducted surveillance sequentially on the following four services: General Internal Medicine, Obstetrics, Intensive Care, and Cardiac Surgery Intensive Care. For each service, we monitored care for 12 weeks using a common method, which was adapted for the service to ensure relevancy to the patients treated. The general method included establishment of surveillance parameters, case finding, peer review and event classification (figure 1).
Establishment of surveillance parameters
Before starting surveillance, we identified several service-specific surveillance parameters describing surveillance location, baseline patient information and triggers.
We prespecified the surveillance location to ensure consistent and representative case identification. On the Obstetrics service, for example, we monitored patients in the case-room, the obstetrical operating room, and the postpartum ward. For Medicine, however, we monitored patients on the ward only. We also prespecified whether we would include care preceding the observation period. For example, for Intensive Care patients we included events leading to unit admission.
We prespecified variables describing baseline patient information to enable consistent case reviews and permit future risk adjustments. For example, when we implemented on the Medicine service, we collected information regarding the presence of diagnoses associated with poor outcomes.30
Most importantly, we prespecified triggers which represented the factors we used to prompt case reviews. Trigger tool methodology has been used successfully in other institutions to address specific AE types such as adverse drug events,18 19 31–37 hospital-acquired infections38 and surgical complications,39–41 so we adapted the concept to create a more generalised model for the identification of all types of AEs in various settings. For each service, we reviewed the literature to identify previously published risk factors for adverse outcomes. A multidisciplinary team specific to each service then determined the face validity of these triggers and discussed additional triggers. Consensus was used to define the final set of triggers. The triggers represent clinical phenomenon, pharmacy orders and system events relevant to the patients treated. For example, ‘stat Caesarean section’ was a system event on the Obstetrics service, while ‘unplanned OR return’ was used on the Cardiac Surgery Intensive Care unit. For all services, we had an ‘Other’ category to account for events that were not predicted a priori. (See online appendix 1) contains the triggers we used for the four services.
Case finding was performed daily Monday to Friday between 08:00 and 16:00 by a trained nurse (observer). The observer training consisted of a patient safety presentation, relevant literature, practice case report generation and reviews, familiarisation with triggers and service-specific integration. We integrated the observer into the daily activities of the specific service to ensure maximal interaction with the providers and opportunities to observe care yet with minimal disruption to the providers' routine. This meant there were slight differences in how the observations were performed. For example, in the Intensive Care Unit, observations occurred at the bedside during morning team rounds, while on the Medicine service they occurred at the nursing station. The observer was encouraged to ask providers for details if they were not volunteered or documented in the chart.
All patients in the surveillance locations were eligible for case finding. The observer captured baseline information on all patients admitted to the service. Once data were entered, the observer monitored the patient until discharge. The observer identified triggers using a combination of three common AE detection methods: daily interactions with clinicians and administrators to enquire about or discuss trigger events (prompted and voluntary reporting), daily chart reviews and direct observation. If triggers occurred when our observer was not present, we identified them based on information volunteered by providers or from the chart.
Once a trigger was identified, we captured standard information describing the case. The observer generated a case summary using a standard format describing the patient and event in detail, the response to the event, the impact of the response and the documented cause of the response if there was one. This information was compiled by the observer through discussions with the relevant staff on the service.
Peer review took place weekly. For all reviews, we used the same definitions and classification systems as those used in the Harvard Medical Practice Study and other patient safety studies (See online appendix 2).1–7 We used a majority opinion to consolidate the ratings of the multiple reviewers.42 Cases were initially classified as outcomes (in which a patient experienced some form of harm, including pain, other distressing symptoms, such as dyspnoea, or reasonable markers for such symptoms such as increases in respiratory rate) or process problems (in which a task did not occur as planned). For example, if a patient experienced a sudden drop in blood pressure, the event was classified as an outcome. On the other hand if a patient missed a dose of an antibiotic but had no discernible change in status, then the event was classified as a process problem. For outcomes, we first assessed whether they were caused by medical care (ie, an AE). If so, reviewers determined whether it was preventable (ie, it was caused by error or was avoidable in some other way). We used six-point Likert scales with a cut-point of 3 for both of these assessments. For processes, reviewers determined whether the procedural or process problem could have caused harm (ie, a potential AE). Again, we used a six-point Likert scale with a cut-point of 3 for this assessment (ie, a score of 4, 5 or 6 indicated a process problem).
Event classification occurred following categorisation of cases as AEs, preventable AEs or potential AEs. For these cases, the observer defined the event's severity and type. For severity, events caused one of the following levels of increasing harm: none; laboratory abnormalities only; symptoms only; temporary disability; permanent disability; or death. Events were also classified according to one or more of the following types: hospital-acquired infection; adverse drug event; procedural complication; operative complication; diagnostic error; management error; anaesthetic complication; obstetrical injury; and system design flaw. As almost all errors could be argued to be system errors, only those events that could not be assigned to another type were deemed to be system design flaws.
We described patient baseline characteristics using median and IQR for continuous variables and frequency distributions for categorical variables. We described events in terms of preventability, severity and type for each service overall. We calculated the risk of experiencing at least one event per hospital encounter. We also calculated the event rate in terms of events per patient day of observation. We used SAS version 9.0 for all data management and analyses.
Table 1 describes the patient populations in terms of their baseline characteristics. There were important differences between services. We observed many more patients in Medicine (n=453) and Obstetrics (n=516) than Cardiac Surgery (n=226) or Intensive Care (n=211). Individual patients spent more days on Medicine than Intensive Care, Obstetrics and Cardiac Surgery, with median lengths of stay on the monitored unit of 8 days, 4.8 days, 2.4 days and 2.1 days, respectively. The indications for admission varied extensively between the four services. Medicine patients were older (median age on Medicine, Cardiac Surgery, Intensive Care and Obstetrics was 74, 70, 66 and 30 years, respectively) and frailer (proportion of patients with a Charlson Index of 3 on Medicine, Cardiac Surgery and Intensive Care was 24%, 3% and 7%, respectively).
Table 2 demonstrates the overall and service-specific AE risk and rate. 192 patients (13.7%) encountered 245 AEs (2.6 events per 100 patient days), 86 patients (6.1%) encountered 88 preventable AEs (0.9 events per 100 patient days), and 75 patients (5.3%) encountered 81 potential AEs (0.9 events per 100 patient days). The greatest number of events was observed on the Medicine service, which accounted for 63% (75/119) of patients with AEs, 56% of (48/86) patients with preventable AEs and 56% (40/75) of patients with potential AEs.
The risk and rate of all AEs and preventable AEs were highest for the Intensive Care unit. Of the 211 patients observed on Intensive Care, 52 (24.6%) experienced at least one AE, and 23 (10.9%) experienced a preventable AE. The corresponding event rates were 4.5 and 1.4 per 100 patient days for AEs and preventable AEs, respectively. The lowest risk and rate of events occurred on the Obstetrics service. Of the 516 patients observed on Obstetrics, 20 (3.9%) experienced at least one AE, and seven (1.4%) experienced a preventable AE. The corresponding event rates were 1.5 and 0.5 per 100 patient days for AEs and preventable AEs, respectively.
Table 3 describes the severity of all AEs detected. Of all 245 events detected, six (2.5%) resulted in/were associated with/potentially led to death, and 16 (6.6%) led to permanent disability. The vast majority of events resulted in symptoms only (n=75, 30.7%) or led to temporary disability (n=112, 45.9%). The majority of events that led to death or permanent disability occurred in the two critical care areas (Cardiac Surgery Intensive Care and Intensive Care). The risk of AEs causing permanent disability or death varied 24-fold across services: Cardiac Surgery Intensive Care (11/226=4.9%), Intensive Care (8/211=3.8%), Medicine (2/453=0.4%), Obstetrics (1/516=0.2%).
Table 3 also describes the types of AEs identified. There were important proportions of all AE types, but overall the most common types were therapeutic errors, procedural complications, and adverse drug events. Within each service, the top three AE types differed, and no service shared the most frequent AE type. For Cardiac Surgery, the top three types were surgical complications (47%), hospital-acquired infections (19%) and procedural complications (11%). For Intensive Care, the top three types were hospital-acquired infections (25%), procedural complications (24%) and therapeutic errors (19%). For Medicine, the top three types were adverse drug events (30%), therapeutic errors (24%) and procedural complications (12%). For Obstetrics, the top three types were therapeutic errors (36%), procedural complications (32%) and adverse drug events (27%).
We used clinical surveillance to detect AEs on four hospital services. AE risk per encounter was 13% with a range of 4% to 25% depending on service. Overall, 36% of AEs were preventable, though this was highly variable between services, ranging from 13% to 56%. General Internal Medicine had the greatest number of AEs but also had the highest number of patients and comorbidities. After controlling for length of stay and the number of patients observed, the risk of AEs, including those with the most severe consequences, was greatest in the two critical care areas. There were differences in the types of AEs identified; however, therapeutic errors were common in all subgroups studied. Finally, clinical surveillance identified an important number of potential AEs.
These findings are important because they identify ample opportunities for quality improvement. We found a greater AE risk than most previous studies of AEs which have tended to rely solely on chart review.1–7 We used a clinical observer to identify events within 24 h of their occurrence, and reviewed them within 7 days of their identification. The timely identification of problems would ultimately mean a more rapid response to rectify them. The reviews were performed by clinicians who knew the clinical setting and were well respected by their local peers. We believe that these factors resulted in classifications with a much greater degree of face validity than existing methods to identify AEs (such as incident reports and chart reviews). These factors also led us to identify a significantly greater proportion of events attributed to diagnostic and therapeutic errors, than prior studies. Furthermore, the systematic nature of our data collection ensured our methods were less biased than morbidity and mortality reviews or closed-claims analysis.15 43 44
Our study supports the concept that improvement strategies must be locally developed and must target specific problems. We found that AE risk and type varied by service. The underlying explanation for this finding likely relates to the frequency of underlying processes occurring within these areas that in turn lead to more opportunities to cause harm. Regardless of the cause, it follows that solutions will also necessarily vary. For example, in the Cardiac Surgery Intensive Care unit, interventions should be directed to preventing certain types of surgical complications. On the Medicine service, interventions that focus specifically on reducing errors related to the medication administration process would be most beneficial. The conclusion that errors and their solutions vary by patient service is intuitively obvious but is often not incorporated in the ‘top down’ approaches to improve quality and safety taken by accreditation organisations or other regulators.
Our study also supports the conclusion that priority setting, as it pertains to safety improvement, must occur in a highly strategic manner. The greatest number of AEs occurred on the service with the frailest patients. Effective interventions to improve safety in these patients would have the largest reduction on the institutional number of events. However, after controlling for the number of patients and length of observation, these patients had a risk of events similar to the least frail patients. The highest event risk and the most severe events occurred in services that had very invasive therapies and where patients were most acutely ill. Therefore, efforts to prevent injuries in these groups might be more efficient and associated with a greater overall impact in terms of reducing the economic burden of AEs. For all these reasons, organisations will have to be very clear about their goals when targeting the safety problems they hope to solve.
Finally, this research has highlighted the feasibility of clinical surveillance. We used a comprehensive strategy involving the use of direct observation, voluntary and prompted reporting, as well as daily chart reviews. This combination of AE detection methods has been recommended by others15 and serves to increase the overall performance of the monitoring system. Our programme was well accepted in the four services studied with very little financial support (the clinical observer was a funded position). However, all other functions required to maintain the process were voluntary, and there was very little infrastructure support required. There was no evidence that providers who were being observed were overtly or covertly averse to the programme or changed their behaviours. In fact, the methodology was so well received that our institution wanted to and is currently taking steps to implement it on other services.
Our findings are important in that we used a reproducible method of case detection and case review applying a standard method of classification to facilitate comparison. However, there are three important limitations which we would like to highlight. First, we are uncertain of the programme's reliability. Threats to reliability include the observer and peer-review processes. We attempted to mitigate these concerns through standardisation of the process of observation using triggers and case report forms; and, for the peer review, the use of multiple reviewers. These approaches have been successful in improving reliability in prior settings. However, we recommend further research before our method be adopted by health systems to compare institutions. Second, the generalisability of our surveillance feasibility is unknown, particularly because it was performed using a single observer. Even though we successfully implemented our programme within multiple settings in different facilities, it is possible that other settings may create challenges. Further work is recommended to establish whether the programme can be successfully adapted within non-teaching hospitals and in different health systems. Third, the observer cannot be everywhere at once and may miss some events as they are happening. This means the programme cannot be used to intervene systematically and directly in patient care when it does not meet the standard. However, this is not the stated purpose of the programme. Furthermore, if cases are detected in which an immediate intervention is warranted, then this appropriate response should be easily performed. For example, we identified a few instances when a critical laboratory abnormality had not been recognised and simply brought it to the attention of the treating team.
In summary, our study supports the conclusion that clinical surveillance appears to be an effective means of detecting patient safety issues. Before making a general recommendation for widely adopting this method, we recommend further evaluations to address four issues. First, we recommend comparing the findings of clinical surveillance with other methods of identifying patient safety events. Second, we recommend an economic analysis to determine the most efficient method of adverse event detection. Third, we would recommend evaluations of this method in different hospitals. Finally, and most importantly, we recommend studies to determine if adverse event detection leads to improvement in patient outcomes.
We would like to thank A Jennings for her assistance throughout the study and in the preparation of the manuscript. This research received financial support from the Canadian Patient Safety Institute, the Canadian Institute for Health Research, the Healthcare Insurance Reciprocal of Canada, the University of Ottawa Heart Institute, and the Ottawa Hospital Centre for Patient Safety.
Funding This research received financial support from the Canadian Patient Safety Institute, the Canadian Institute for Health Research, the Healthcare Insurance Reciprocal of Canada, the University of Ottawa Heart Institute and the Ottawa Hospital Centre for Patient Safety.
Competing interests None.
Ethics approval Ethics approval was provided by the Ottawa Hospital Research Ethics Board.
Provenance and peer review Not commissioned; externally peer reviewed.