Statistics from Altmetric.com
As in most other areas of human experience, fashions in healthcare and in healthcare research come and go. They come about because a good idea is articulated, often as a sound bite that makes intrinsic sense, and they go because a few intrepid researchers are prepared to investigate the substance behind the sound bite and they come to realise that the logic is not supported by the reality. The original logic is supplanted.
Over the last 20–30 years, there have been some notable fashions in healthcare that have lit up the globe like a comet and then died. Just as there have been passing fashions in surgery (eg, routine tonsillectomies) and medications (eg, hormone replacement therapy), so also have there been passing fashions in quality improvement (QI) and QI research. Kieran Walshe in his bibliometric research tracked the frequency paths of different QI terms used between 1988 and 2007.1 ‘Patient safety’ was a relatively low frequency term (along with ‘clinical governance’ and ‘six sigma’) from 1988 to 1999, when suddenly it took off following the release of the Institute of Medicine's ‘To Err is Human’2 report, surpassing almost all the other QI terms, along with ‘six sigma’.
The rise in evidence-based medicine, arguably itself a fashion, is largely responsible for deposing fashionable good ideas. Evidence-based medicine, although not immune to criticism,3 is different from other fashions in that it has infiltrated every part of every health system—the idea, at least, if not always the practice—and it has not gone away. It is a concept that has captured health systems globally and become embedded as a value, an aspiration and an expectation of healthcare, even more now than it was when the idea first lit up the world more than 40 years ago. Epidemiological research that measures, compares, tests and exposes is the key scientific approach used to create most of the evidence underpinning evidence-based medicine. Epidemiology is so important to medicine now that we are challenged to imagine life before the evidence-based medicine comet.
There may be some justification for viewing the patient safety phenomenon as a paradigm changing one, like evidence-based medicine, but we wonder if the relatively recent focus on incident reporting systems in primary care patient safety research may be a passing (or even passed) fashion. We propose that it is time for patient safety incident reporting to move from being a research tool to being embedded in health systems.
Governments, organisations and institutions have invested huge financial, personnel and time resources in establishing incident reporting systems with a view to identify and remedy threats to patient safety. Processes have been redesigned to make healthcare safer and tested (eg, physician order entry4), and efforts have been made to understand why known safe practices (eg, handwashing5) are not always used. However, there is little evidence to show that, overall, health systems are safer places for patients to engage with now than they were 10 years ago.6
The early studies identifying particular patient harms, their likelihood of occurrence and degree of preventability were all based in hospitals.7–14 These studies have had an impact far in excess of what would seem possible from their size or design. They raised awareness of the harm that healthcare processes can generate and forced critical reviews of health services. They established an epidemiological foundation for threats to patient safety in hospital care, provided an essential foundation for focused patient safety research and prompted health services to investigate and learn from safety management processes in other industries. The aviation industry, in particular, had a history of incident reporting systems that seemed transferable to the health sector and such systems became widely adopted into hospital safety management practices.
Primary care escaped the attentions of patient safety promoters in the early 2000s. The step of establishing an epidemiological description of the patient safety landscape was skipped in primary care. For some, this meant that patient safety was not a primary care issue; for others, it meant that the research just had not been done yet. Taking the latter view, a cohort of primary care researchers, including the authors, set about trying to find out whether people involved in primary care provision should be included when patient safety was being talked about. Emulating the hospital records review studies was considered not only ‘cumbersome and costly’15 but also impractical. Primary care does not operate in time-limited events like a hospital admission. Primary care episodes may unfold over months or years16 and primary care patient records may be dispersed across many providers.
However, patient safety reporting systems appeared to be transferable from hospitals into primary care settings. Reporting systems, where providers (in primary care, often doctors) recount the details of the patient safety incidents they observe, have an obvious role to play in contributing to safer healthcare. They provide invaluable insights about the more serious and rare threats to patient safety and give opportunities for clinicians to reflect on care processes they might modify for the good of their patients. They also have disadvantages. While they were initially introduced with a widespread expectation that they could be used to count patient safety incidents and to measure change, in practice they have become more useful as a tool to raise awareness of safety issues and to engage front-line providers in safety management practices. They cannot provide robust epidemiological information that can be used to quantify patient safety risks and measure their health and economic impacts. They cannot monitor patient safety improvements.
Primary care patient safety reporting systems have now been tested in many countries. In primary care, patient safety research has become dominated by research based on reporting systems.17 We know now from the extensive repetitions of this research method that they do operate in primary care settings—no further evidence of that is needed. The research reported by O'Bierne et al is the latest example of primary care reporting systems research, this time from Canada.18 However, this research report provides two critical pieces of information that we suspected but did not previously know for sure: first, reporting systems are expensive and, second, people do not like using them.
The reported research was conducted under a finite research grant, so costs could be calculated—$Can 1000 per report. In our view, this is a lot. The actual reports contained little novel information that could not be drawn from existing research and there does not appear to be any reason why family practices in Alberta, Canada, should be broadly different in their experience of patient safety events from any other group of family practices in other parts of the world. However, we appreciate that it is a subjective view.
It is clear from this study that people do not like reporting patient safety incidents to reporting systems and that they therefore provide only a very biased snapshot of all the incidents that actually happen in primary care clinics. Even among practices volunteering to participate in this research and even with the supportive infrastructure that must be in place when research is done, 104 people in the study family practices made only 264 reports in 36 months (<1 report per person per year). In a direct observational study, Elder et al19 reported that in 351 family practice visits where data were collected after each consultation, errors and preventable adverse events were identified in 24% of encounters. Applying this rate to the Canadian practices suggests that there were, conservatively, around 50 000 patient safety events occurring in the Canadian study practices, of which only 0.5% were reported.
Direct observation is bound to be a more expensive research method than independent report making and therefore impractical to institute on a long-term basis, but we think that primary care patient safety reporting systems have a limited future as a method of research intending to produce epidemiological data. Although they may be a relatively cheap research method to apply, the difficulty in engaging reporters means that they perform poorly for the purpose they are mainly used for—describing primary care patient safety incidents. Moreover, the descriptions that have been reported from them are consistent enough to tell us that the same sorts of problems are encountered in family medicine and general practice clinics wherever reporting systems have been instituted. If they are to be used in future, it should be to address questions relating to their role in behaviour change and making healthcare safer for patients. O'Bierne et al18 suggest that this is where they are heading with their research.
There is now enough evidence that reporting systems, at least primary care reporting systems, are costly and fail to provide the sorts of epidemiological results that really matter. If patient safety incident reporting systems are to become embedded in health systems, following the model of evidence-based medicine, we need to show if and how they make healthcare safer. To date, there has been little research on the role of incident reporting systems as catalysts for culture changes that lead to safer healthcare. This research area needs to be explored further and may be where reporting systems are shown to be most useful. The epidemiology of patient safety incidents in primary care remains a research gap. To explore this, we need to develop other better research methods.
Competing interests None.
Provenance and peer review Commissioned; internally peer reviewed.
If you wish to reuse any or all of this article please use the link below which will take you to the Copyright Clearance Center’s RightsLink service. You will be able to get a quick price and instant permission to reuse the content in many different ways.