Article Text
Abstract
Clinical guidelines often make recommendations on the use of diagnostic tests. Compared with sensitivity and specificity, the use of pre and posttest probabilities allows a more explicit and rational selection and interpretation of diagnostic tests. Ideally, clinical guidelines relating to diagnosis should routinely incorporate this information to enhance individualised decision making. We report our experience of incorporating pre and posttest probabilities into a guideline on the investigation of women with postmenopausal bleeding developed by the Scottish Intercollegiate Guidelines Network. Issues relating to their application are highlighted, including the limitations of available evidence on diagnostic tests and prevalence of disease, acceptability to guideline users, and the uncertain impact on actual clinical decision making. Despite these potential difficulties, the incorporation of data on pre and posttest probabilities into the development and presentation of guideline recommendations may offer an important opportunity to make clinical decision making more transparent for both clinicians and patients.
Statistics from Altmetric.com
Clinical guidelines are “systematically developed statements to assist practitioner and patient decisions about appropriate health care for specific clinical circumstances”.^{1} Guidelines often make recommendations on the use of diagnostic tests. Parameters essential to the evaluation of diagnostic test performance, such as sensitivity and specificity, are neither intuitive nor readily transferable to clinical situations.^{2} Much has been written highlighting the superiority of likelihood ratios in providing information more relevant to clinical decisions concerning individual patients,^{3–}^{7} yet these parameters are seldom presented in clinical guidelines. The Scottish Intercollegiate Guideline Network (SIGN) guidelines are developed by multidisciplinary groups according to rigorous criteria, including explicit methods of appraising evidence and grading recommendations.^{8} None of 17 SIGN guidelines published between 1996 and 2001 which dealt substantially with diagnostic processes mentioned the use of likelihood ratios in the development of recommendations.
We were members of a SIGN guideline development group that addressed the selection and interpretation of diagnostic procedures for the assessment of women with postmenopausal bleeding. During the course of the group’s deliberations we were struck by the discomfort expressed by other group members and the target audience (mainly encompassing gynaecologists, radiologists and general practitioners) when “newer” approaches to understanding diagnostic tests were introduced. This paper draws on our deliberations to demonstrate how the development of diagnostic recommendations within clinical guidelines can be improved and highlights potential difficulties with their application.
LIMITATIONS OF SENSITIVITY AND SPECIFICITY IN THE CLINICAL ENCOUNTER
Postmenopausal bleeding represents one of the most common reasons for referral to gynaecological services, largely because of the need to detect or exclude endometrial carcinoma. The most rigorously evaluated investigation is transvaginal ultrasonography^{9,}^{10} which measures endometrial thickness. The diagnostic rationale is that, for an individual woman, the greater the measured endometrial thickness, the higher the probability that cancer is present. Transvaginal ultrasonography is attractive as an initial investigation as it is noninvasive and well tolerated. Sufficient reassurance from a negative result can help avoid unnecessary and more invasive procedures seeking a tissue diagnosis. The Guideline Development Group had to decide whether or not to recommend the use of this as a first line test and, if so, what diagnostic threshold to recommend.
Although transvaginal ultrasonography obtains an actual measurement of endometrial thickness, test results are typically reported simply as “positive” or “negative” depending on whether the thickness is above or below a specified threshold. In such circumstances, performance of the diagnostic test across the study group is usually summarised in terms of sensitivity and specificity. Sensitivity is the probability of testing positive if the disease is truly present. Specificity is the probability of testing negative if the disease is truly absent.
Box 1 presents an illustrative example of a woman referred to a gynaecology outpatient department for assessment of postmenopausal bleeding. She and her general practitioner were concerned about the possibility of endometrial cancer. Transvaginal ultrasound was performed using a threshold of >5 mm to define abnormal endometrial thickening. Based upon findings from a recent metaanalysis, transvaginal ultrasonography using this threshold has a sensitivity of 91% and specificity of 58% for the detection of endometrial carcinoma.^{10} On hearing her test result was negative, the woman questioned her gynaecologist:
Box 1 Use of pretest and posttest probabilities to judge the usefulness of a diagnostic test for an individual patient
A woman aged 55 was referred to a gynaecology outpatient department for assessment of postmenopausal bleeding. She and her general practitioner were concerned about the possibility of endometrial cancer. She had no history of using hormone replacement therapy. Transvaginal ultrasound was performed using a threshold of 5 mm or less to define a negative result. The sensitivity of the test is 91% and the specificity is 58%.
Using pre and posttest probabilities, the pretest probability that this woman has endometrial cancer is estimated as 10%. Following a positive test, her probability of cancer is 19%, and following a negative test, her probability of cancer is 1.7%.
Woman: This means I can’t have cancer then?
Gynaecologist: It’s unlikely. The test picks up 91 out of every 100 cancers.
Woman: So I have a 9 in 100 chance of actually having cancer then.
Gynaecologist: Mmm . . . not exactly. It depends . . . Following a negative test, both the woman and her gynaecologist may wish to have a more individualised estimate of risk, an idea of how likely it is that she has cancer. Apart from highlighting some of the problems clinicians encounter with interpreting sensitivity and specificity,^{11} this also shows that, regardless of the test result, if only these parameters are known it is not always clear whether further investigation is justified.
ENHANCING THE INTERPRETATION OF DIAGNOSTIC TEST RESULTS
A more informed decision involves the use of Bayes’ theorem. The pretest probability of disease is combined with test performance to estimate the posttest probability of cancer. For each of the two possibilities—a positive or negative transvaginal ultrasonography result—a formula involving the sensitivity and specificity of the test can be used to calculate the corresponding likelihood ratio. The posttest probability of cancer is then derived from the pretest probability and the relevant likelihood ratio using either a formula or nomogram. In the examples that follow the actual likelihood ratios are not presented, largely because clinicians are likely to be more conversant with probabilities.^{12}
The pretest probability (also known as prior probability or prevalence) quantifies what proportion of patients has the condition of interest—in this case the proportion of women presenting with postmenopausal bleeding who have endometrial cancer. The positive posttest probability estimates the probability that a patient has disease given a positive test result. Conversely, the negative posttest probability estimates the probability that a patient has disease given a negative test result.
This is also illustrated in box 1. Assuming a pretest probability of 10% and using a test threshold of 5 mm, the woman’s posttest probability of cancer following a positive test is estimated at 19%. Her posttest probability of cancer given a negative test result is 1.7% (approximately 1 in 60). Therefore, if the woman asks about the possibility that she has cancer given a negative result, her gynaecologist can reply: “About 1 in 60”. It is uncertain whether she and her gynaecologist would be sufficiently reassured by a 1.7% probability to render further investigation unnecessary. However, using the posttest probability provides a quantifiable estimate of the probability of disease being present for the patient in question, rather than largely relying upon clinical intuition.
CLARIFYING THE RATIONALE FOR THE SELECTION OF DIAGNOSTIC TESTS BY CLINICIANS
Some clinicians prefer lower test thresholds to determine a negative result. Using a threshold of 3 mm is more sensitive (98%) but less specific (53%) than 5 mm. What would using the 3 mm threshold mean for the woman in the clinical scenario in box 1? Based upon a pretest probability of 10%, her posttest probability of cancer following a negative test result is 0.4% (instead of the 1.7% when a less sensitive threshold was used). Both the woman and her doctor may be satisfied that this probability of cancer (1 in 250) is sufficiently low that further investigation is unnecessary unless symptoms recur.
If ultrasonography can be requested using a specified thickness threshold, each threshold for defining a positive or negative result represents, in effect, a different diagnostic test. The use of posttest probabilities can inform individual decisions about which threshold value for transvaginal ultrasonography is most appropriate. Despite a negative test result using the 5 mm threshold, the woman in this case might prefer further investigation and hence initial transvaginal ultrasonography may not alter subsequent clinical management. The use of posttest probabilities indicates that a more sensitive 3 mm threshold test might be necessary to ensure that a negative result provided adequate reassurance. Calculating posttest probabilities can therefore prompt an explicit consideration of whether an investigation (at a specified threshold) is worth performing.
POTENTIAL PITFALLS AND POSSIBLE SOLUTIONS
The incorporation of pretest and posttest probabilities into clinical guidelines may allow more rational selection and interpretation of diagnostic tests. However, guideline developers and users need to be aware of potential difficulties associated with their application.
Quality of evidence
Unfortunately, the evidence base for a 3 mm threshold is sparse and less reliable than that for the 5 mm threshold.^{10} Poorer quality studies may overestimate the accuracy of diagnostic tests.^{13} The Guideline Development Group faced the dilemma of whether to recommend transvaginal ultrasonography at a 5 mm threshold which would probably be too insensitive as a diagnostic test in most women, or to recommend a 3 mm threshold based on evidence possibly overestimating its accuracy. It was decided to make approximate adjustments to the likelihood ratio to account for bias (described in Appendix). After consideration of the issues there was a consensus decision that the 3 mm threshold should be recommended. For this the best and worst cases for negative posttest probabilities ranged from 0.6% to 0.8%. While acknowledging that this was an extrapolation, it was considered to be a reasonable pragmatic decision.
Uncertainty around estimates of test performance
Confidence intervals (CIs) demonstrate the degree of statistical uncertainty around point estimates of diagnostic accuracy. The robustness of recommendations can be checked if confidence intervals are used to provide a “worst case scenario” for test performance and subsequent posttest probabilities.^{10}
In the case of the woman with a pretest probability of 10% tested using the 3 mm threshold, her posttest probability of cancer given a negative result is estimated as 0.8% (using the most cautious likelihood ratio negative of 0.07). However, if uncertainty is allowed for (95% confidence), the probability of cancer may be as high as 2.1%. Some women and clinicians may feel inadequately reassured even by a negative transvaginal ultrasonography result if aware of the level of prevailing uncertainty.
Box 2 Definitions and formulae
Pretest probability = prevalence (the proportion of the population with the disorder)
Pretest odds = prevalence/(1 – prevalence)
Likelihood ratio for a positive test = sensitivity/(1 – specificity)
Likelihood ratio for a negative test = (1 – sensitivity)/specificity
Posttest odds = pretest odds × likelihood ratio
Posttest probability = posttest odds/(posttest odds + 1)
Estimates of posttest probability should reflect statistical uncertainty around estimates of both pretest probability and test performance. Although guideline users may feel uncomfortable at having to confront the uncertainty around test results, this does prompt a more explicit consideration of the potential limitations of diagnostic approaches.
Estimating pretest probabilities
Information on pretest probability is central to the likelihood ratio approach to interpretation of diagnostic test results. However, there is a paucity of reliable prevalence data. The Guideline Development Group estimated the overall pretest probability of endometrial cancer in women referred with postmenopausal bleeding to be 10%.^{14–}^{16} Estimating pretest probabilities for more specific subgroups is problematic. The Group used extrapolated data to estimate a pretest probability of 1% of endometrial cancer for women on sequential hormone replacement therapy presenting with unscheduled bleeding.^{17,}^{18} Such estimates may be contentious, particularly if they have a major impact on estimates of posttest probability and subsequent criteria for test selection and interpretation. If there is substantial doubt as to the pretest probability that applies, then the posttest probability can be calculated for a range of pretest values, particularly for the plausible upper limit for pretest probability. If new data on relevant pretest probabilities become available after publication, the revised posttest probability can then be calculated more accurately. In both cases this depends on guidelines explaining the formula used (box 2). More generally, pretest probabilities can be raised or lowered according to levels of clinical suspicion, thus providing more pertinent posttest probabilities.^{2}
Shifting pretest probabilities
Pretest probability may also alter over time within either setting as one or more of background risk factors for disease, consultation patterns, or referral thresholds change. For example, pretest probabilities for patients consulting general practitioners may differ from those for patients referred to a hospital clinic. General practitioners may “filter” out lower risk patients from referrals to secondary care. Introducing direct access to hospital investigations may reduce referral thresholds so that patients with lower pretest probabilities of disease are investigated. Subsequently, the posttest probabilities change, as can test performance.^{19} There is therefore a need for up to date epidemiological studies on the prevalence of disease associated with indications for investigation in various settings such as primary and secondary care. In particular, more studies of diagnostic approaches are required in primary care as this is usually where the most critical decision—whether to refer for further investigation—is made.^{20}
Acceptability to guideline users
Although pretest and posttest probabilities offer a more transparent basis for clinical decision making, clinicians may be deterred from using them because of actual and perceived complexities in their application. When the draft guideline on postmenopausal bleeding was pretested on a range of clinicians, concern was expressed that using pretest and posttest probabilities might be too complex (SIGN National Meeting, Edinburgh, 12 May 2000). This view is difficult to reconcile with the fact that probabilities are being used commonly to support and individualise treatment decisions.^{21} In coronary heart disease prevention, the absolute benefits of treatment vary according to pretreatment risk.^{22} Risk assessment charts, now commonplace in coronary heart disease guidelines, enable clinicians to assess an individual patient’s absolute risk based on a number of identifiable risk factors.
An advantage of pretest and posttest probabilities is that they avoid the “one size fits all” approach, the basis for widespread scepticism of guidelines among clinicians.^{23} In developing its recommendations, the Guideline Development Group suggested that a less than 1% probability of having cancer given a negative result would be sufficiently reassuring to justify the avoidance of further more invasive investigations. Individual women and other clinicians may hold different views about what constitutes a “safe” probability. Ideally, the selection of diagnostic tests should be driven by the acceptability of corresponding posttest probabilities. These data should be presented to allow individualised decisions to be made. We recommend extending the use of pretest and posttest probabilities in diagnosis, for example, to criteria for referral and investigation of malignancies other than endometrial cancer.^{24}
Impact on planning services
Sensitivity and specificity are still useful in the overall planning of healthcare programmes. Lower transvaginal ultrasonography thresholds tend to be more sensitive and hence miss fewer cancers. Such thresholds tend to be less specific and will result in more false positives and hence more patients being unnecessarily investigated. Guideline developers therefore need to balance the needs of individual patients against population needs in formulating recommendations. Health economic techniques can help make the costs and benefits involved in such tradeoffs more explicit.
Impact on clinical decision making
There is a substantial body of research on how clinicians make decisions^{25} which we cannot address in full here. Nonspecific diagnostic guideline recommendations are more likely to result in inappropriate or potentially harmful decisions.^{26} Presenting clinicians with information on likelihood ratios can improve their interpretation of diagnostic tests compared with information on sensitivity and sensitivity.^{11} It is not known whether knowledge of pre and posttest probabilities improves clinical decision making further. Teaching clinicians to make better judgements about disease probability may not alter treatment decisions.^{27} However, it is unrealistic to expect that—by itself—enabling clinicians to estimate posttest probabilities of disease more accurately will lead to more rational decision making. Evidence on changing professional practice, for example, indicates that the simple distribution of clinical guidelines seldom changes clinical practice.^{28} Therefore, as with any clinical guideline, active strategies are required to support the implementation of recommendations within diagnostic guidelines.
IMPLICATIONS FOR GUIDELINE DEVELOPMENT PROGRAMMES
The SIGN Guideline Development Group agreed to use pre and posttest probabilities in the development and presentation of its recommendations. A summary of the relevant recommendations from the Quick Reference Guide is shown in fig 1. In England and Wales the National Institute for Clinical Excellence (NICE) has embarked on a guideline development programme and will need to consider how to develop and present recommendations about diagnostic tests or processes. For potential users of guidelines, available generic checklists do not offer specific advice on assessing guidelines concerned with diagnostic processes.^{29}
Key messages

The use of pretest and posttest probabilities allows a more explicit and rational selection and interpretation of diagnostic tests.

Clinical guidelines relating to diagnosis should routinely incorporate this information to enhance decision making.

More epidemiological research is needed on the probability of disease being present across a range of settings to better inform pretest probabilities.
We recommend that diagnostic clinical guidelines routinely present posttest probabilities, with some indication of the uncertainty around these estimates, since these are the statistics most relevant to patient management. We further recommend that the pretest probabilities and likelihood ratios from which the posttest probabilities have been derived are also presented, for four main reasons:

the joint impact on decision making of test performance and the patient’s preexisting risk of disease is made explicit;

upstream to guideline development, further research priorities are highlighted. This includes reporting of pre and posttest probabilities for different subgroups of patients in future diagnostic studies;

revised posttest probabilities can be calculated if pretest probabilities change or better performing tests become available; and

clinicians need to be exposed to pre and posttest probabilities through guidelines more frequently if they are to gain confidence in their application.
APPENDIX: ADJUSTMENTS MADE TO SENSITIVITY AND SPECIFICITY FOR 3 MM THRESHOLD
The systematic review of transvaginal ultrasonography identified two primary studies evaluating a 3 mm threshold, neither of which was judged to be of high quality.^{10} The review identified 21 primary studies evaluating a 5 mm threshold, four of which were judged to be of high quality.^{10} Pooled sensitivity was only slightly higher for the 21 studies (91% v 89%) and pooled specificity was lower (58% v 68%). The differences in sensitivity and specificity between all 21 studies and the four high quality studies of the 5 mm threshold were applied to data reported for the 3 mm studies.
The resulting estimates of sensitivity and specificity were used to calculate the positive and negative likelihood ratios. This allowed an exploration of the extent to which the lower quality data for the 3 mm threshold might be misleading. The adjustments were applied in three ways: exactly as observed at 5 mm (sensitivity decreased by 2%, specificity increased by 10%); half the change (–1% and +5%); and the worst combination of these (–2% and +5%). The adjusted likelihood ratio negatives ranged from 0.05 to 0.07 compared with 0.04 for the unadjusted likelihood ratio.
Acknowledgments
We are grateful to other members of the SIGN Guideline Development Group for contributing to the development of the views expressed in this paper, to Jayne Parry, David Weller and Phil Hannaford for comments on earlier drafts, and to Patrick Chien and Janesh Gupta for providing data.
REFERENCES
Request permissions
If you wish to reuse any or all of this article please use the link below which will take you to the Copyright Clearance Center’s RightsLink service. You will be able to get a quick price and instant permission to reuse the content in many different ways.