Skip to main content
  • Research article
  • Open access
  • Published:

Do coder characteristics influence validity of ICD-10 hospital discharge data?

Abstract

Background

Administrative data are widely used to study health systems and make important health policy decisions. Yet little is known about the influence of coder characteristics on administrative data validity in these studies. Our goal was to describe the relationship between several measures of validity in coded hospital discharge data and 1) coders' volume of coding (≥13,000 vs. <13,000 records), 2) coders' employment status (full- vs. part-time), and 3) hospital type.

Methods

This descriptive study examined 6 indicators of face validity in ICD-10 coded discharge records from 4 hospitals in Calgary, Canada between April 2002 and March 2007. Specifically, mean number of coded diagnoses, procedures, complications, Z-codes, and codes ending in 8 or 9 were compared by coding volume and employment status, as well as hospital type. The mean number of diagnoses was also compared across coder characteristics for 6 major conditions of varying complexity. Next, kappa statistics were computed to assess agreement between discharge data and linked chart data reabstracted by nursing chart reviewers. Kappas were compared across coder characteristics.

Results

422,618 discharge records were coded by 59 coders during the study period. The mean number of diagnoses per record decreased from 5.2 in 2002/2003 to 3.9 in 2006/2007, while the number of records coded annually increased from 69,613 to 102,842. Coders at the tertiary hospital coded the most diagnoses (5.0 compared with 3.9 and 3.8 at other sites). There was no variation by coder or site characteristics for any other face validity indicator. The mean number of diagnoses increased from 1.5 to 7.9 with increasing complexity of the major diagnosis, but did not vary with coder characteristics. Agreement (kappa) between coded data and chart review did not show any consistent pattern with respect to coder characteristics.

Conclusions

This large study suggests that coder characteristics do not influence the validity of hospital discharge data. Other jurisdictions might benefit from implementing similar employment programs to ours, e.g.: a requirement for a 2-year college training program, a single management structure across sites, and rotation of coders between sites. Limitations include few coder characteristics available for study due to privacy concerns.

Peer Review reports

Background

Administrative data including hospital discharge abstract data have been extensively used to examine health and health systems and have provided insights into health care practices. However, the validity of such administrative data for use in research has been questioned [1]. As a result substantial efforts are being made to thoroughly assess these data for quality [27]. Many investigators [35, 810] have conducted validation studies of International Classification of Disease 9th Revision Clinical Modification (ICD-9-CM) coding focusing on individual data items, certain clinical conditions and complications of substandard care. These studies have found that administrative data are accurately coded for severe or life-threatening conditions such as myocardial infarction and cancer, but that some non-specific conditions like rheumatologic disease, are less accurately coded. The introduction of the International Classification of Disease 10th Revision (ICD-10) coding system, gave rise to additional validation studies which concluded that although the ranking in leading causes of death in the US [11] and trends in cause-specific mortality in Europe [12] changed between the two systems, coding validity by and large remained the same [13].

The quality of coded data is influenced by two major factors: first how clearly, precisely and completely health care workers (primarily physicians) document diagnoses and treatments in the patient chart and second, how accurately and consistently the charts are coded by health records coders. The mechanisms by which both of these processes affect the quality of administrative data remain unclear. However, the former (physician charting) is difficult to assess, being highly variable and dependent on many factors not elucidated in administrative data, such as communication between physician and patient and the experience and expertise of the physician [14]. The variability and validity of coding by health records technicians was the focus of this study. Specifically, variability and validity of coding of hospital discharge data in relation to coders' demographics and experience. Studies of coding validity have mostly attempted to assess the quality of coded medical information by relying on a code-recode or reabstraction methodology that determines the difference in coding outcomes between career coders and a panel of "expert" coders. These studies have invariably shown differences in coding between the career coders and the expert coders [15]. According to Iezzoni [1] and O'Malley [14], the process of assigning ICD codes is complicated, more of an art than a science. The many steps in the process of coding a diagnosis introduce numerous opportunities for error [14, 16].

There may be many reasons, typically not elucidated in reabstraction studies, why interpretation of data can vary between coders. For instance Green and Benjamin [17], in their 1986 study of hospitals in Illinois, USA, found improved coding quality in hospitals that employed credentialed coders, provided feedback to enhance quality and offered continuing education to coding staff. More recently, Callen et al. [18] in comparing error rates of procedure coding between trained coders and operating theatre staff in an Australian hospital demonstrated high error rates among operating theatre staff, thereby highlighting the need for formal coding training. Additionally, Lorenzoni et al. [19], in a study of an Italian hospital, showed that a program of ongoing training, continuous monitoring and feedback proved to be successful in improving the quality of material abstracted from the medical chart. Furthermore, Santos et al. [20] elicited coding managers' perceptions and reported that coding quality could be improved by higher staffing levels, continuing education for coders and increased interaction with medical staff.

Previous studies [1720] examining associations of coder characteristics and coding validity have generally assessed error rates only for individual data items and in addition did not specifically examine groups of similarly trained and credentialed coders. Our study was conducted to address the quality of coded administrative data by coders' characteristics related to employment and experience, and unfolded in two parts. First, we examined validity of coding in all coded health records across a large number of diagnoses and co-morbidities over 5 years to establish face validity of the coded data in terms of coder characteristics. Second, we investigated the validity of coding in relation to coder characteristics by assessing agreement between the coded discharge data and data obtained by chart review and reabstraction in a subset of our data. The goal was to provide evidence to inform quality improvement initiatives in coders' training.

Methods

Study Design

This was a descriptive study of ICD-10-Canadian Version (CA) coding validity in Calgary, Alberta, Canada, a city of approximately 1.2 million people. The study was approved by the Conjoint Health Research Ethics Board at the University of Calgary.

Part 1: Assessment of face validity of coding

Data Sources and Linkage

The health records coders described in this study are professionally trained and worked at four hospitals in Calgary. One coordinator supervises and manages the coding practice at the four sites to establish a consistent approach to coding. Each discharge record contained a unique identification number for each admission and coder and up to 16 coded diagnoses and 10 coded procedures. All the data in the discharge record was coded and no data was available in free text format. Using the coder identification number, hospital discharge records of patients admitted to Calgary hospitals between April 1st 2002 and March 31st 2007 were linked to coders' employment data (100% linkage), see Figure 1. Before linkage, coders working at Calgary facilities other than the four major hospitals were excluded (32 coders), as were coders that coded less than 10 records in total (7 coders). The final dataset for Part 1 of the study consisted of 422,618 discharge records coded by 59 coders. The coders excluded from this analysis were less often part-time employees (approximately 27% compared to 44%) and a higher proportion were high volume coders (27% compared to 20%).

Figure 1
figure 1

Flow chart of record linkage.

Indicators of Face Validity

Face validity of coding was assessed by looking at 6 indicators in each record: (1) number of diagnoses, (2) number of procedures, (3) number of complications, (4) number of Z codes (which pertain to factors influencing health status and contact with health services) and (5 and 6) number of unspecified diagnoses (those codes ending in .8 or .9). Theoretically, a greater number of diagnoses, procedures, complications and Z codes noted in the discharge record provide a more detailed picture of the patient. Conversely, a greater number of unspecified codes in the record could indicate lack of detail and potentially lack of validity.

In addition, the mean number of diagnoses coded per admission was determined for different levels of complexity of the major condition. We expected that the number of co-morbid diagnoses would increase with the complexity of the major condition. For this analysis we identified 6 common conditions of varying complexity which were coded as the major diagnosis.

The coder characteristics that were assessed included volume of coding (high volume vs. low volume), employment status (full-time vs. part-time), and hospital site (tertiary hospital vs. others). We defined high volume coders as those who were at the 80th percentile or higher of coding volume. This corresponded to ≥13, 000 records coded during the study period. Low volume coders made up the balance. We also described the number of records coded by year and because temporal trends in coding have been previously observed [19]. We also hypothesized that there may be seasonal variations in coding due to time pressures on coders at certain times of the year; therefore we described the number of records coded by season.

Part 2: Assessment of agreement between coded record and chart review

Data Sources and Linkage

To assess agreement, a subset of data was assembled from the dataset in Part 1. Two nurses, who had previously worked as coding professionals, reviewed the charts and reabstracted data on 32 predefined conditions from a random sample of 2002 inpatient charts, as previously described [13]. These 32 conditions make up two commonly used risk adjustment taxonomies, the Charlson and the Elixhauser methods [2124]. These reabstracted data were linked to the Part 1 dataset using the unique identification number for each admission. Of 2002 reabstracted charts 710 records were excluded because they were coded by coders that were excluded in Part 1 of the study. The final dataset for Part 2 of the study consisted of 1,292 discharge records coded by 18 coders, see Figure 1.

Next, the same 32 predefined conditions that were reabstracted by the nurse reviewers, were defined as being present or absent in the coded hospital discharge abstract record using ICD-10 coding algorithms [25].

Statistical Analysis

For assessment of face validity (Part 1), descriptive statistics were employed to report the mean number of diagnoses, procedures, complications, unspecified diagnoses (ie. diagnoses ending in 8 or 9) and Z codes. The kappa statistic was used to assess agreement between the coded hospital discharge record and chart review data [26]. Kappa values were compared for high vs. low volume coders, full-time vs. part-time coders and by hospital site, i.e. tertiary care centre vs. all other sites.

Results

Part 1: Assessment of face validity

A total of 422,618 discharge records were coded by 59 coders during the study period. On average each coder coded 7,163 records during the study period, see Table 1. The number of records coded increased steadily over time, but did not vary substantially by season, indicating a constant workload throughout the year with no obvious reduction in coding in the summer months. Coders were worked predominantly at site B, the tertiary care site (37%) and mostly worked full-time (56%). Low volume coders coded up to 12,587 records while high volume coders coded 13,039 to 24,631 records.

Table 1 Number of records coded and characteristics of coders in a Canadian Health Region from fiscal year 2002 to 2006

The mean number of diagnoses per record coded decreased over time from 5.2 in 2002/03 to 3.9 in 2006/07, see Table 2. Conversely, the number of records coded per year increased over the same period as seen in Table 1. When the data were stratified by season, no variation in coding across the 6 indicators of face validity was seen. Coders at the tertiary care site recorded the highest mean number of diagnoses, 5.0. No substantial variation was demonstrated over the other 5 indicators.

Table 2 Mean number of coded diagnoses, procedures, complications, unspecified and Z codes by year, season, site and coder characteristics

In relation to complexity of the major diagnosis, the mean number of diagnoses increased from 1.5 for a singleton born in hospital (simplest condition examined), to a maximum of 7.9 for diabetes with complications (most complex condition examined), see Table 3. The number of diagnoses coded did not vary substantially by the coder's employment status or coding volume. Coders working at the tertiary hospital coded substantially more diagnoses for both diabetes with and without complications, than did coders working at other sites.

Table 3 Mean number of coded diagnoses by complexity of main diagnosis, and by coding volume and employment status of coders and hospital level

Part 2: Assessment of agreement between coded record and chart review

The kappa values ranged from a low of 0 for blood loss anemia in non-tertiary hospitals to a high of 1 for HIV in low volume coders and part time coders, and in tertiary hospitals, see additional file 1 (Additional Table S4). Of note, these extreme kappas occurred in low prevalence conditions. Since kappa is a function of prevalence, the prevalence of each condition as determined for both the hospital discharge data and the re-abstracted chart data is also presented in Additional file 1, Additional Table S4 [27]. For most conditions kappa was similar across employment status, coding volume and site. Differences in kappa that existed across coder characteristics did not show a consistent direction in favor of any particular coder characteristics, see Figure 2.

Figure 2
figure 2

a-c: Difference in agreement (Kappa) of coders' data with abstraction data by coder and hospital characteristics for Elixhauser [22]and Charlson [21]comorbidities, listed hereafter: 1) Myocardial infarction, 2) Cerebrovascular disease, 3) Rheumatic disease, 4) Dementia, 5) Cardiac arrhythmias, 6) Pulmonary circulation disorders, 7) Valvular disease, 8) Hypertension, 9) Hypothyroidism, 10) Lymphoma, 11) Solid tumour without metastasis, 12) Renal failure, 13) Blood loss anemia, 14) Deficiency anemia, 15) Coagulopathy, 16) Fluid and electrolyte disorders, 17) Weight loss, 18) Obesity, 19) Alcohol abuse, 20) Drug abuse, 21) Psychoses, 22) Depression, 23) Congestive heart failure, 24) Peripheral vascular disease, 25) Paralysis, 26) Chronic pulmonary disease, 27) Diabetes with complications, 28) Diabetes uncomplicated, 29) Peptic ulcer disease, 30) Metastatic cancer, 31) Liver disease, 32) HIV/AIDS.

Discussion

This study examined the relationship between the quality of coded administrative data health record coders' characteristics, specifically employment status and experience. First, we assessed the face validity of coded health records by coder characteristics across a large number of diagnoses over 5 years. Second, we investigated the agreement between the coded hospital discharge record and reabstracted chart review data in relation to coder characteristics in a subset of the data. Overall, our results suggest that these coder characteristics do not influence the validity of hospital discharge data. Notably, coding site (tertiary vs. other hospitals) is related to the number of diagnoses coded per record but not to the agreement (kappa) between coded record and chart review data.

This study assessed validity of coded data by looking at 6 indicators of face validity. We theorized that a greater number of diagnoses, procedures, complications and Z codes in the discharge record would provide a more detailed picture of the patient, while a greater number of unspecified codes in the chart would indicate lack of detail. Our results demonstrate that a greater number of diagnoses were coded by coders working at a tertiary care centre, while the number of diagnoses coded was remarkably similar across employment status (full-time vs. part-time) and volume (high volume vs. low volume).

Coding variation by site has also been described by Lorence et al., [28] who recorded lower oncology coding error rates in rural areas compared to metropolitan areas in the USA. Besides rural location other site related organizational factors are thought to influence coding validity. In a recent mixed methods study by Santos et al., [20] structural features related to the physical setting of the workplace and the coders' proximity to coding managers and clinicians were identified as influencing coding quality in the qualitative analysis. However, the quantitative portion of the Santos study did not reveal any significant relationship between physical location of coders and quality of coding. Interestingly Rangachari [29] showed that large teaching hospitals performed poorly compared to small, rural, non-teaching hospitals when it came to coding accuracy. While we did not specifically record coding errors in this study, a greater number of diagnoses were coded at the tertiary care centre than at other sites in our region. This could be explained by variation in structural features of the Health Records Department at the 4 sites as described above or in other organizational factors that we did not interrogate in this study, such as level of staffing and availability of resources. Another potential explanation for our result may be differences in the severity of illness of patients admitted to the various hospitals in our region. The tertiary centre cares for more acutely ill patients and is a referral centre for trauma patients; therefore on average we would expect patient records to contain more diagnoses.

Santos et al. [20] have demonstrated that workforce issues may affect coding quality at times when coders may have to perform multiple tasks and have tight deadlines for their work. Our results showed that the coding workload is unevenly distributed among coders in Calgary, with approximately half the coders (29 coders) coding less than 10% of the records, while the other half (30 coders) coded the remaining 90% of the records (data not shown). Local coding managers have reported staff shortages at times during the study period [30]. However, our results do not indicate variation in coding validity by season, which would be expected if the Health Records Department was consistently understaffed at specific times of the year or if coders had especially high workloads before the end of the fiscal year in March. Therefore our results do not support the premise that staff shortages are responsible for the uneven distribution of workload.

Besides variation in the mean number of diagnoses coded by hospital site, no other important variation in coding across the other 5 indicators of face validity was demonstrated. Moreover, our analysis by complexity of disease showed (as expected) that the number of diagnoses coded increased with increasing complexity of the major condition. Combined, these findings indicate coding consistency among coders, and provide the administrative data with face validity. In our region coding consistency is achieved through employing coders who are similarly trained and through a regionalized management structure [30]. Most coders undergo 2 years of training in an accredited college program that incorporates work experience. In addition, coders in the 4 major hospitals are managed under one Health Records Department and rotate between hospitals, thereby achieving consistent coding.

Our results demonstrate that while the volume of coding has risen over time, the number of diagnoses coded has decreased. Temporal trends in number of diagnoses coded have also been demonstrated by Lorenzoni et al. [19] in an Italian hospital. These authors showed an increase in diagnoses coded from 1994 to 1997 and attributed this improved trend to training, continuous monitoring and feedback to coders. While the trend to decreasing numbers of diagnoses coded over time in this study could point to less detailed coding and may seem to be of concern, especially to health researchers who look to administrative data to provide more rather than less detail, there are several other possible explanations for this result. First, the Health Records Department in our region may have been understaffed during the study period. This would be consistent with our result that the number of records increased as time went by. As increasing numbers of records are coded each year, coders may have less time to code each one and might therefore decrease the number of codes per record. However, as noted above there was no seasonal variation in mean number of diagnoses even at fiscal year end, which would argue against this explanation. Another possibility is that coding practices may have changed over time in relation to coding of diagnosis types. According to coding managers working at our institution, the coding of certain diagnoses designated as "type 3 diagnoses" is likely to have decreased over time [30]. These diagnoses do not satisfy the requirements to be considered a co-morbidity for administrative data purposes, and may or may not require treatment [31]. However, a change in coding of type 3 diagnoses is unlikely to affect assignment of records to a Case Mix Group (CMG) and as a result will impact minimally on the conclusions drawn by researchers from analyses of administrative data. Second, the documentation provided by physicians may have become less detailed overtime. This is difficult to assess but would not be related to coder characteristics. Third, the severity of illness of patients admitted to hospital in our region may have changed over time, again this would not be related to coder characteristics.

Finally, in investigating the validity of coding using reabstraction methodology in a subset of our data, we saw that for most conditions variations in kappa were similar across employment status, coding volume and site. Psychosis, HIV, Paralysis and Blood loss anemia were the conditions most often associated with wide variations in kappa across coder characteristics; however these differences in kappa did not show a consistent direction in favor of either high or low volume coders, full-time or part-time coders or hospital site.

This study has limitations. First, we had less information about health records coders' characteristics than we would have liked because of privacy concerns. Ideally we would have liked more comprehensive information, such as life time work experience and continuing education, because Santos and others have identified an important impact of ongoing education on coding quality [7, 17, 18, 20, 32]. Second, we did not include rural and small hospitals in our study, although variation in error rates of coding have been previously demonstrated in these settings [28, 29]. Third, it is well known that inadequate documentation provided by clinicians may lead to poor quality coding; unfortunately it was not within the scope of this study to examine clinician documentation, which could have produced some of the variations in the coded data that was seen in this study.

On the other hand, a strength of this study was the use of reabstracted chart review data, which is considered by many researchers to be the "gold standard" when assessing coding validity [15]. In addition, we investigated coding validity in a very large number of hospital discharge abstracts, across a comprehensive list of major diagnoses.

Conclusion

In summary, the quality of administrative data has implications for health records managers, health policy makers and for the reimbursement of healthcare expenditures as well as for academic researchers. Our study provided a comprehensive analysis of coding validity by year, coder characteristics and disease complexity. Our results suggest that coder characteristics do not influence the validity of hospital discharge data and that coding can be remarkably consistent. In Calgary this was achieved by employing coders with professional training and through a consistent management structure that included rotation of coders between sites. Other jurisdictions might benefit by employing these measures to ensure coding validity, thereby improving the quality of administrative data.

Abbreviations

ICD-10-CA:

International classification of disease version 10, Canadian version

ICD-9:

International classification of disease version 9

USA:

Untied States of American

HIV:

Human immune deficiency virus

CMG:

Case mix groups.

References

  1. Iezzoni L: Risk adjustment for measuring health care outcomes. 2003, Chicago: Health Administration Press, 3

    Google Scholar 

  2. Jurrlink D, Preyra C, Croxford R, Chong A, Austin P, Tu JV: Canadian Institute for Health Information Discharge Abstract Database: A Validation Study. 2006, Toronto: Institute for Clinical and Evaluative Sciences

    Google Scholar 

  3. Quan H, Parsons GA, Ghali WA: Validity of information on comorbidity derived rom ICD-9-CCM administrative data. Med Care. 2002, 40 (8): 675-85. 10.1097/00005650-200208000-00007.

    Article  PubMed  Google Scholar 

  4. Quan H, Parsons GA, Ghali WA: Validity of procedure codes in International Classification of Diseases, 9th revision, clinical modification administrative data. Med Care. 2004, 42 (8): 801-9. 10.1097/01.mlr.0000132391.59713.0d.

    Article  PubMed  Google Scholar 

  5. Quan H, Parsons GA, Ghali WA: Assessing accuracy of diagnosis-type indicators for flagging complications in administrative data. J Clin Epidemiol. 2004, 57 (4): 366-72. 10.1016/j.jclinepi.2003.01.002.

    Article  PubMed  Google Scholar 

  6. Roos LL, Gupta S, Soodeen RA, Jebamani L: Data quality in an information-rich environment: Canada as an example. Can J Aging. 2005, 24 (Suppl 1): 153-70. 10.1353/cja.2005.0055. Spring

    Article  PubMed  Google Scholar 

  7. De Coster C, Quan H, Finlayson A, Gao M, Halfon P, Humphries KH, et al: Identifying priorities in methodological research using ICD-9-CM and ICD-10 administrative data: report from an international consortium. BMC Health Serv Res. 2006, 6: 77-10.1186/1472-6963-6-77.

    Article  PubMed  PubMed Central  Google Scholar 

  8. Iezzoni LI, Burnside S, Sickles L, Moskowitz MA, Sawitz E, Levine PA: Coding of acute myocardial infarction. Clinical and policy implications. Ann Intern Med. 1988, 109 (9): 745-51.

    Article  CAS  PubMed  Google Scholar 

  9. Lee DS, Donovan L, Austin PC, Gong Y, Liu PP, Rouleau JL, et al: Comparison of coding of heart failure and comorbidities in administrative and clinical data for use in outcomes research. Med Care. 2005, 43 (2): 182-8. 10.1097/00005650-200502000-00012.

    Article  PubMed  Google Scholar 

  10. Weingart SN, Iezzoni LI, Davis RB, Palmer RH, Cahalane M, Hamel MB, et al: Use of administrative data to find substandard care: validation of the complications screening program. Med Care. 2000, 38 (8): 796-806. 10.1097/00005650-200008000-00004.

    Article  CAS  PubMed  Google Scholar 

  11. Anderson RN, Rosenberg HM: Disease classification: measuring the effect of the Tenth Revision of the International Classification of Diseases on cause-of-death data in the United States. Stat Med. 2003, 22 (9): 1551-70. 10.1002/sim.1511.

    Article  PubMed  Google Scholar 

  12. Janssen F, Kunst AE: ICD coding changes and discontinuities in trends in cause-specific mortality in six European countries, 1950-99. Bull World Health Organ. 2004, 82 (12): 904-13.

    PubMed  Google Scholar 

  13. Quan H, Li B, Saunders LD, Parsons GA, Nilsson CI, Alibhai A, et al: Assessing validity of ICD-9-CM and ICD-10 administrative data in recording clinical conditions in a unique dually coded database. Health Serv Res. 2008, 43 (4): 1424-41. 10.1111/j.1475-6773.2007.00822.x.

    Article  PubMed  PubMed Central  Google Scholar 

  14. O'Malley KJ, Cook KF, Price MD, Wildes KR, Hurdle JF, Ashton CM: Measuring diagnoses: ICD code accuracy. Health Serv Res. 2005, 40 (5 Pt 2): 1620-39. 10.1111/j.1475-6773.2005.00444.x.

    Article  PubMed  PubMed Central  Google Scholar 

  15. Lorence DP, Ibrahim IA: Benchmarking variation in coding accuracy across the United States. J Health Care Finance. 2003, 29 (4): 29-42.

    PubMed  Google Scholar 

  16. Stausberg J, Lehmann N, Kaczmarek D, Stein M: Reliability of diagnoses coding with ICD-10. Int J Med Inform. 2008, 77 (1): 50-7. 10.1016/j.ijmedinf.2006.11.005.

    Article  PubMed  Google Scholar 

  17. Green E, Benjamin C: Impact of the medical record credential on data quality. J Am Med Rec Assoc. 1986, 57 (11): 29-38.

    CAS  PubMed  Google Scholar 

  18. Callen J, Rust J, Hines B, Robertson l, Killen J: Is coding in theatres a viable option? A review of current theatre coding process at Royal Prince Alfred Hospital. Journal of the Australian Confederation of Operating Room Nurses. 1998, 11 (2): 34.

    Google Scholar 

  19. Lorenzoni L, Da Cas R, Aparo UL: The quality of abstracting medical information from the medical record: the impact of training programmes. Int J Qual Health Care. 1999, 11 (3): 209-13. 10.1093/intqhc/11.3.209.

    Article  CAS  PubMed  Google Scholar 

  20. Santos S, Murphy G, Baxter K, Robinson KM: Organisational factors affecting the quality of hospital clinical coding. Him J. 2008, 37 (1): 25-37.

    PubMed  Google Scholar 

  21. D'Hoore W, Bouckaert A, Tilquin C: Practical considerations on the use of the Charlson comorbidity index with administrative data bases. J Clin Epidemiol. 1996, 49 (12): 1429-33. 10.1016/S0895-4356(96)00271-5.

    Article  PubMed  Google Scholar 

  22. Charlson ME, Pompei P, Ales KL, MacKenzie CR: A new method of classifying prognostic comorbidity in longitudinal studies: development and validation. J Chronic Dis. 1987, 40 (5): 373-83. 10.1016/0021-9681(87)90171-8.

    Article  CAS  PubMed  Google Scholar 

  23. Elixhauser A, Steiner C, Harris DR, Coffey RM: Comorbidity measures for use with administrative data. Med Care. 1998, 36 (1): 8-27. 10.1097/00005650-199801000-00004.

    Article  CAS  PubMed  Google Scholar 

  24. Charlson M, Szatrowski TP, Peterson J, Gold J: Validation of a combined comorbidity index. J Clin Epidemiol. 1994, 47 (11): 1245-51. 10.1016/0895-4356(94)90129-5.

    Article  CAS  PubMed  Google Scholar 

  25. Quan H, Sundararajan V, Halfon P, Fong A, Burnand B, Luthi JC, et al: Coding algorithms for defining comorbidities in ICD-9-CM and ICD-10 administrative data. Med Care. 2005, 43 (11): 1130-9. 10.1097/01.mlr.0000182534.19832.83.

    Article  PubMed  Google Scholar 

  26. Cohen J: A coefficients of agreement for nominal scales. Edu and Psych Meas. 1960, 37-46. 10.1177/001316446002000104. 20

  27. Chen G, Faris P, Hemmelgarn B, Walker RL, Quan H: Measuring agreement of administrative data with chart data using prevalence unadjusted and adjusted kappa. BMC Med Res Methodol. 2009, 9: 5-10.1186/1471-2288-9-5.

    Article  PubMed  PubMed Central  Google Scholar 

  28. Lorence DP, Spink A, Jameson R: Manager's reports of variation in coding accuracy across U.S. oncology centers. J Oncol Manag. 2002, 11 (6): 20-6.

    PubMed  Google Scholar 

  29. Rangachari P: Coding for quality measurement: the relationship between hospital structural characteristics and coding accuracy from the perspective of quality measurement. Perspect Health Inf Manag. 2007, 4: 3.

    PubMed  PubMed Central  Google Scholar 

  30. Bulloch S: Health Records Department Staffing in the Calgary Calgary. 2009

    Google Scholar 

  31. CIHI: Coding variation in CIHI discharge abstract database data. 2003, Canadian Institute of Health Research

    Google Scholar 

  32. McKenzie K, Walker S, Spallek M: Australian coder workforce survey 2002--managers' responses. Journal (Inst Health Rec Inf Manag). 2003, 44 (2): 8-14.

    Google Scholar 

Pre-publication history

Download references

Acknowledgements

The authors would like to acknowledge Sally Bulloch, a member of the health records management team who was involved in discussions that gave rise to this study and who discussed the results of this study with the authors. The authors would also like to acknowledge Dr. Stafford Dean who helped with acquiring the data for the study and Bing Li who provided the reabstracted chart data to the authors.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Cynthia A Beck.

Additional information

Competing interests

The authors declare that they have no competing interests.

Authors' contributions

DH: Made a substantial contribution to the concept and design of this study, she analyzed and interpreted the data, she drafted the manuscript, revised it critically and gave final approval for it to be published. HQ: Made a substantial contribution to the concept and design of this study, he acquired the data, was involved in interpretation of the data, critically revised the manuscript and gave final approval for it to be published. PF: Made a substantial contribution to the concept and design of this study, he was involved in interpretation of the data, critically revised the manuscript and gave final approval for it to be published. CB: Made a substantial contribution to the concept and design of this study, she was involved in interpretation of the data, critically revised the manuscript and gave final approval for it to be published.

Electronic supplementary material

12913_2009_1236_MOESM1_ESM.DOC

Additional file 1: Additional table S4. Agreement (kappa) between coded record and chart review data for Elixhauser[22] and Charlson[21] co-morbidities, by coding volume and employment status of coders and coding site (DOC 60 KB)

Authors’ original submitted files for images

Rights and permissions

Open Access This article is published under license to BioMed Central Ltd. This is an Open Access article is distributed under the terms of the Creative Commons Attribution License ( https://creativecommons.org/licenses/by/2.0 ), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.

Reprints and permissions

About this article

Cite this article

Hennessy, D.A., Quan, H., Faris, P.D. et al. Do coder characteristics influence validity of ICD-10 hospital discharge data?. BMC Health Serv Res 10, 99 (2010). https://doi.org/10.1186/1472-6963-10-99

Download citation

  • Received:

  • Accepted:

  • Published:

  • DOI: https://doi.org/10.1186/1472-6963-10-99

Keywords