Hospital discharge diagnostic and procedure codes for upper gastro-intestinal cancer: how accurate are they?
© Stavrou et al.; licensee BioMed Central Ltd. 2012
Received: 21 May 2012
Accepted: 13 September 2012
Published: 21 September 2012
Population-level health administrative datasets such as hospital discharge data are used increasingly to evaluate health services and outcomes of care. However information about the accuracy of Australian discharge data in identifying cancer, associated procedures and comorbidity is limited. The Admitted Patients Data Collection (APDC) is a census of inpatient hospital discharges in the state of New South Wales (NSW). Our aim was to assess the accuracy of the APDC in identifying upper gastro-intestinal (upper GI) cancer cases, procedures for associated curative resection and comorbidities at the time of admission compared to data abstracted from medical records (the ‘gold standard’).
We reviewed the medical records of 240 patients with an incident upper GI cancer diagnosis derived from a clinical database in one NSW area health service from July 2006 to June 2007. Extracted case record data was matched to APDC discharge data to determine sensitivity, positive predictive value (PPV) and agreement between the two data sources (κ-coefficient).
The accuracy of the APDC diagnostic codes in identifying site-specific incident cancer ranged from 80-95% sensitivity. This was comparable to the accuracy of APDC procedure codes in identifying curative resection for upper GI cancer. PPV ranged from 42-80% for cancer diagnosis and 56-93% for curative surgery. Agreement between the data sources was >0.72 for most cancer diagnoses and curative resections. However, APDC discharge data was less accurate in reporting common comorbidities - for each condition, sensitivity ranged from 9-70%, whilst agreement ranged from κ = 0.64 for diabetes down to κ < 0.01 for gastro-oesophageal reflux disorder.
Identifying incident cases of upper GI cancer and curative resection from hospital administrative data is satisfactory but under-ascertained. Linkage of multiple population-health datasets is advisable to maximise case ascertainment and minimise false-positives. Consideration must be given when utilising hospital discharge data alone for generating comorbidity indices, as disease burden at the time of admission is under-reported.
KeywordsValidation study Cancer Comorbidity Administrative data
The assessment of health services utilisation and associated patient outcomes are fundamental to improving health care performance. However, traditional investigation methods such as clinical cohort investigations are resource-intensive and costly. Increasingly, population-level health administrative data, such as hospital discharge and registry data, used alone or linked with other datasets, are being used as a cost-effective and resource-efficient alternative to investigate population treatment patterns, health service utilisation and outcomes of care [1, 2] across a range of conditions [1, 3–9].
Analyses using health administrative data is generally based on the assumption that the data sets have high levels of accuracy in identifying medical conditions and associated treatments and services. In particular there is widespread use of hospital discharge data in this context, yet there are relatively few published studies reporting the accuracy of hospital discharge diagnostic and procedure codes. Some of the well documented limitations are missing data, abstraction errors and misclassification errors . Therefore investigations using these data requires high level expertise from the perspective of the analysts and in the interpretation of findings [11, 12].
In Australia, there have been series of validation studies investigating the accuracy of hospital discharge data in identifying obstetric conditions and outcomes [13–16] but there are fewer studies examining the accuracy of cancer related diagnoses [17–19] and treatment in discharge data [7, 20]. Clearly the identification of accurate population level cancer-related measures from hospital discharge data is essential to improved understanding of treatment processes and outcomes of care. This is particularly important in circumstances where discharge data is used as the only information source and not linked to other population datasets such as cancer notifications.
Upper gastro-intestinal (upper GI) cancers account for 7% of all incident cancers and 15% of all cancer deaths in Australia . Surgical resection for curable upper GI cancers is the standard treatment, with or without adjuvant chemotherapy. We previously reported on patient outcomes following curable surgical resection for oesophageal cancer in New South Wales (NSW), the largest jurisdiction in Australia, using linked administrative health data . We expand on this work by examining the accuracy of hospital administrative discharge diagnostic codes in identifying site-specific upper GI cancer cases and procedure codes for those undergoing curative resection for site-specific cancer. We also examine the accuracy of specific comorbidities as recorded in hospital discharge data compared with those listed in patient medical records.
Australia has a publicly-funded universal health care system. All Australian citizens and permanent residents are entitled to subsidised treatment from medical practitioners and fully subsidised (free) treatment in public hospitals . NSW is the largest jurisdiction in Australia, and until 2011 comprised eight area health services.
Our study population was potential patients with data indicative of a primary incident upper GI cancer (International Classification of Diseases v10 [ICD-10] codes C15, C16, C22 and C25) in the period July 2006 to June 2007 in an area health service (AHS) clinical database. Cases were confirmed using data extracted from their medical records.
The South Eastern Sydney Illawarra Area Health Service commenced capture of the diagnostic and treatment details of all patients diagnosed with cancer after 1st January 2006 or receiving part or all of their treatment within a health service facility. However, the AHS clinical database does not distinguish between primary and secondary cancer diagnoses. Nevertheless, this was the most systematic and cost-effective approach available to us for identifying potential upper GI incident cancer cases from which data could be extracted from patient medical records. Data extracted from the medical records of confirmed cases was considered to be ‘the gold standard’.
Hospital discharge database
The Admitted Patient Data Collection (APDC) is a census of all inpatient separations (discharges) from all public, private and repatriation hospitals, private day procedures centres and public nursing homes in NSW. Hospital medical coders abstract data from patient medical records following discharge and submit details to the NSW Health Information Exchange for every episode of care. A separate record is processed for each period of inpatient care, irrespective of the time interval between the date of separation and subsequent readmission.
Case record extraction
A data extraction form was developed and pilot tested by the research team in consultation with a medical registrar and gastroenterologist. The final version of the abstraction tool had 12 items regarding patient characteristics, cancer-specific diagnostic and treatment characteristics and the presence of common comorbidities (eg hypertension, diabetes, ischaemic heart disease) as suggested for inclusion by the consultants. Most of the items had response options whereby the trained extractor was required to indicate the presence of specific characteristics. To assess inter-rater reliability, a second trained researcher extracted data from a random selection of at least 10% of medical records (n = 38; 12%) independently. Both extractors were health care professionals trained in data abstraction by the medical registrar, and were blind to the diagnostic and treatment details of patients as described in the APDC.
We calculated sensitivity and positive predictive value (PPV) of the APDC data against the case record data (gold standard) for the following: 1) diagnosis for site-specific upper GI cancer; 2) curative resection for site-specific cancer and 3) comorbid conditions. Sensitivity was calculated as the proportion of cases/procedures of cancer or comorbidity reported in the APDC as compared with the true diagnosis/procedure as determined by the case data. We did not calculate specificity or negative predictive value as the denominator (the population without a GI cancer diagnosis) was not ascertained.
Of the persons with site-specific cancer diagnosis/procedure or comorbid condition in the entire APDC (ie true and false report), the PPV was calculated as the proportion with a matching case/procedure from abstracted data (ie true report).
We determined the agreement between the case medical record data and the APDC using the kappa statistic, which adjusts for the agreement that would be observed on the basis of chance. A κ-value >0.75 is an indication of excellent agreement whilst that between 0.40 and 0.75 represents fair to good agreement .
We identified surgical resections in the APDC using the Medicare Benefits Schedule-Extended classification of the International Classification of Diseases (ICD_10_AM) procedural block codes for oesophagectomy (0858–0860), gastrectomy (875–879), pancreatectomy and excision of lesions of pancreas (978 and 979 90294–01, 30578–00) and excision procedures of the liver (953) . Upper GI cancer classification recorded in the APDC as a reason for the episode of care, or comorbidities were identified using ICD_10_AM diagnostic codes from the primary and up to 10 secondary diagnostic fields. Comorbidity ICD-10 diagnostic codes were as follows: hypertension (I10-I15), diabetes (E10.1, E10.5, E10.9, E11.1, E11.5, E11.9, E13.1, E13.5, E13.9, E14.1, E14.5, E14.9), ischemic heart disease (I2-I25), GORD (K21.0, K21.9), alcohol abuse (F10.1, F10.2, K70, Z72.1, Z86.41), hepatitis B or C (B16, B18, B19, K73), chronic obstructive pulmonary disease (J44) and dementia (F01-F03).
We also used the kappa statistic to calculate inter-rater reliability of the extracted data for cancer diagnosis, curative resection and comorbidity. Inter-rater reliability between the record extractors was κ = 0.91 and κ = 0.74 for overall site-specific cancer diagnosis and curative surgery respectively. Agreement for specific comorbidities was κ =1.00 for diabetes, ischaemic heart disease and hypertension and κ =0.64 for gastro-oesophageal reflux disease (GORD). Hence we felt the data obtained by the main extractor could be analysed with confidence.
All statistical analyses were performed using SAS software, version 11.2 (SAS Institute Inc, Cary, NC).
The study was approved by the NSW Population and Health Services Research Ethics Committee (Ref 2010/05/253) and site specific approvals from Prince of Wales, St George, Sutherland and Wollongong Hospitals.
Of the 472 potential cases identified in AHS clinical database, 337 (71%) were available for review; however due to mismatching errors during the data linkage process, four records did not link to the APDC. Of the remaining 333 medical records 240 patients had an incident diagnosis of upper GI cancer; this constituted our study cohort. The majority of patients without incident upper GI cancer (n = 91/333) had another primary cancer type (eg lung, renal cell, unknown primary site) and there was insufficient information in the medical records or associated notes to determine cancer diagnosis type for the remaining two patients.
Over half of the 240 patients with upper GI cancer as classified in the medical chart review were >70 years of age in 2006 (55%), with 35% aged 51–70 years; comparable to patients with no primary diagnosis of upper GI cancer (63% and 27% respectively). 32% (n = 76) of the cases had an incident diagnosis of pancreatic cancer, 30% (n = 71) gastric cancer, 25% (n = 59) oesophageal cancer and 14% (n = 34) liver cancer. Only 27% (64/240) of all upper GI cancer patients had curative surgery; the majority being for gastric cancer (42%) and pancreatic cancer (27%). The most common comorbidity reported in the study cohort was hypertension (50% of patients), with fewer reports of diabetes (32%), GORD (20%), ischaemic heart disease (IHD: 18%), alcohol abuse (9%) or Hepatitis B or C (7%). Similar comorbidity profiles were found in the patients not classified with upper GI cancer from the medical chart review.
Cancer diagnosis and surgical resection
Accuracy of upper GI diagnosis and curative resection reporting, plus reporting of comorbidities in APDC discharge data
Medical records -Condition present
Sensitivity (95% CI)
Positive predictive value
UGI cancer diagnosis
<0.01 (−−, 0.09)
Hepatitis B or C
Misclassification was the most common reason for false-negative reports in the hospital discharge data; for diagnosis 11 oesophageal cancers were classified as gastric cancer, whilst seven gastric, two liver and four pancreatic cancers were misclassified as another primary cancer (such as lung cancer or renal cell carcinoma). Two surgeries for oesophageal cancer were misclassified in the discharge data as occurring for gastric cancer, three pancreatic and three gastric cancer surgeries were misclassified as occurring for liver cancer and one liver cancer was assigned to another primary cancer (renal cell).
PPV for diagnosis of incident cancer in the APDC ranged from 42% for liver cancer to 81% for pancreatic cancer. With the exception of liver cancer, PPV for curative surgery was reasonably high (ranging from 79% to 93%).
Hence, although incident cases and procedures were under-ascertained, fair to very good agreement between the two data sources was shown for cancer diagnosis (κ-coefficient ranging from 0.42 to 0.84) and curative surgery (κ-coefficient ranging from 0.68 to 0.83) (Table 1).
Sensitivity was highest for dementia (70.0%, 95%CI 35-92%) and lowest for GORD (8.5%, 95%CI 3-21%). PPV was variable, ranging from 17% (95%CI 6-38%) for GORD to 98% (95%CI 87-100%) for diabetes. Agreement between the two data sources was low to fair. Comorbidities were underreported in the APDC.
There is widespread and increasing use of population-level hospital discharge diagnostic and procedural codes to monitor processes and outcome of care for health services research. The validation of coding in health administrative datasets has been identified as a priority in health services research by an international consortium . However, validation studies remain uncommon . This study contributes to the body of knowledge regarding the accuracy of administrative hospital discharge data (APDC) for cancer-related diagnoses, associated resection and comorbidity. This study also adds to the current literature on the nature and extent of reporting of comorbid diseases in population administrative data.
The distribution of site-specific cancers reported in our study was similar to the distribution of upper GI cancers in the Australian population . The proportion of patients undergoing curative resection was also consistent with other studies [5, 27, 28]. We demonstrated that the APDC records cancer diagnosis and procedures at an acceptable level when compared to the medical record gold standard. We obtained similar levels of sensitivity for diagnosis compared to previous validation studies examining other cancers [17–19]. Previous validation studies examining breast and prostate cancer diagnoses have also shown surgical procedure records in administrative data to be well reported both in New South Wales [7, 20] and internationally [29, 30]. Nevertheless, true incident cancer cases and procedures from administrative data are still under-reported [31, 32] and hence linkage with other datasets, such as registry-based data is recommended for case-ascertainment. Conversely, using hospital discharge data as the source for detecting incident upper GI cancer will over-estimate the incident rate, as false-positives are reported.
Our study also demonstrated the lower validity of common comorbidities reported in hospital discharge data. Despite an improvement in the accuracy of comorbid coding (with the introduction of ICD-10 coding) in administrative data over recent years, sensitivity when compared with medical records is generally low . Our finding is consistent with a previous Australian study reporting under-ascertainment in more than 80 of 100 conditions in hospital discharge data . The under-ascertainment of comorbidities in administrative data has been attributed to incompleteness of data transfer from medical records in individual hospitals to administrative databases [34–36]. Hospital coders are required to report medical conditions that affect the specific admission, however financial incentives may also impact on comorbidities which are reported. For example, recording of certain comorbidities over others may occur due to their effect on patient length-of-hospital-stay and procedures performed resulting in greater financial re-imbursement to the hospital. Misclassification of similarly related diseases (eg Barrett’s oesophagus versus GORD, COPD versus emphysema) may also occur. Clearly, the under-reporting of common comorbidities has implications for researchers using hospital discharge data as the sole source for assessing incidence, procedures, health outcomes and patient comorbidity as our study and others have identified that case ascertainment is likely to be incomplete.
Population dataset linkage is a cost-effective and attractive method for undertaking health services research, as large representative cohorts can be investigated efficiently when compared with traditional methods of recruiting participants from the population of interest. However, the validation of codes used to identify patients with particular health states and/or undergoing hospital procedures is essential to avoid misclassification bias which has the potential to undermine the internal validity and interpretation of study findings. This validation study showed that hospital discharge data has some limitation in the reporting of cancer, related curative resection and comorbidity. The outcomes from this validation study on cancer-related hospital administrative data will be important for consideration of future cancer surgical prevalence and outcomes by researchers and policy makers.
However, this validation study was limited to one clinical cancer group in an urban AHS in NSW, of which only 70% of the cases could be reviewed. Nevertheless, as cases were from four different hospital sites and the capture of cases is extremely high, it was thought that selection bias would be minimal and that the medical record data were indicative of records in the AHS clinical database. However, future studies should assess data accuracy across multiple jurisdictions and across several cancers. In addition, we used an AHS clinical database to assist in identifying potential GI cancer cases, as doing otherwise would have been significantly more time and cost intensive. This method does not allow for estimates of the NSW population without upper GI cancer, hence specificity and negative predictive value of the hospital discharge data could not be measured.
Hospital administrative data provide a valid method of investigating health outcomes. However cases, procedures and comorbidity in population-level hospital discharge data is under-ascertained and hence researchers and policy-makers should acknowledge this in research and health planning assessments. Linkage across multiple datasets is recommended to improve case ascertainment.
We acknowledge Doctors Shelanah Fernando and David Williams in helping with the design and piloting of the data extraction form, Ms Annabelle Drew for assisting with data abstraction and the hospitals’ medical record teams for assisting with accessing the medical records.
- Ben-Tovim DI, Pointer SC, Woodman R, Hakendorf PH, Harrison JE: Routine use of administrative data for safety and quality purposes - hospital mortality. Med J Aust. 2010, 193: S100-S103.PubMedGoogle Scholar
- Board N, Watson DE: Using what we gather - harnessing information for improved care. Med J Aust. 2010, 193: S93-S94.PubMedGoogle Scholar
- Semmens JB, Platell C, Threlfall TJ, Holman CDJ: A population-based study of the incidence, mortality and outcomes in patients following surgery for colorectal cancer in Western Australia. ANZ J Surg. 2000, 70 (1): 11-18. 10.1046/j.1440-1622.2000.01734.x.View ArticleGoogle Scholar
- Spilsbury K, Semmens JB, Saunders CM, Holman CDJ: Long-term survival outcomes following breast cancer surgery in Western Australia. ANZ J Surg. 2005, 75 (8): 625-630. 10.1111/j.1445-2197.2005.03478.x.View ArticlePubMedGoogle Scholar
- Stavrou EP, Smith GS, Baker DF: Surgical outcomes associated with oesophagectomy in New South Wales; an investigation of hospital volume. J Gastrointest Surg. 2010, 14: 951-957. 10.1007/s11605-010-1198-7.View ArticleGoogle Scholar
- Salz T, Sandler RS: The effect of Hospital and surgeon volume on outcomes for rectal cancer surgery. Clin Gastroenterol Hepatol. 2008, 6 (11): 1185-1193. 10.1016/j.cgh.2008.05.023.View ArticlePubMedPubMed CentralGoogle Scholar
- Goldsbury DE, Smith DP, Armstrong BK, O'Connell DL: Using linked routinely collected health data to describe prostate cancer treatment in New South Wales, Australia: a validation study. BMC Health Services Res. 2011, 11: 253-259. 10.1186/1472-6963-11-253.View ArticleGoogle Scholar
- Taylor LK, Simpson JM, Roberts CL, Olive EC, Henderson-Smart DJ: Risk of complications in a second pregnancy following caesarean section in the first preganancy: a population-based study. Med J Aust. 2005, 183: 515-519.PubMedGoogle Scholar
- Roberts CL, Algert CS, Morris JM, Ford JB, Henderson-Smart DJ: Hypertensive disorders in pregnancy: a population-based study. Med J Aust. 2005, 182: 332-335.PubMedGoogle Scholar
- Bright RA, Avorn J, Everitt DE: Medicaid data as a resource for epidemiologic studies: strengths and limitations. J Clin Epidemiol. 1989, 42: 937-945. 10.1016/0895-4356(89)90158-3.View ArticlePubMedGoogle Scholar
- van Walraven C, Bennett C, Forster AJ: Administrative databse research infrequently used validated diagnostic or procedural codes. J Clin Epidemiol. 2011, 64: 1054-1059. 10.1016/j.jclinepi.2011.01.001.View ArticlePubMedGoogle Scholar
- Terris DD, Litaker DG, Koroukian SM: Health state information derived from secondary databases is affected by multiple sources of bias. J Clin Epidemiol. 2007, 2007: 734-741.View ArticleGoogle Scholar
- Lain SJ, Roberts CL, Hadfield RM, Bell JC, Morris JM: How accurate is the reporting of obstetric haemmorhage in hospital discharge data? A validation study. ANZ J Obstetr Gynaecol. 2008, 48: 481-484.Google Scholar
- Roberts CL, Bell JC, Ford JB, Hadfield RM, Algert CS, Morris JM: The accuracy of reporting of the hypertensive disorders of pregnancy in population health data. Hypertens Pregnancy. 2008, 27: 285-297. 10.1080/10641950701826695.View ArticlePubMedPubMed CentralGoogle Scholar
- Roberts CL, Ford JB, Lain SJ, Algert CS, Sparks CJ: The accuracy of reporting of general anaesthesia for childbirth: a validation study. Anaesth Intensive Care. 2008, 36: 418-424.PubMedGoogle Scholar
- Korst LM, Gregory KD, Gornbein JA: Elective primary caesarean delivery: Accuracy of administrative data. Paediatr Perinat Epidemiol. 2004, 18: 112-119. 10.1111/j.1365-3016.2003.00540.x.View ArticlePubMedGoogle Scholar
- Cooper GS, Yuan Z, Stange KC, Dennis LK, Amini SB, Rimm AA: The sensitivity of Medicare claims data for case ascertainment of six common cancers. Med Care. 1999, 37: 436-444. 10.1097/00005650-199905000-00003.View ArticlePubMedGoogle Scholar
- Couris CM, Schott AM, Ecochard R, Morgon E, Colin C: A literature review to assess the use of claims databases in identifying incident cancer cases. Health Serv Outcomes Research Method. 2003, 4: 49-63. 10.1023/A:1025828911298.View ArticleGoogle Scholar
- Wang PS, Walker AM, Tsuang MT, Orav EJ, Levin R, Avorn J: Finding incident breast cancer cases through US claims data and a state cancer registry. Cancer Causes Control. 2001, 12: 257-265. 10.1023/A:1011204704153.View ArticlePubMedGoogle Scholar
- McGeechan K, Kricker A, Armstrong B, Stubbs J: Evaluation of linked cancer registry and hospital records of breast cancer. Aust NZ J Public Health. 1998, 22: 765-770. 10.1111/j.1467-842X.1998.tb01490.x.View ArticleGoogle Scholar
- Australian Institute of Health and Welfare: Cancer in Australia: an overview. 2010, CanberraGoogle Scholar
- Medicare - background brief. [http://www.aph.gov.au/library/intguide/sp/medicare.htm].
- Centre for Health Record Linkage (CHeReL): Quality Assurance: http://www.cherel.org.au/quality-assurance.
- Fleiss JL: Statistical methods for rates and proportions. 1981, John Wiley, New York, 2Google Scholar
- National Centre for Classification in Health: ICD-10-AM/ACHI/ACS. 2008, Sydney Uo, Sydney, 6Google Scholar
- De Coster C, Quan H, Finlayson A, Gao M, Halfon P, Humphries KH, et al: Identifying priorities in methodological research using ICD-9-CM and ICD-10 administrative data: report from an international consortium. BMC Health Serv Res. 2006, 6: 77-10.1186/1472-6963-1186-1177.View ArticlePubMedPubMed CentralGoogle Scholar
- Smith GS, Minehan E: The SUGSS database:a web based upper GI cancer database. Aust NZ J Surg. 2009, 79 (Suppl 1): A41.View ArticleGoogle Scholar
- van Heek NT, Kuhlmann KFD, Scholten RJ, de Castro SMM, Busch ORC, van Gulik TM, Obertop H, Gouma DJ: Hospital volume and mortality after pancreatic resection - a systematic review and an evaluation of intervention in The Netherlands. Ann Surg. 2005, 242 (6): 781-790. 10.1097/01.sla.0000188462.00249.36.View ArticlePubMedPubMed CentralGoogle Scholar
- Malin JL, Kahn KL, Adams J, Kwan L, Laouri M, Ganz PA: Validity of cancer registry data for measuring the quality of breast cancer care. J Natl Cancer Inst. 2002, 94: 835-844. 10.1093/jnci/94.11.835.View ArticlePubMedGoogle Scholar
- Pinfold SP, Goel V, Sawka C: Quality of hospital discharge and physician data for type of breast cancer surgery. Med Care. 2000, 38: 99-107. 10.1097/00005650-200001000-00011.View ArticlePubMedGoogle Scholar
- Bernal-Delgado EE, Martos C, Martinez N, Chirlaque MD, Marquez M, Navarro C, et al: Is hospital discharge administrative data an appropriate source of information for cancer registries purposes? Some insight from four Spanish registries. BMC Health Serv Res. 2010, 10: 9-14. 10.1186/1472-6963-10-9.View ArticlePubMedGoogle Scholar
- Couris CM, Seigneurin A, Bouzbid S, Rabilloud M, Perrin P, Martin X, Colin C, Schott AM: French claims data as a source of information to describe cancer incidence: predictive values of two identification methods of incident prostate cancers. J Med Syst. 2006, 30: 459-463. 10.1007/s10916-006-9028-x.View ArticlePubMedGoogle Scholar
- Januel J-M, Luthi J-C, Quan H, Borst F, Taffe P, Ghali WA, Burnand B: Improved accuracy of co-morbidity coding over time after the introduction of ICD-10 administrative data. BMC Health Serv Res. 2011, 11: 194-207. 10.1186/1472-6963-11-194.View ArticlePubMedPubMed CentralGoogle Scholar
- Preen DB, Holman CDJ, Lawrence DM, Baynham NJ, Semmens JB: Hospital chart review provided more accurate comorbidity information than data from a general practitioner survey or an administrative database. J Clin Epidemiol. 2004, 57: 1295-1304. 10.1016/j.jclinepi.2004.03.016.View ArticlePubMedGoogle Scholar
- Powell H, Lim LL-Y, Heller RF: Accuracy of administrative data to assess comorbidity in patients with heart disease: an Australian perspective. J Clin Epidemiol. 2001, 54: 687-693. 10.1016/S0895-4356(00)00364-4.View ArticlePubMedGoogle Scholar
- Hawker GA, Coyte PC, Wright JG, Paul JE, Bombardier C: Accuracy of administrative data for assessing outcomes after knee replacement surgery. J Clin Epidemiol. 1997, 50: 265-273. 10.1016/S0895-4356(96)00368-X.View ArticlePubMedGoogle Scholar
- The pre-publication history for this paper can be accessed here:http://www.biomedcentral.com/1472-6963/12/331/prepub
This article is published under license to BioMed Central Ltd. This is an Open Access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/2.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.