- Research article
- Open Access
- Open Peer Review
This article has Open Peer Review reports available.
How to limit the burden of data collection for Quality Indicators based on medical records? The COMPAQH experience
© Corriol et al; licensee BioMed Central Ltd. 2008
Received: 16 November 2007
Accepted: 21 October 2008
Published: 21 October 2008
Our objective was to limit the burden of data collection for Quality Indicators (QIs) based on medical records.
The study was supervised by the COMPAQH project. Four QIs based on medical records were tested: medical record conformity; traceability of pain assessment; screening for nutritional disorders; time elapsed before sending copy of discharge letter to the general practitioner. Data were collected by 6 Clinical Research Assistants (CRAs) in a panel of 36 volunteer hospitals and analyzed by COMPAQH. To limit the burden of data collection, we used the same sample of medical records for all 4 QIs, limited sample size to 80 medical records, and built a composite score of only 10 items to assess medical record completeness. We assessed QI feasibility by completing a grid of 19 potential problems and evaluating time spent. We assessed reliability (κ coefficient) as well as internal consistency (Cronbach α coefficient) in an inter-observer study, and discriminatory power by analysing QI variability among hospitals.
Overall, 23 115 data items were collected for the 4 QIs and analyzed. The average time spent on data collection was 8.5 days per hospital. The most common feasibility problem was misunderstanding of the item by hospital staff. QI reliability was good (κ: 0.59–0.97 according to QI). The hospitals differed widely in their ability to meet the quality criteria (mean value: 19–85%).
These 4 QIs based on medical records can be used to compare the quality of record keeping among hospitals while limiting the burden of data collection, and can therefore be used for benchmarking purposes. The French National Health Directorate has included them in the new 2009 version of the accreditation procedure for healthcare organizations.
Medical records are a key instrument in the coordination of patient care. They facilitate diagnosis and information sharing, reduce medical errors, and serve an important medical-legal function , regardless of the type of healthcare organisation (HCO) – public or private – or clinical specialty. However, when the first accreditation results for French HCOs were made public in 2003 their quality was found to be in need of substantial improvement. Two frequent shortcomings were no written record of the information provided to patients and unsigned drug prescriptions . Ever since the French health authorities announced that each person with national health insurance coverage will have a single electronic medical record shared between patient and health practitioners, they have become a hot topic .
Medical records are also used as a source of clinical data for documenting Quality Indicators (QIs) [4–6], although other methods providing more relevant data, such as standardized patients or vignettes [7–9], are also available. However, since the latter have not been used on a wide scale to compare hospitals, collecting data for QIs from medical records remains the standard despite the workload and expense of chart abstraction. The estimated total cost of copying and reviewing 8 000 charts manually is US$10 millions .
Our study focuses on the need to improve the quality of medical records because of their key role in the coordination of care. We developed and tested a set of four QIs for medical records (record conformity, traceability of pain assessment, screening for nutritional disorders, and time elapsed before sending discharge letters) and collected data in a panel of 36 volunteer hospitals. We wanted to establish whether these QIs could be used by all types of hospital while limiting the burden of data collection. We determined QI feasibility, reliability, and ability to discriminate among the hospitals.
The study was run by the COMPAQH project, a French national initiative for the development of QIs, coordinated by the French Institute for Health and Medical Research (INSERM), and sponsored by the Ministry of Health and National Health Directorate . The project's overall objectives are to select and test QIs in order to monitor quality and performance in French hospitals, and design ranking methods and Paying-for-Quality programs.
In 2003, the French Ministry of Health and the French National Authority for Health (HAS), listed 8 priority areas in need of quality improvement: continuity of care, pain management, management of patients with nutritional disorders, iatrogenic risks, patient satisfaction, follow-up of practice guidelines, management of human resources, and access to care. A set of 42 QIs relating to these areas was established by COMPAQH. Four of these QIs were based on hospital medical records and corresponded to 3 priority areas: QI 1: conformity of medical records (continuity of care), QI 2: traceability of pain assessment (pain management), QI 3: screening for nutritional disorders (management of patients with nutritional disorders), and QI 4: time elapsed before sending a copy of discharge letters to general practitioners (continuity of care).
Medical record conformity
Composite score describing compliance with 10 items: presence of surgical report, delivery report, anaesthetic record, transfusion record, outpatient prescription, outpatient record, admission documents, care and medical conclusions at admission, drug prescriptions during stay, and overall organisation of record
Traceability of pain assessment
Proportion of records containing at least one pain assessment result (Number of records containing at least one result/N)
Screening for nutritional disorders
Proportion of records giving body weight at admission (Number of records giving weight at admission/N)
Time elapsed before sending discharge letters
Proportion of records containing a letter sent within 8 days (Number of records containing a letter/N)
A panel of 36 volunteer hospitals (16 public, 7 private not-for-profits, and 13 private profit-making) took part in collecting data on the 4 selected QIs in 2005. They comprised 24 acute care establishments, 4 cancer clinics, 4 rehabilitation centres, and 4 psychiatric hospitals. The 4 QIs except for QI 2 (traceability of pain assessment) were applicable to all establishments; QI 2 applied to acute care hospitals only.
There were 6 steps to the data collection process: (1) Diffusion of an instructions brochure for each QI describing the protocol and items for which data was to be collected; (2) Nomination of a person within each hospital who would oversee data collection; (3) Random selection of 80 medical records; (4) Data collection by 6 CRAs who used each medical record to complete the quality assessment grid under the supervision of a physician; (5) Calculation of results; (6) Summary of the strengths and weaknesses of each QI, and diffusion of the validated instructions brochures to the bodies responsible for generalising QI use.
The burden of data collection was reduced by using: (i) a single sample of medical records for all 4 QIs. The person who oversaw data collection in each hospital randomly drew the records; (ii) a small sample size (80 medical records for 4 QIs) that was nevertheless large enough to reveal differences among hospitals [12, 13]. However, because of exclusions, a second and even a third set of 30 records, was also drawn. The mean number of records abstracted in each hospital was 93 (range 80–132, median 86); (3) a limited number of items for the QI for medical record conformity. The initial list of 16 items was cut down to 10 items after a working group had identified feasibility problems relating to some items (see below).
None of the 4 QIs required adjustment.
We drew up a list of problems that the CRAs might encounter during data collection. A working group of 5 physicians and 5 hospital managers then examined whether the list was complete by checking it against a number of medical records and made observations on the wording used for each item in the list. Their comments were used to revise the instructions brochure prior to its testing by the CRAs. The final list comprised 29 pre-identified problems. The CRAs completed the 29-item form for each QI in each hospital and also recorded all non-listed problems they had to deal with (e.g. problems relating to clarity of instructions or to time spent in collecting data on a specific item). In each hospital, they also estimated the time taken to sample 80 medical records, retrieve the medical records from archives, record the data on paper, enter it on computer, and control its quality.
The working group validated the amended version of the instructions brochure which was then used to assess QI reliability and discriminatory power.
In 6 hospitals, a sample of 20 out of the 80 medical records was coded by 2 different CRAs (double-blind data capture). Reliability was given by the Kappa (κ) coefficient. A κ value greater than 0.6 indicates satisfactory reliability; a value above 0.8 indicates very satisfactory reliability . For QI 1 (medical record conformity), we also computed the Cronbach α coefficient to assess the score's internal consistency and calculated inter-item correlations to establish whether any items might not provide redundant information .
The hospitals were ranked into 5 categories using the Ontario Hospital Report method . We chose this method because of the need to introduce uncertainty (i.e. confidence intervals, CI) into the ranking as the sample of medical records supporting the data was small . The discriminatory power of each QI was assessed by comparing the results for each hospital. The hospitals were ranked into 5 groups on the basis of the overall mean 90% and 99% confidence intervals (CI) as indicated in the footnote to Table 3.
Burden of data collection
The estimated time taken to collect data on the 4 QIs in each hospital was 8.5 days, broken down as follows: 0.5 days to sample the medical records, 2 days to retrieve the medical records from archives, 4 days to abstract the sample, 1 day to enter the data on computer, and 1 day to check data quality.
Feasibility problems encountered
Before data collection
After data collection
Low implication/motivation of institution
Low hands-on implication/motivation
Low implication/motivation due to external events unrelated to the project
Staff unavailable over the phone
Staff unavailable to make an appointment
Staff did not turn up at appointment
Staff unavailable by email
Staff unavailable owing to an unexpected event
Staff misunderstood QI description in instructions brochure
Staff misunderstood QI in assessment grid
Staff misunderstood QI in assessment grid instructions
Non-compliance with instructions by archives department
Non-compliance with instructions by department of medical information
Non-compliance with instructions by person in charge of data collection
Non-compliance with instructions by medical secretariat
Non-compliance with instructions by the physician
Non-compliance with protocol instructions for other reasons
Difficulty for CRA to access all data in the medical records
The overall rates of occurrence of feasibility problems for each QI were 16.1% for medical record conformity, nutritional disorders, and discharge letters, and 16.7% for pain assessment. No CRA reported a problem that was an in-built limitation on feasibility. No CRA and no person in charge of data collection in each hospital reported a critical feasibility problem. An analysis of each problem with the working group led to improvements in the wording used in the amended version of the instructions.
Inter-observer reliability was satisfactory for 3 of the 10 items of the QI for medical record conformity and very satisfactory for the remaining 7 items (κ: 0.59 – 0.97). Reliability was very satisfactory for the single item which could be computed for the 3 other QIs (κ: 0.86 for pain; 0.93 for nutritional disorder, and 0.96 for discharge letter). The internal consistency of the composite score of 10 items was acceptable (Cronbach α coefficient: 0.72). There were too few inter-item correlations to be able to delete any item (only 3 coefficients were significantly above 0.50 in absolute value).
QI discriminatory power
Comparisons among hospitals based on the quality of medical records in 2005
Overall mean (range)
Number of hospitals in each category
Medical record conformity
Traceability of pain assessment by department
- General medicine
Screening for nutritional disorders
Time elapsed before sending copy of discharge letter
We developed 4 practicable and acceptable QIs based on chart abstraction and covering different aspects of quality of care (continuity of care, staff coordination, coordination between hospitals and general practitioners, pain management and awareness of nutritional disorders). The feasibility and reliability of the 4 QIs were good. The motivation of the 36 hospitals which had volunteered to take part in the test might partly account for this result. The 4 selected QIs were also able to discriminate among hospitals, suggesting that they could be used nationwide for benchmarking purposes, to identify best performance and analyze best practice.
QI reliability was demonstrated by double-blind data collection. Since the validity of using data collected by in-house hospital staff for performing comparisons among HCOs may be questioned, external quality control of data is necessary. This is for instance the case when medical and administrative data are collected to take decisions on funding at a national level. In our study, the data was collected by external CRAs.
The burden of data collection was minimised by collecting data for all 4 QIs from a single small sample of medical records and by restricting the number of items for the QI relative to the completeness of the medical record. Sample size for 4 QIs was hardly any larger than for a single QI. Data on clinical QIs relating to specific diseases needs to be abstracted by a physician, thus increasing costs. However, our data collection did not require technical medical knowledge and could be carried out by a CRA under the supervision of a physician.
Additional improvements could further reduce the workload. Working prospectively could save the time taken to retrieve records from archives (2 days). Abstracting them directly on computer using logical and quality tests could also save time (2 days). The workload could thus be reduced from 8.5 to 4.5 days. If and when electronic medical records replace manual records [10, 18], data collection and the computing of results could become fully automated.
As regards generalisation of the 4 QIs to all French hospitals, our study revealed several limitations. The first was that COMPAQH CRAs and not hospital staff assessed QI feasibility. This was so that we could assess true QI reliability and ensure good hospital participation. For hospitals to collect data there will be a need for greater transfer of knowledge and staff training. In addition, hospitals will need to find the resources to cover the cost of the 8.5 working days needed to collect data.
A second limitation concerns the items used to assess the completeness of the medical record (QI 1). These items are legal requirements and will change with changes in the law. Some legal requirements, however, are more difficult to implement than others, in particular those arising from a precautionary principle applied under public pressure. Results will then be poor and benchmarking will be impossible. An umbrella institution should thus oversee the generalization of these QIs. It could be the French National Health Directorate, which oversees quality assessment and improvement in many fields of healthcare, from general practice to HCO accreditation.
The relevance of chart abstraction itself is controversial: does the QI reflect the quality of care or just the reporting of information? If the quality improvements we have studied just boil down to improvements in reporting, this may turn out to be an inadequate response to the need to improve quality of care [19, 20]. Differences in the reporting of items required by law may then have accounted for the huge differences we observed in hospital QI rates. There is, however, evidence showing that data recording does contribute to quality of care . If this is indeed the case, then improving quality of care requires improving the completeness of medical records.
Finally, a sample of fewer than 100 medical records may suffice to compare hospitals, but not hospital departments, with acceptable accuracy. A department may be a more relevant unit of study than a hospital.
In conclusion, this is to our knowledge the first set of QIs using a leverage effect and lowering the burden of data collection through pooling. By using a representative panel of hospitals, trained CRAs and appropriate tools, the COMPAQH project has shown that these QIs can reveal significant differences among hospitals. The National Health Directorate has decided to include them in the next version (2009) of the accreditation procedure in which all the 3000 or so HCOs in France have to take part. Widespread use of these QIs should inspire a culture of quality measurement and the development of further QIs.
We thank the statistics team (P. Gérard-Dematons, F. Capuano, M. Couralet), the members of the working group (J. Béranger, C. Billiard-Decré, S. Calmus, M. El-Khammar, Q. Escartin, C. Gardel, S. Goubet, S. Morin, B. Saiman), and the hospital representatives who took part in the project. The list of hospitals participating in the COMPAQH project and further details are available (in French) on the COMPAQH website: http://ifr69.vjf.inserm.fr/compaqh/
Several institutions have supported the COMPAQH project financially: National Health Directorate, French Ministry of Health, Directorate for Hospitals and Organization of Care, National Health Insurance, Federations of Private and Public Hospitals, National Federation of French Sickness Funds, French Federation of Cancer Centres.
- Wood DL: Documentation guidelines: evolution, future direction, and compliance. Am J Med. 2001, 110: 332-334. 10.1016/S0002-9343(00)00748-8.View ArticlePubMedGoogle Scholar
- Daucourt V, Michel P: Results of the first 100 accreditation procedures in France. Int J Qual Health Care. 2003, 15: 463-471. 10.1093/intqhc/mzg071.View ArticlePubMedGoogle Scholar
- French Health Insurance Act, n°2004-810, 13 August 2004, Article n°5 (Loi relative à l'assurance maladie du 13 août 2004). Accessed November 4th 2008, [http://www.legifrance.gouv.fr/jopdf/common/jo_pdf.jsp?numJO=0&dateJO=20040817&numTexte=2&pageDebut=14598&pageFin=14626]
- Schneider KM, Wiblin RT, Downs KS, O'Donnell BE: Methods for evaluating the provision of well child care. Jt Comm J Qual Improv. 2001, 27: 673-682.PubMedGoogle Scholar
- Keating NL, Landrum MB, Landon BE, Ayanian JZ, Borbas C, Guadagnoli E: Measuring the quality of diabetes care using administrative data: is there bias?. Health Serv Res. 2003, 38: 1529-1545. 10.1111/j.1475-6773.2003.00191.x.View ArticlePubMedPubMed CentralGoogle Scholar
- Gilbert EH, Lowenstein SR, Koziol-McLain J, Barta DC, Steiner J: Chart reviews in emergency medicine research: where are the methods?. Ann Emerg Med. 1996, 27: 305-308. 10.1016/S0196-0644(96)70264-0.View ArticlePubMedGoogle Scholar
- Luck J, Peabody JW, Dresselhaus TR, Lee M, Glassman P: How well does chart abstraction measure quality? A prospective comparison of quality between standardized patients and the medical record. Am J Med. 2000, 108: 642-649. 10.1016/S0002-9343(00)00363-6.View ArticlePubMedGoogle Scholar
- Dresselhaus TR, Peabody JW, Lee M, Wang MM, Luck J: Measuring Compliance with preventive care guidelines:standardized patients, clinical vignettes and the medical record. J Gen Intern Med. 2000, 15: 782-788. 10.1046/j.1525-1497.2000.91007.x.View ArticlePubMedPubMed CentralGoogle Scholar
- Peabody JW, Luck J, Glassman P, Dresselhaus TR, Lee M: Comparison of vignettes, standardized patients, and chart abstraction: a prospective validation study of 3 methods for measuring quality. JAMA. 2000, 283: 1715-1722. 10.1001/jama.283.13.1715.View ArticlePubMedGoogle Scholar
- Weiner M, Stump TE, Callahan CM, Lewis JN, McDonald CJ: Pursuing integration of performance measures into receptor electronic medical records: beta-adrenergic antagonist medications. Qual Saf Health Care. 2005, 14: 99-106. 10.1136/qshc.2004.011049.View ArticlePubMedPubMed CentralGoogle Scholar
- Grenier-Sennelier C, Corriol C, Daucourt V, Aronica E, Michel P, Minvielle E: Developing quality indicators in hospitals: the COMPAQH project. Rev Epidemiol Sante Publique. 2005, 53 (Spec No 1): 1S22-30.PubMedGoogle Scholar
- McGlynn EA, Kerr EA, Adams J, Keesey J, Asch SM: Quality of health care for women: a demonstration of the quality assessment tools system. Med Care. 2003, 41: 616-625. 10.1097/00005650-200305000-00009.PubMedGoogle Scholar
- Arkin CF, Wachtel MS: How many patients are necessary to assess test performance?. JAMA. 1990, 12 (2): 275-278. 10.1001/jama.263.2.275.View ArticleGoogle Scholar
- Streiner D, Norman G: Health Measurement Scales. A Practical Guide to their Development and Use. 1995, Oxford University Press, 2Google Scholar
- Nunnally JC: Psychometric Theory. 1978, New York: McGraw-Hill, 2Google Scholar
- Ontario Hospital Report, Technical Summary Patient Satisfaction 2005. Accessed November 4th 2008, [http://www.hospitalreport.ca/downloads/2005/AC/AC_CUO_TechReport_FINAL.pdf]
- Bird SM, Cox D, Farewell VT, Goldstein H, Holt T, Smith PC: Performance indicators: good, bad, and ugly. J R Statistical Soc. 2005, 168: 1-27. 10.1111/j.1467-985X.2004.00333.x.View ArticleGoogle Scholar
- Morris L, Taylor M, Campbell LM, Sullivan FM: How will practices cope with information for the new GMS contract? Coronary heart disease data recording in five Scottish practices. Inform Prim Care. 2003, 11: 121-127.PubMedGoogle Scholar
- Fairbrother G, Hanson KL, Friedman S, Butts GC: The impact of physician bonuses, enhanced fees, and feed-back on childhood immunization coverage rates. Am J Public Health. 1999, 89: 171-175. 10.2105/AJPH.89.2.171.View ArticlePubMedPubMed CentralGoogle Scholar
- Lofgren JP, Brown HR, Lesinger C, Sims W: Increase in Alabama public health immunization levels coinciding with monthly distribution of county immunization levels comparisons and financial rewards for accomplishments. Conference abstracts from the 31st National Immunization Conference; May 19–22. 1997, ; Detroit, MichGoogle Scholar
- Solomon DH, Schaffer JL, Katz JN, Horsky J, Burdick E, Nadler E, Bates DW: Can history and physical examination be used as markers of quality? An analysis of the initial visit note in musculoskeletal care. Med Care. 2000, 38: 383-391. 10.1097/00005650-200004000-00005.View ArticlePubMedGoogle Scholar
- The pre-publication history for this paper can be accessed here:http://www.biomedcentral.com/1472-6963/8/215/prepub
This article is published under license to BioMed Central Ltd. This is an Open Access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/2.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.