- Research article
- Open Access
- Open Peer Review
Keys to successful implementation of a French national quality indicator in health care organizations: a qualitative study
BMC Health Services Researchvolume 16, Article number: 553 (2016)
Several countries have launched public reporting systems based on quality indicators (QIs) to increase transparency and improve quality in health care organizations (HCOs). However, a prerequisite to quality improvement is successful local QI implementation. The aim of this study was to explore the pathway through which a mandatory QI of the French national public reporting system, namely the quality of the anesthesia file (QAF), was put into practice.
Seven ethnographic case studies in French HCOs combining in situ observations and 37 semi-structured interviews.
A significant proportion of potential QAF users, such as anesthetists or other health professionals were often unaware of quality data. They were, however, involved in improvement actions to meet the QAF criteria. In fact, three intertwined factors influenced QAF appropriation by anesthesia teams and impacted practice. The first factor was the action of clinical managers (chief anesthetists and head of department) who helped translate public policy into local practice largely by providing legitimacy by highlighting the scientific evidence underlying QAF, achieving consensus among team members, and pointing out the value of QAF as a means of work recognition. The two other factors related to the socio-material context, namely the coherence of information systems and the quality of interpersonal ties within the department.
Public policy tends to focus on the metrological validity of QIs and on ranking methods and overlooks QI implementation. However, effective QI implementation depends on local managerial activity that is often invisible, in interaction with socio-material factors. When developing national quality improvement programs, health authorities might do well to specifically target these clinical managers who act as invaluable mediators. Their key role should be acknowledged and they ought to be provided with adequate resources.
Recently, health authorities in many countries have launched public reporting campaigns to improve the quality of care based on validated quality indicators (QIs) QIs should be understood here as statistical measures that assess output or processes. The premise is that public disclosure of QI scores will increase transparency and encourage quality improvements in health care organizations (HCOs) [1–3].
Quality improvements necessitate the development not only of valid QI measures but also the means to enable the introduction of corrective interventions that befit the local context . Indeed, QI measurement alone does not ensure corrective interventions that will improve QI scores . Accounting is a matter of “style” rather than “substance” . Thus, the value of a QI depends on its adoption by HCO staff and in the changes in practice or work organization that will improve quality. For instance barriers arising during implementation may lead to suboptimal QI assessment (lower scores than expected)  and become divorced from operational activities or induce unwelcome effects (for example, defensive responses, a decrease in quality, and gaming practices [7–9]).
A recent review on causal pathways between performance measure use and effect stressed the need to develop research in the area of QI implementation . In a national public reporting system a mandatory assessment tool (the QI) is used to acquire quality improvement results, either at the macro level to support investment decisions or at the micro level to help frontline workers improve service . This contrasts with the usual global way local QI implementation is considered in terms of total quality management tools or continuous quality improvement approaches and strategies [12–15].
According to the emerging field of the implementation sciences , this study aims to explore the pathway through which measurement tools come to be operationalized in practice.
To do so we link empirical findings to a pragmatic approach reminiscent of cultural-historical activity theory . This approach focuses on the impact of a tool (here, a QI) on human actions and practice. A QI does not operate in an isolated context but within an action scheme (i.e. within an interpretative framework described by an implicit “how to use” guide for the tool) that is constantly being updated through individual practitioners’ experiences.
To understand the factors influencing effective QI use, and the interaction between these factors and the local context, we undertook a qualitative analysis of a QI - the quality of the anesthesia file (QAF).- over a two-year period (2011–2013).
Choice of QI
In order to focus on the issue of implementation in a local context (rather than its relevance), we selected a QI that had characteristics that were in strong alignment with professional practice. Anesthestists have developed a strong culture of data traceability  and consider the anesthesia file as representing a major quality issue. Their work requires planning  and collaboration [20, 21]. This is especially true in France where any patient planning to have surgery must consult an anesthetist one month prior to the surgery date, as well as a few hours before surgery at admission. These consultations are often made by different doctors working in a large team. Thus, anesthetists’ activity relies heavily on the data included in a patient’s anesthesia file. The quality of the anesthesia file (QAF) is one of the 62 QIs that are part of the French national public reporting system introduced in 2006 and coordinated by the French Ministry of Health and French National Authority for Health (Haute Autorité de Santé, HAS) . The anesthesia file is part of the patient file, is used primarily by anesthesia physicians and nurse anesthetists (sometimes by surgeons), and remains a paper document in most French HCOs.
The QAF assesses whether certain elements of information using 13 key criteria (Table 1) are present in the anesthesia file over 3 distinct periods (pre-operative, per-operative, and post-operative). The 13 criteria were selected as a result of research studies and validated by professionals, members of the French Anesthesia and the Intensive Care Society (Société Française d’Anesthésie et de Réanimation, SFAR), . The premise is that information shared among anesthetists provides higher quality care.
Since 2008, each of the 1300 acute care HCOs in France have had to evaluate their compliance with the 13 criteria in a random sample of 60 files for patients having undergone surgery the preceding year. A composite score is calculated and expressed as a percentage of average score of the 60 files. The national target as defined in 2009 is 80 %. Assessment was conducted yearly until 2011 and was then reduced to every 2 years to reduce the burden of data collection.
We attended three 3-h work sessions organized by the HAS and SFAR to learn about QAF design. We retrieved the minutes of earlier meetings as well as official documents and reports to acquire the background knowledge needed for our case studies (2012–2013). In order to learn more about the local contexts that could affect how the QAF was implemented we selected 6 diverse HCOs (Table 1), distinct in size, types (teaching/nonteaching hospitals), status (public/private), and location. Each HCO works as a “polar case” , providing an opportunity to explore significant phenomena. Before contacting the individual anesthesia teams we first sent a formal e-mail to the management of each of the HCOs outlining our research objectives and to obtain institutional approval.
In order to focus on the implementation process without normative consideration, in a first step our case selection did not take into account the two available QAF scores (2010/2011) obtained by the HCOs. Two of the six HCOs (A & E) had in fact achieved excellent QAF scores in their last assessment. The four other HCOs had slightly lower scores but nevertheless showed marked progress. In a second step, a contrasting HCO case study was selected whose QAF score had fallen by 24 % between 2010 and 2011 (G in Tables 1 and 2).
In each HCO observations were made and semi-structured interviews were held by at least 2 researchers in order to ensure standard reporting. Observations (1–5 days) involved tracking the anesthesia file. The anesthesia file was approached as “work practice centered” as noticed elsewhere . It allowed us to identify the different professional actors interacting with the anesthesia file. We observed how it was completed and used in the anesthesia consultation offices and in the operating room area in five of the seven HCOs. During our observations staff were asked what they knew about the QAF and what they thought about any changes that had been introduced since the QAF has been implemented. Several informal “hallway” interviews were also carried out during this time. We conducted 37 semi-structured interviews with staff involved in the QAF implementation and/or data collection (4 to 9 members/HCO according to HCO type) (Table 1): 18 anesthetists (all full-time), 10 nurses, 3 quality managers, and 6 stakeholders (e.g., head of Information Systems Department). The interview guide developed by the researchers was used to question their knowledge of QAF and the local aspects of QAF implementation. Overall, 35/37 interviews were digitally recorded and fully transcribed. Two interviews were not audio recorded, but we took detailed notes during interviews. Documents (whether in house or not) were collected on the QAF scores and improvement initiatives in each HCO.
The data from the case studies were analyzed in several ways. Following an inductive approach, we first used open coding to summarize segments of data. In a second step, as data collection progressed, the codes with common elements were merged into categories in line with the issue of QAF implementation (use of electronic health record, improvement actions, interpersonal ties…). These categories were then compared through cross case analysis during monthly meeting with the research team, allowing us to identify factors influencing QAF implementation according to the local context. At each step of the data analysis validity was increased by the use of multiple data sources .
Limited perceived use of QAF and quality Improvements
Very few anesthetists and nurses were aware of the QAF. In general, QAF scores were not perceived to hold any special relevance to how well an anesthesia unit functioned: “I flicked through the information on the indicator and went on to my next task” (Chief anesthetist, HCO A); “The anesthesia teams do not show much interest in the yearly scores” (Quality manager, HCO B). Our observations in the operating area underscored these comments. On presenting our study to the anesthesia teams, we nearly always had to explain the meaning of QAF. Awareness of QAF was restricted to staff members who had participated in data collection (the chief anesthetist in HCO D had even forgotten despite his involvement) and, in the contrasting case (HCO G), to the chief anesthetist who was the only person to mention QAF scores during our visit.
However staff members were largely involved in quality improvement initiatives to meet QAF criteria. Anesthetists and nurse anesthetists in HCO D considered the changes made to the anesthesia file to be improvements in practice but never mentioned that they considered the QAF criteria to be at the root of these changes. This was indeed confirmed by the quality manager. Over the course of our study the QAF scores across the first 6 HCOs improved steadily. Changes were made to the anesthesia file in all 7 HCOs to facilitate documenting criteria, ensure better quality of information, and achieve higher QAF scores. In HCO G, the newly appointed chief anesthetist planned to use the QAF to underscore work organization problems. In HCO B, a task force was tackling difficult intubation, an important but often poorly documented safety criterion..In HCO C, two boxes restricting choice (standard risk or high risk) were added to the blank space in the last section of the file.
Key role of clinical managers in QAF implementation
The anesthetists in charge of data collection in all 7 HCOs developed an interest in QAF. In HCO C, the anesthetist who collected data on all 60 files decided to analyze the data himself, made a note of potential problems, and submitted his analysis for discussion with other staff. In HCO D, whilst collecting data, the chief anesthetist noted that the QAF scores reflected differences in post-anesthesia care in facilities at two different sites (Table 1, criteria 10–12) (criterion 10 was documented in 40 % of files at the smaller site versus 98 % at the larger site). The physician’s signature authorizing patient discharge from the recovery room was present in only 60 % of files at the smaller site. “I was not aware that documenting was poor at this site (…). The process enabled us to evaluate our work better. If there had been a court case and a judge had discovered from the file that the patient had been discharged without the physician’s signature, where would we have been?” (Chief anesthetist, HCO D)Participating in data collection is a means not only of producing statistics for external accountability but also of generating awareness of the quality of the work done. An anesthetist in HCO D, who was proud to discover that the unit’s files were well kept, was motivated to devote her time and energy to the implementation of improvement initiatives to meet QAF criteria.
All the chief anesthetists and some of the anesthetists involved in data collection were also involved in promoting corrective actions. The reasoning adopted by these clinical managers (chief anesthetists) and the improvement initiatives they proposed to meet QAF objectives were seen to be legitimate as they were in line with local concerns and needs. We identified 3 lines of reasoning:
Scientific evidence: Health professionals are more likely to accept quality improvement initiatives if they are supported by scientific evidence. The QAF score in HCO A was poor because the criterion for patient discharge from the recovery room (criterion 11) was not completed because it was inconvenient (distance from operating area). Despite the poor score, the anesthetists remained adamantly against signing the discharge sheet. It was only when the chief anesthetist pointed out a published randomized trial indicating that the signature had a positive impact on quality that the staff agreed to support the improvement initiative.
Consensus: Scientific evidence needs to be supported by consensus. “Even if the corrective action seems obvious or is evidence based, collective decision-making is needed in order to avoid resistance” (Chief anesthetist, HCO A).
Work recognition: Because QAF scores are made public they are a means for the hospital and staff to gain recognition. Recognition was particularly important to practitioners working in private sector hospitals competing for clients: “If you know yourself that you are doing a good job, you want others to know as well. You’re not going to shoot yourself in the foot” (Chief anesthetist, HCO E).
The major role of information system
A major issue for QAF implementation appeared to be the use of an electronic information system. Interviewees in all 7 HCOs considered to be a positive relationship between the spread of electronic information systems and the completeness of the anesthesia files. The use of software had many advantages. For instance, it could bring file completion to a halt after an inappropriate answer; it could complete information (e.g. replace physicians’ signed initials by their full signature (HCO A) or declare peri-anesthetic incidents by default (HCO B). The IT system in HCO C included the initial consultation and the final document in the patient record to improve QAF score. However, although digitalizing may facilitate data collection, there is little incentive to make any changes if the existing system is coherent and paper sheets are easy to trace. In 4 HCOs, the anesthesia file was still a paper file. The HCO G was encountering technical difficulties but was planning on developing an electronic information system with compulsory steps (Table 3).
Imbrication of social and material constraints during QAF implementation
The improvement of QAF scores is not a mechanical result of a new technology’s implementation. In HCO E, the chief anesthesiologist had to design a flexible software that met both the objectives of QAF and the specific needs and values of the anesthesia team. “Physicians will agree to changes in their routines as long as they do not lose time, I update the software constantly”. Adjustments are made mutually between human actions and technologies. Changes to local practice always require stakeholder involvement: “expert systems can produce such actions only to the extent that the people intended to use them actually do so” . Our observations also revealed that a number of factors relating to acceptance of improvement initiatives were associated with the social environment. According to the department head of HCO D, it was only when the more senior team members left and two teams were combined that improvement action were able to be implemented in his department. In HCO G, the implementation of a new technogoly was also in opposition to the values of some senior anesthetists. They refused the intrusion of a computer screen during their pre-anesthesia consultations, leaving the onus of subsequent data entry from written notes to others, increasing the likelihood of errors. Moreover, opposition to the new technology was exacerbated by high stress levels due to the departure of the chief anesthetist and ensuing high staff turnover (physicians and nurses). The poor QAF scores in this HCO did not in fact mean, as was initially thought, that clinical managers had not made attempts to undertake improvement initiatives to meet QAF criteria. In different social environments (HCOs A and E), new QAF software for the electronic information system was adapted to user needs before being taken onboard. The recently nominated department head in HCO E did not immediately introduce a compulsory all-electronic anesthesia file but focused first on a single priority step (signatures for discharge from the recovery room) in order not to antagonize the senior anesthetists. Full digitalization was only introduced later.
Overall, our study revealed that factors related to changes in the electronic information systems and social environment played a role during QAF implementation that were at least as crucial as any action undertaken by the clinical managers. The interaction between these 3 factors – clinical manager involvement, material and social factors – were shown to influence QAF impact on the ground (Fig. 1).
Our study highlights three related findings. First, it has confirmed that a significant proportion of potential QI users, whether anesthetists or other health professionals, are often unaware of quality data [27, 28]. The engagement of team members in improvement actions was provided by the activity of clinical managers (senior anesthetists or heads of department who have professional legitimacy among anesthetists) and was often unnoticed by staff. Studies in the field of implementation sciences often highlight the importance of local leadership involvement in the implementation of quality improvements [29, 30]. Our pragmatic perspective allowed us to specify the nature of this local leadership involvement. The clinical managers have a boundary spanning role, consisting in sharing knowledge across intra-organizational and extra-organizational boundaries [31–34]. They take on the responsibility of a challenging reframing process  by translating public policies in terms of local practice. They act as mediators aware of the potentialities of QAF and push the QAF cursor into the realm of clinical practice. Third, our study has emphasized the benefit of considering local managerial activity in terms of the emerging socio-material approach in management science [36, 37, 39]. There was strong overlap between three socio-material constraints, namely, local mediator involvement, coherence of the information system for data collection, and the quality of the interpersonal ties among staff members. Staff climate, the role of information systems, and the potential added value of an electronic information system are each established key factors in quality improvement and safety issues [38, 39]. We have established that local managerial efficacy and QAF implementation requires positive interactions between all three of these factors (see high QAF scores recorded for HCOs A and E). Initiatives undertaken by clinical managers have little chance of success if they come up against unfavorable social environmental factors (see HCO G) .
Implications for policy and practice
Our findings have at least three implications in terms of public management. First, public policies tend to focus on the metrological reliability of QIs as an attribute of the quality mesure , on QI relevance to health professionals, and on ranking methods, but tend to overlook local QI implementation. Second, health authorities have chosen to promote local QI implementation by propagating the principles of QI development and the QI scores achieved. However, a public policy is not effective just because it is made known and explained widely. We have clearly shown that widespread knowledge of a QI is not a prerequisite to its implementation. A more effective communication strategy might be for health authorities to target clinical managers who have the legitimacy and clout to put forward improvement initiatives adapted to the local context. Third, local implementation requires the involvement of many professionals. Health authorities give little recognition to the role of clinical managers as mediators both inside and outside the HCO and do not provide them with adequate resources to encourage improvement initiatives. This issue could be addressed by including their role as mediators in their professional development appraisal and by paying for improved quality.
Our study has the limitations of a qualitative study on a single QI. One limitation was a risk of a social desirability bias because the professionals’ relationship to the QI and its appropriation were discussed during interviews in the anesthesia units . This risk was lessened by systematically cross-referencing the interview data with observations, written documents, official QAF scores as well as data from interviews with staff occupying other posts within the unit. Such cross-referencing showed that high scores were not due to stakeholders being highly committed to quality issues. Staff tended to be honest in their recall of QI use. The scope of our findings might be limited by our choice of a single discipline that, moreover, has long been highly alert to quality issues . According to Hamblin’s typology, the positive conduct of anesthetists is more of the ‘saints” and “honest triers” type than the “reactive gamers” and “rational maniacs” type . Studies in disciplines where the link between QIs and stakeholders is more complex, which relate to more than one discipline, and which involve professionals less well versed in quality issues are needed.
This study highlights the strong overlap between three key factors (coherence of information systems, fine-tuning of the QI to clinical practice by clinical managers, and the socio-material context) in the implementation of a national QI. In light of our findings, public policies for national QIs should consider not only metrological QI validation but also the context of QI implementation and the crucial mediatory role of clinical managers.
Epstein A. Performance reports on quality—prototypes, problems and prospects. N Engl J Med. 1995;333:57–61.
Epstein A. Rolling down the runway; the challenges ahead for quality report cards. JAMA. 1998;279:1691–6.
Marshall MN, Shekelle PG, Davies HTO, Smith PC. Public reporting on quality in the United States and the United Kingdom. Health Aff. 2003;22–3:134–48.
Chassin MR, Loeb J, Schmaltz SP, Wachter RM. Measures — using measurement to promote quality improvement. N Engl J M. 2010;363–7:683–8.
Berwick DM, James B, Coye MJ. Connections between quality measurement and improvement. Med Care. 2003;41(Suppl1):130–8.
Pflueger D. Accounting for quality: on the relationship between accounting and quality improvement in healthcare. BMC Health Serv Res. 2015;15–1:178.
Fourcade A, Blache JL, Grenier C, Bourgain JL, Minvielle E. Barriers to staff adoption of a surgical safety checklist. BMJ Qual Saf. 2012;21–3:191–7.
Bevan G, Hood C. What’s measured is what matters: targets and gaming in the English public health care system. Public Adm. 2006;84–3:517–38.
Lester HE, Hannon KL, Campbell SM. Identifying unintended consequences of quality indicators: a qualitative study. BMJ Qual Saf. 2011;20:1057–61.
Contandriopoulos D, Champagne F, Denis JL. The multiple causal pathways between performance measures’use and effect. Med Care Res Rev. 2014;71–1:3–20.
Cacace M, Ettelt S, Brereton L, Pedersen J, Nolte E. How health systems make available information on service providers: experience in seven countries. Santa Monica: RAND Europe and London School of Hygiene and Tropical Medicine; 2011.
Shortell SM, O’Brien JL, Carman JM, Foster RW, Hughes EF, Boerstir H, O’Connor EJ. Assessing the impact of continuous quality improvement/total quality management: concept versus implementation. Health Serv Res. 1995;30–2:377–401.
Kaplan HC, Brady PW, Dritz MC, et al. The influence of context on quality improvement success in health care: a systematic review of the literature. Milbank Q. 2010;88:500–59.
Conry MC, Humphries N, et al. A ten year (2000–2010) systematic review of interventions to improve quality of care in hospitals. BMC Health Serv Res. 2012;12:275. doi:10.1186/1472-6963-12-275.
Secanell M, Groene O et al. Deepening our understanding of quality improvement in Europe (DUQuE): overview of a study of hospital quality management in seven countries. Int J Qual Health Care. 2014; http://dx.doi.org/10.1093/intqhc/mzu025
Kirk MA, Kelley C, Yankey N, Birken SA, Abadie B, Damschroder L. A systematic review of the use of the consolidated framework for implementation research. Implement Sci. 2016;11–1:72.
Engeström Y, Miettinen R, Punamäki RL. Perspectives on activity theory. Cambridge: Cambridge University Press; 1999.
Cabridain M. Managerial procedures and hospital practices: a case study of the development of a new medical discipline. Soc Sci Med. 1985;2–2:167–72.
Wong A. From the front lines: a qualitative study of anesthesiologists’ work and professional values. Can J Anaesth. 2005;58–1:108–17.
Hindmarsh J, Pilnick A. The tacit order of teamwork: collaboration and embodied conduct in anesthesia. Sociol Quart. 2002;43–2:139–64.
Mort M, Goodwin D, Smith AF, Pope C. Safe asleep? Human–machine relations in medical practice. Soc Sci Med. 2005;2027–2037:61–9.
Haute Autorité de santé (French National authority for Health). http://www.has-sante.fr/portail/jcms/c_970481/fr/ipaqss-2014-a-2016-recueils-des-indicateurs-de-qualite-et-de-securite-des-soins (Accessed 1 Dec 2015).
Couralet M, Leleu H, Capuano F, Marcotte L, Nitenberg G, Sicotte C, Minvielle E. Method for developing national quality indicators based on manual data extraction from medical records. BMJ Qual Saf. 2013;22–2:155–62.
Flyvberg B. Five misunderstandings about case study research. Qualitative Inquiry. 2006;12–2:219–45.
Østerlund C. Mapping medical work: documenting practices across multiple medical settings. J Cent Inf Stud. 2004;5–3:35–43.
Eisenhardt K. Building theories from case study research. Acad Manag Rev. 1989;14–4:532–50.
Greenhalgh T, Stones R, Swinglehurst D. Choose and book: a sociological analysis of ‘resistance’ to an expert system. Soc Sci Med. 2014;104:210–9.
Ginsburg LS. Factors that influence line managers’ perception of hospital performance data. Health Serv Res. 2003;38:261–86.
Hafner JM, Williams SC, Koss RG, Tshurtz BA, Schmaltz SP, Loebe JM. The perceived impact of public reporting hospital performance data: Interviews with hospital staff. International J Qual Health Care. 2011;23:697–704.
Guerrero EG, Padwa H, Fenwick K, Harris LM, Aarons GA. Identifying and ranking implicit leadership strategies to promote evidence-based practice implementation in addiction health services. Implement Sci. 2016;11–1:69.
Nadeem E, Olin SS, Hill LC, Hoagwood KE, Horwitz SM. Understanding the components of quality improvement collaboratives: a systematic literature review. Milbank Q. 2013;91–2:354–94.
Carman JM, Shortell SM, Foster RW, Hughes EFX, Boerstler H, O’brien JL, O’Connor EJ. Keys for successful implementation of total quality management in hospitals. Health Care Man Rev. 2010;35–4:283–93.
Bohmer RMJ. Leading clinicians and clinicians leading. N Eng J Med. 2013;368–16:1468–70.
Mc Fadden KL, Stock GN, Gowen CR. Leadership, safety climate and continuous quality improvement: impact on process quality and patient safety. Health Care Man Rev. 2015;40–1:24–34.
Kellogg KC, Orlikowski WJ, Yates J. Life in the trading zone: structuring coordination across boundaries in postbureaucratic organizations. Org Sci. 2006;17–1:22–44.
Boxenbaum E. Lost in translation, the making of Danish diversity management. Am Behav Scient. 2006;49–7:939–48.
Orlikowski W, Scott SV. Sociomateriality: challenging the separation of technology, work and organization. Acad Manag Annals. 2008;2–1:433–74.
Leonardi PM, Nardi BA, Kallinikos J, editors. Materiality and organizing: social interaction in a technological world. London: Oxford University Press; 2012.
Bates DW, Gawande AA. Improving safety with information technology. N Engl J Med. 2003;348:2526–34.
Benzer JK, Young G, Stolzmann K, et al. The relationship between organizational climate and quality of chronic disease management. Health Serv Res. 2011;46–3:691–711.
Allen D. From boundary concept to boundary object: the practice and politics of care pathway development. Soc Sci Med. 2009;69–3:354–61.
Tutorials on Quality Measures: Desirable Attributes of a Quality Measure. In: National Quality Measures Clearinghouse (NQMC) [Web site]. Rockville: Agency for Healthcare Research and Quality (AHRQ); [cited 2016 Jun 1]. Available: http://www.qualitymeasures.ahrq.gov.
Cohen MM, Duncan PG, Pope WD, Biehl D, Tweed WA§Merchant RN. The canadian four-center study of anaesthetic outcomes: can outcomes be used to assess the quality of aneasthesia care? Canadian Journal of Anaesthesia. 1992;39(5):430–439.
Haller G, Stoelwinder J, Myles PS, Mc NJ. Quality and safety indicators in anesthesia: a systematic review. Anesthesiology. 2009;110–5:1158–75.
Durand-Zaleski I, Alberti C, Durieux P, et al. Informed consent in clinical research in France: assessment and factors associated with therapeutic misconception. J Med Ethics. 2008;34(9):e16.
The authors wish to thank the staff in all the participating anesthesia departments and the president (Professor Dan Benhamou) and members of SFAR.
The study was supported by a grant (PRIQS-ANR Blanc 2010) from the French National Agency for Research (ANR).
MW and EM were involved in the conception, design, and analysis of data. MW, MLG, and AZ were involved in data collection. MW, EM, CS, PL, MLG, AZ were involved in drafting the article and revising it critically for important intellectual content, and approved the final version to be published. EM is the guarantor.
EM was commissioned by the French national Authority for Health (HAS) to provide metrological evidence on national quality indicators, including QAF, and for this reason did not participate in data collection.
Consent for publication
Ethics approval and consent to participate
The Design of the research was ethically approved by the French National Agency for Research (ANR- 2010-BLAN-1820-01). All participants provided oral consent since in France this type of study does not require formal informed consent of participants (i.e. clinical research does require informed patient consent, but not if the study employs the use of surveys in an administrative context ). Administrative review and clearance for staff interviews was obtained before data collection in each hospital. Lastly, the researchers were particularly sensitive to issues of confidentiality, during and after interviews.
Data sharing statement
Data is available upon request from the author.