Skip to main content
  • Research article
  • Open access
  • Published:

Unpacking organizational readiness for change: an updated systematic review and content analysis of assessments



Organizational readiness assessments have a history of being developed as important support tools for successful implementation. However, it remains unclear how best to operationalize readiness across varied projects or settings. We conducted a synthesis and content analysis of published readiness instruments to compare how investigators have operationalized the concept of organizational readiness for change.


We identified readiness assessments using a systematic review and update search. We mapped individual assessment items to the Consolidated Framework for Implementation Research (CFIR), which identifies five domains affecting implementation (outer setting, inner setting, intervention characteristics, characteristics of individuals, and implementation process) and multiple constructs within each domain.


Of 1370 survey items, 897 (68%) mapped to the CFIR domain of inner setting, most commonly related to constructs of readiness for implementation (n = 220); networks and communication (n = 207); implementation climate (n = 204); structural characteristics (n = 139); and culture (n = 93). Two hundred forty-two items (18%) mapped to characteristics of individuals (mainly other personal attributes [n = 157] and self-efficacy [n = 52]); 80 (6%) mapped to outer setting; 51 (4%) mapped to implementation process; 40 (3%) mapped to intervention characteristics; and 60 (4%) did not map to CFIR constructs. Instruments were typically tailored to specific interventions or contexts.


Available readiness instruments predominantly focus on contextual factors within the organization and characteristics of individuals, but the specificity of most assessment items suggests a need to tailor items to the specific scenario in which an assessment is fielded. Readiness assessments must bridge the gap between measuring a theoretical construct and factors of importance to a particular implementation.

Peer Review reports


The rapid growth of multi-disciplinary fields, including implementation science, brings along with it the propagation of more terminology [1, 2]. While some of these terms may represent unique ideas, there are also many examples of the Jingle and Jangle Fallacies [3, 4]. The Jingle Fallacy, also known as synonymy, occurs when multiple names are used to refer to the same concept or thing (e.g., practice facilitation and coaching). Conversely, the Jangle Fallacy, or polysemy, occurs when the same name is used for different concepts or things. For instance, a “practice” in healthcare could refer to a medical organization (e.g., there are three doctors at this practice) or a strategy or process (e.g., a care management practice to manage chronic illness).

The seemingly self-explanatory concept of “organizational readiness for change” actually falls prey to both the Jingle and Jangle Fallacies. In the former case, we do not yet have good distinctions between assessing “organizational readiness for change,” “needs,” “barriers and facilitators,” or “factors affecting implementation” [5]. An earlier systematic review on organizational readiness for change found that relevant literature, in addition to discussing “readiness”, used terms like “preparedness”, “willingness”, “commitment” and “acceptance” [6].

The Jangle Fallacy also applies in that “organizational readiness for change” has been defined and measured in different ways. Some definitions and measures focus on the characteristics of individuals within an organization, as demonstrated by this definition from Weiner and colleagues: “the extent to which organizational members are psychologically and behaviorally prepared to implement organizational change” [7]. Others focus on macro-level factors, such as collective commitment or collective efficacy, and define organizational readiness for change as “a comprehensive attitude” that incorporates factors at an organizational level [8].

In the absence of a consensus on a conceptual framework for organizational readiness for change, knowing what needs to be included in such an assessment may remain a challenge [9]. Theorists in implementation science have an interest in refining and standardizing the measurement of organizational readiness for change to improve conceptual clarity, comparison across sites and studies, and predictive validity. In practice, however, using an existing measure may be challenging. Some assessments are developed with a particular setting or intervention in mind [6], for example, specific to addiction treatment [10], or describing transitions related to a hospital relocation [11] which can makes them less generalizable. On the other hand, broader assessments, in their attempts to be inclusive, may be lengthy or imprecise and thus require adaptation to meet the needs of a given context.

Our work began as part of the US Department of Veterans Affairs Health Services Research and Development (HSR&D) Care Coordination Quality Enhancement Research Initiative (QUERI) program. One of our aims was to use readiness assessments across three different projects to improve care coordination in VA and compare their predictive validity regarding implementation outcomes. We began by searching for existing assessments and discovered that a team at St. Michael’s Hospital in Toronto had created the Ready, Set, Change! decision support tool to help researchers identify existing assessments that would be best suited for their studies [12]. The Ready, Set, Change! team included assessments from a 2014 systematic review [6] that met pre-determined criteria for validity and reliability. The recommended assessments from the decision support tool, however, were not suitable for our needs without adaptation, due to their length and lack of relevance to our specific context and intervention details.

In response to this experience, we set out to review existing measures of organizational readiness for change to see how others had operationalized the concept. We then engaged in content analysis to identify core concepts, mapping them to the Consolidated Framework for Implementation Research (CFIR) [13]. CFIR provides a broad range of constructs relevant to implementation research and allowed for comprehensive description and comparison of the explicit and implicit definitions and frameworks underlying identified readiness assessments. Because we anticipated a range of organizational readiness definitions and measurement approaches, we chose CFIR as a broad framework that would likely capture the various permutations organizational readiness assessments were likely to take, even when they did not overlap with each other or any one organizational readiness for change framework. In so building on prior work [6, 7, 12, 14], our objective is pragmatic: to support developers of readiness assessments in determining key topics they may want to keep in mind when tailoring or developing an assessment outside the purview of existing assessments.


Our approach involved multiple steps. First, we used systematic review methods to update the database searches conducted by a prior review of organizational readiness for change assessments to identify any additional relevant assessments. Then, we built an item bank composed of individual items included in the readiness assessments identified. Finally, we used directed content analysis to sort items into categories using CFIR as our initial foundation [13]. This systematic review is reported according to Preferred Reporting Items for Systematic Reviews and Meta-Analyses (PRISMA) guidelines, see Additional file 1 for PRISMA checklist [15].

Literature search

We built upon the literature search conducted by Gagnon and colleagues as part of their 2014 systematic review of organizational readiness instruments [6]. Because this review of organizational readiness assessments used a search conducted in 2012, we updated the search through June 14, 2017. This broad search was based on terms related to readiness, change, and health or social services within six databases: Web of Science, Sociological Abstracts, PubMed, PsycINFO, Embase, and CINAHL (see Additional file 2 for full search strategy). We found additional studies by mining identified literature for relevant references, as well as by expert suggestion.

Study selection

Two team members (IML, DMD) independently screened all identified titles and abstracts in duplicate. For potentially relevant abstracts, we retrieved full-text articles and reviewed them independently in duplicate as well, with discrepancies reviewed by the full team. To be included, the actual assessment used, with a full list of individual items, needed to be available for each full-text publication. This assessment needed to be relevant to healthcare delivery settings and to measure organizational readiness for change. Because, as noted above, organizational readiness for change is a nebulous concept, the measure had to capture a general sentiment of willingness, readiness, or acceptance for an organizational or collective change or innovation (rather than personal behavior change, e.g., for smoking cessation). Multiple studies using the same assessment could be included if they represented unique data collection with separate samples of participants, since each use constituted an operationalization that could inform our research objective. By including duplications and variations, we were better able to describe the uses of each assessment, including contexts in which each assessment was used, if the assessment was altered, and whether assessments were collected alongside additional measures.

Data abstraction

We transcribed all individual questions or items from included publications into a database that served as an item bank. We captured information about each included publication, including the name of the assessment used (when reported), total number of items in that assessment or assessments, study setting, study sample, type of intervention, and any additional data collected for the study (e.g., other screeners or surveys, interviews, patient records). For items that appeared multiple times, we made separate entries in the database for each unique appearance (i.e., when one assessment was used by multiple studies in part or in whole). We did not conduct a quality assessment of the included studies, since our analysis was not focused on the validity or robustness of study findings.

Synthesis and analysis

We used directed content analysis to identify themes within the readiness assessment items in our database. Directed content analysis builds from existing theory, models, or frameworks, which can provide the initial coding structure [16]. Beginning with these predetermined codes, all data is coded to the extent possible. Analysts then identify data that cannot be captured by the existing coding structure and develop new codes, or sub-codes of existing codes, to better capture how the existing theory, model, or framework is supported and extended by the data.

Because of the conceptual fuzziness surrounding organizational readiness for change, we sought a comprehensive framework to which we could map items in the item bank, and selected CFIR, which includes five domains within which 39 constructs are nested [13]. The “intervention characteristics” domain includes eight constructs such as relative advantage and cost of the intervention. The “outer setting” domain includes four constructs for factors outside an organization (e.g., external policy and incentives). Within the “inner setting” domain are five constructs: structural characteristics, networks and communications, culture, implementation climate, and readiness for implementation. These last two constructs are also broken down into sub-constructs, with six sub-constructs nested under implementation climate and three under readiness for implementation. The fourth domain is “characteristics of individuals,” which houses five constructs. The final domain of “process” is comprised of four constructs: planning, engaging (which has sub-constructs for four different groups of individuals who may be involved in the implementation), executing, and reflecting and evaluating. For a delineation of how the framework was applied in this analysis, and exemplar items from the item bank, see the codebook in Additional file 3. We iteratively developed the codebook based on the existing framework to clarify our application of the CFIR construct definitions and any modifications we made. For example, based on the CFIR definitions, we limited certain CFIR constructs to intervention-specific items (e.g., the construct “available resources” was used for project-specific resources), whereas other CFIR constructs were exclusively used for items that described general characteristics (e.g., the construct of “structural characteristics” was applied for items describing organizational resources more broadly).

Two members of the study team independently coded each item with a CFIR construct, or sub-construct where possible. All discrepancies were reconciled by these two members or the larger team when necessary. We categorized nearly all items under a CFIR construct or sub-construct. We developed one new construct-level code to capture items related to leadership qualities that were not intervention-specific. These items did not fit into the CFIR categorizations, as the existing representation of leadership within CFIR was in sub-constructs related to engagement of leadership with a specific intervention, as opposed to a more general description of an organization’s leaders. Some additional items that were project-specific were excluded from coding (e.g., “12-step theory (AA/NA) is followed by many of the counselors here” [17]). When more than 50 items were coded to a CFIR construct that did not have specific sub-constructs, a pile- sort methodology was used to develop new sub-constructs; this allowed us to better characterize the diversity within these large constructs.

In the case of the networks and communications construct, we used an additional model from Lanham and colleagues to classify the sub-constructs, since emerging sub-codes aligned with characteristics of work relationships that Lanham and colleagues had previously identified [18,19,20]. CFIR defines the networks and communications construct as being about relationships: “the nature and quality of webs of social networks and the nature and quality of formal and informal communications within an organization” [13]. Specifying sub-constructs using an established model for work relationships therefore had face validity.

The Lanham model was developed with a focus on relationships in healthcare delivery settings; applications of the model suggest that these relationship characteristics should be considered during improvement efforts or redesign [19, 20]. The model includes seven characteristics, of which five emerged within these data and were therefore applied: relatedness, trust, respectful interaction, heedfulness, and mindfulness. Full descriptions of these five characteristics are provided in Additional file 3. We generated additional inductive sub-constructs to capture emergent themes in the items within the networks and communications construct that fell outside the relationship model.

In coding each item, we relied on the most granular code appropriate (e.g., using subcodes where appropriate), and noted the unit of measurement: “self,” “staff,” “leadership,” or “organization.” “Organization” was the default if the unit of measurement was ambiguous. Additionally, we recorded information on whether the item referenced implementation of a specific intervention, rather than a general question about the state of the organization or individual. See Additional file 4 for the coding form. Once all items were coded, we narratively summarized our findings to describe the operationalization of organizational readiness for change within the included assessments and studies.


Literature flow

The total number of publications included in our analysis is 27, which represents 29 uses of readiness assessments. From the 29 of organizational readiness assessment uses, 1370 individual assessment items were included in the item bank. See Fig. 1 for literature flow.

Fig. 1
figure 1

Literature Flow

Description of included studies

Table 1 provides details of the published uses of readiness assessments. The first instance was in 1988 [37], with the next one a decade later [29]. Published uses of readiness assessments steadily increased from 2007 onward, with 23 of the 29 assessment uses published in 2007 or later. Readiness assessment data were supplemented with additional data collection efforts in 14 uses, which commonly took the form of interviews or other complementary instruments, such as a job satisfaction survey [22].

Table 1 Evidence Table for Included Studies

Assessments ranged in size from nine to 134 items (of note, both of these extremes were adaptations of the same original survey, discussed below), with a median of 30 items (interquartile range: 44 items). Whereas the first few assessments were larger, and large assessments with more than 50 items continue to be used, smaller assessments with fewer than 30 items, and often fewer than 20 items, began to appear in 2004. Overall these smaller assessments comprise roughly half of the assessment uses (52%, 15/29 uses).

The majority of uses represent separate assessments tailored to a unique context (62%, 18/29 uses), but three assessments were used more than once. Variations of the Texas Christian University Organizational Readiness for Change Treatment assessment were used six times. Originally developed for use in addiction treatment settings, this assessment includes both a director and a staff version [10]. All of the additional uses adapted or used supplemental data collection, and all but one use (in a child welfare setting) occurred in mental health/substance use settings. The second assessment to be used more than once was developed by Holt and colleagues, and was developed to better understand the use of information systems in a variety of organizations [8]. The second use of this assessment used the same questions in a primary care setting related to eHealth tool deployment, in addition to a supplemental questionnaire [24]. Finally, the Evidence-Based Practices Beliefs scale was used in three hospital-based settings, the first time when it was being developed [25] and later in two samples of hospital-based nurses as part of a battery of assessments [26, 27].

The 18 individually developed assessments were fielded in a variety of settings, including many types of clinical settings: primary care, long-term care, hospitals, rehabilitation, and mental health. Two assessments were not conducted in healthcare delivery settings [8, 30], but were included by Gagnon’s earlier work and had items that were deemed flexible enough to apply in healthcare delivery settings. All but six uses described a particular intervention for which the readiness assessment was conducted.

Eleven uses employed the specific phrase “Organizational Readiness for/to Change” [8, 10, 17, 21,22,23,24, 35, 41, 43], with an additional eight uses of “Organizational Readiness for [a specific intervention]” (e.g. e-Health) [11, 26, 27, 32, 38, 39, 42]. Other variations included “Practice Capacity for Change,” [31] “Preparedness for Change,” [34] and “Readiness to Engage in EBP.” [25]

Items mapped to CFIR

Of the 1370 readiness items, 1310 were mapped to CFIR; the remaining 60 items (4%) did not match with any of the CFIR constructs, largely due to the specificity of the question related to particular programming. The majority of items mapped to the CFIR domain of inner setting (68%, n = 897) (Fig. 2). The second most heavily mapped CFIR domain was characteristics of individuals (18%, n = 242), followed by outer setting (6%, n = 80), implementation process (4%, n = 51), and intervention characteristics (3%, n = 40).

Fig. 2
figure 2

Items Mapped to CFIR

Seven CFIR constructs had 50 or more items mapped to them, together accounting for 82% of the total items. Within the domain of inner setting, items most often mapped to the CFIR constructs of readiness for implementation (n = 220), networks and communications (n = 207), implementation climate (n = 204), structural characteristics (n = 139), and culture (n = 93). Within the domain of characteristics of individuals, items most often mapped to the CFIR constructs of other personal attributes (n = 157) and self-efficacy (n = 52).

We generated one additional construct-level code to capture a unique theme outside of the CFIR structure: leadership qualities (n = 34). These items did not fit into the CFIR categorizations, as the existing representation of leadership within CFIR was in sub-constructs related to engagement of leadership with a specific intervention, as opposed to a more general description of an organization’s leaders. See Fig. 2 for the mapping of items to CFIR domains and constructs.

Assessments mapped to CFIR

Figure 3 is a heat map that displays the concentration of items within each assessment use across the various CFIR constructs and subconstructs. Twenty constructs had at least one item mapped from our bank of assessment items. The same seven CFIR constructs identified in the prior section as having most items mapped to them also occurred mostly frequently across assessment uses, however their order of popularity shifted: implementation climate was included in 25 uses; readiness for implementation was included in 22 uses; structural characteristics and personal attributes both included in 16 uses; networks and communications was included in 15 uses; culture was included in 14 uses; and self-efficacy was included in 13 uses. The number of constructs included in assessment uses ranged from one construct (readiness for implementation) in a 12-item assessment [41], to 13 constructs in a 57-item assessment [42]. The median number of constructs included in any assessment use was 6 constructs (interquartile range of 4 to 10), with median of 10 items per construct (interquartile range of 3 to 15).

Fig. 3
figure 3

Heat Map of Assessment Uses to CFIR

Top seven CFIR constructs

Highlighted here are the seven constructs that received the most attention from assessment developers and/or users, both in terms of the individual item analysis and the assessment heat map. Definitions and example items for all sub-constructs derived in the analysis are included in Additional file 3; here we focus on synthesizing findings for each construct.

Readiness for implementation

Readiness for implementation was mapped to the most individual items of any construct – 220 items – and ranked second in number of assessment uses, with 22 uses. Defined as “tangible and immediate indicators of organizational change,” readiness for implementation includes sub-constructs for leadership engagement, available resources, and access to knowledge and information [13]. CFIR defines this construct and its sub-constructs as specific to an intervention’s implementation, rather than describing the organization more generally. We identified 97 items as related to the subconstruct of available resources. The leadership engagement sub-construct was represented in 46 items, while access to knowledge and information about the intervention was represented by 13 items. The other 64 items in the readiness for implementation construct were judged to be “immediate indicators of organizational change” that fell outside available resources, leadership engagement, or access to knowledge and information [13]. These included items like “people who work here want to implement this change” [41], related to buy-in from staff members other than leadership, as well as items that described the meeting of pre-conditions for implementation (e.g., “how confident are you that most physicians can use e-prescribing instead of handwritten or printed prescriptions?” [40]).

Implementation climate

Implementation climate is defined as “the absorptive capacity for change, shared receptivity of involved individuals to an intervention and the extent to which use of that intervention will be rewarded, supported, and expected within their organization” [13]; in keeping with this definition, this set of codes was also specific to an intervention. This construct ranked first in assessment uses with 25 uses and third in total number of mapped items, with 204 items. Two main sub-constructs absorbed most of the items: compatibility (n = 97) and tension for change (n = 88). Compatibility of the intervention could be with the organization broadly, leadership, a workgroup or team, or the respondent’s beliefs or job. Items in the tension for change sub-construct took the form of needs assessments (e.g., “my unit needs guidance in developing services to address alcohol and drug behaviors presented by our patients” [21]), or descriptions of pressures for changes. The final 19 items in implementation climate were in the sub-constructs of relative priority (n = 7), organizational incentives and rewards (n = 6), goals and feedback (n = 4), and other (n = 2).

Other personal attributes

With 157 items, the other personal attributes construct ranked fourth in individual item mapping, and had 16 assessment uses. This broad construct within the characteristics of individuals domain served as catch-all for non-specific items about the respondent (e.g., self-efficacy or knowledge and beliefs about the intervention). Our inductive sorting produced seven sub-constructs, including a respondent’s engagement with or aptitude for learning new skills or job-related content (n = 51), descriptive information like identifying your position or department within the organization (n = 31), and whether respondents saw themselves as a self-described leader (n = 22).

Structural characteristics

Tied for third highest uses with other personal attributes and featured in slightly fewer individual items (n = 139), structural characteristics included five sub-constructs detailing various aspects of an organization’s workforce, physical and information technology infrastructure, and operational structure. These items were also not specific to any particular intervention.

Networks and communications

The networks and communications construct was also non-intervention specific, ranked second highest in individual items (n = 207), and had 15 assessment uses. We identified five sub-constructs related to relationships using definitions provided by the Lanham and colleagues model [18].

Mindfulness, which is defined as “openness to new ideas and different perspectives, fully engaged presence, rich discriminating awareness, or seeking novelty (even in routine situations)” [18], was represented by 91 items like “People in this team are always searching for fresh, new ways of looking at problems” [29]. Heedfulness is described as interactions “where individuals are sensitive to the task at hand (the job they are doing) and are paying attention to the way their roles and actions fit into (affect) the roles and actions of the entire group” [18]. “Everyone knows how their work will affect the work of the next person or the quality of the final product or service” [37] was an exemplar of items related to this sub-construct (n = 34). Respectful interaction items (n = 31), which represent “honest, self-confident, and appreciative interaction among individuals; often creating new meaning” [18], included this example: “Different parts of the organization work together well; when conflict arises, it is often productive” [37]. The Trust sub-construct, or the “willingness of an individual to be vulnerable to another individual” [18], contained 16 items, like “to what extent do you feel at ease with the members of your team?” [29]. The final sub-construct from the relationships model by Lanham and colleagues represented the idea of relatedness, which is “characterized by work- or non-work-related conversations and activities” [18]. Items in the relatedness sub-construct (n = 15) described communications like “staff share common goals about the care of residents at the beginning of and throughout each shift” [10].

In addition to the sub-constructs defined by the relationships model, two inductive sub-constructs were developed that captured communication activities that did not delve into underlying relationship traits: organization-level communications (n = 16), and cross-departmental communication (n = 4).


The construct for self-efficacy within the characteristics of individuals domain included 52 items; because they were consistent and coherent as a group and did not warrant subdivision, no further sub-constructs were developed.


The culture construct, which generally describes “norms, values, and basic assumptions of a given organization” [13] contained 93 items that were never specific to a particular implementation and described shared characteristics about the group/organization as a whole. This construct fell within the inner setting domain.


Our examination of organizational readiness for change assessments identified both significant variation and important commonalities in how scale developers operationalize this topic. Originally we had hoped to generate a master organizational readiness assessment with modular elements, in order to draw upon this assessment for our different projects. In reality, we found that the existing assessments were so tailored to the specific study, intervention, or setting that this was not possible. No gold standard exists within the realm of organizational readiness for change assessments; every use we identified was tailored to some degree, whether through modification or elimination of items from an existing assessment, supplemental data collection, or the building of an assessment de novo. And while the definition of organizational readiness for change can be either intervention-specific or more general, [7, 14, 44, 45] in developing measures authors chose to be intervention-specific for 23 of the 29 assessment uses we identified.

Use of the CFIR to compare content across assessments revealed several constructs that emerge frequently in readiness assessments, particularly implementation climate and readiness for implementation. These seem like core constructs, given their consistency in the data and conceptual overlap with the various definitions of organizational readiness for change. These constructs were never the only ones represented, however; every assessment also included a unique constellation of items across the other constructs and sub-constructs of CFIR. Structural characteristics, other personal attributes, and networks and communications constructs also appeared frequently, indicating the important role played by an organization’s structure, its people, and the connections between them in influencing readiness.

Using a broad framework like CFIR, rather than any specific organizational readiness for change framework, proved helpful because we were able to capture the full range of contextual information assessments were designed to gather, even when they did not directly overlap with one another. There were only minor adjustments needed to classify items from the assessments using CFIR, which suggests that organizational readiness for change is captured within the large framework of constructs “associated with effective implementation” that CFIR offers. The one notable area where we made additions to CFIR relates to teams. CFIR has domains for individual level (micro) and organizational level (macro) constructs, but no domain specific to an intermediate (meso) level, like a team. The prevalence of items we coded with the “team” unit of analysis suggests that this level may be distinct, but this distinction was somewhat obscured between the domain of inner setting, which often describes much more than one team, and the domain of characteristics of individuals, which is more granular. For instance, the wording of items like the following illustrate a team orientation: “The implementation team members have staff support and other resources required for the project” [35]. Capturing this team (meso) level as distinct from the macro level may be helpful in more clearly distinguishing the role of teams in readiness for change.


Several limitations may affect the interpretation of the presented findings. Potential publication bias must be acknowledged as an issue, but given that we were not focused on a particular outcome, we could not use standard statistical methods, like a funnel plot, for detecting potential bias. It is also likely that organizational readiness for change assessments, or assessments with the same purpose, are used either formally or informally in practice without being published. No unpublished assessments were included in this work, nor did we search gray literature for assessments, and as such, our conclusions may not be applicable to these additional assessments.

In addition, breaking down validated instruments into individual items can be problematic from the perspective of psychometric properties. However, our aim was not to produce valid instruments for use, but rather to conduct qualitative analysis, the aims of which are to describe themes and better understand commonalities and differences between assessments, rather than to test construct validity.

Finally, some items appear multiple times within our data set, which may impact our assessment of item and construct frequencies. In four cases these items came from identical assessments, which were tested in different populations. In addition, some adaptations included subsets of items from the original survey. Each use of a readiness assessment, regardless of duplication, was included to get a better understanding of our main objective, which was to describe the operationalization of organizational readiness for change as it appeared in the literature.


A better understanding of organizational readiness for change may require the streamlining of terminology, but it is also important to describe why these assessments are done in the first place, and what developers felt was important to assess. While using a broad framework to identify commonalities suggests prevalent themes throughout the assessments, it is important we also understand how these constructs can be used to predict desired outcomes or serve as a diagnostic for tailoring implementation approaches. In addition, analysis of these assessments revealed an implicit expectation that team relationships are a critical component of readiness. There was no single construct within CFIR to capture this sentiment, but we found that Lanham and colleagues’ model for characteristics of work relationships was highly compatible with the content of relationship items. Future work more closely examining the role of relationships in organizational readiness may be required to fully explicate their impact.


The readiness assessments reviewed in this article revealed significant commonalities; however, the specificity of many items suggests most assessments will need to be customized or tailored prior to use. The continued proliferation of new assessments, meanwhile, signals that there is no current gold standard assessment for organizational readiness for change. Consensus around a definition of organizational readiness for change may allow future developers to focus on a more parsimonious and better-specified set of constructs. Work testing the relationship between organizational readiness for change and implementation outcomes [46] will help to better specify the underlying mechanisms of readiness and may lead to more adaptable assessments. In the meantime, readiness assessments must often bridge the gap between measuring a theoretical construct and evaluating factors specific to a particular implementation.

Availability of data and materials

The datasets used and/or analyzed during the current study are available from the corresponding author on reasonable request. All literature included in the was obtained from publicly available sources.



Consolidated framework for implementation research


Health services research and development


Preferred reporting items for systematic reviews and meta-analyses


Quality enhancement research initiative


Department of Veterans Affairs


  1. Larsen KR, Voronovich ZA, Cook PF, Pedro LW. Addicted to constructs: science in reverse? Addiction. 2013;108(9):1532–3.

    Article  PubMed  Google Scholar 

  2. Greenhalgh T, Robert G, Macfarlane F, Bate P, Kyriakidou O. Diffusion of innovations in service organizations: systematic review and recommendations. Milbank Q. 2004;82(4):581–629.

    Article  PubMed  PubMed Central  Google Scholar 

  3. Thorndike EL. An introduction to the theory of mental and social measurements. New York: Teacher's college, Columbia University; 1913.

  4. Kelley TL. Interpretation of educational measurements; 1927.

    Google Scholar 

  5. Chaudoir SR, Dugan AG, Barr CH. Measuring factors affecting implementation of health innovations: a systematic review of structural, organizational, provider, patient, and innovation level measures. Implement Sci. 2013;8(1):22.

    Article  PubMed  PubMed Central  Google Scholar 

  6. Gagnon M-P, Attieh R, Ghandour EK, Légaré F, Ouimet M, Estabrooks CA, et al. A systematic review of instruments to assess organizational readiness for knowledge translation in health care. PLoS One. 2014;9(12):e114338.

    Article  PubMed  PubMed Central  CAS  Google Scholar 

  7. Weiner BJ. A theory of organizational readiness for change. Implement Sci. 2009;4(1):67.

    Article  PubMed  PubMed Central  Google Scholar 

  8. Holt DRAA, Feild HS, Harris SG. Readiness for organizational change: the systematic development of a scale. J Appl Behav Sci. 2007;43(2):232–55.

    Article  Google Scholar 

  9. Rabin BA, Lewis CC, Norton WE, Neta G, Chambers D, Tobin JN, et al. Measurement resources for dissemination and implementation research in health. Implement Sci. 2015;11(1):42.

    Article  Google Scholar 

  10. Lehman WE, Greener JM, Simpson DD. Assessing organizational readiness for change. J Subst Abus Treat. 2002;22(4):197–209.

    Article  Google Scholar 

  11. Gray CS, Wilkinson A, Alvaro C, Wilkinson K, Harvey M. Building resilience and organizational readiness during healthcare facility redevelopment transitions: is it possible to thrive? Herd-Health Env Res Des J. 2015;9(1):10–33.

    Google Scholar 

  12. Timmings C, Khan S, Moore JE, Marquez C, Pyka K, Straus SE. Ready, set, change! Development and usability testing of an online readiness for change decision support tool for healthcare organizations. BMC Med Inform Decis Mak. 2016;16(1):24.

    Article  PubMed  PubMed Central  Google Scholar 

  13. Damschroder LJ, Aron DC, Keith RE, Kirsh SR, Alexander JA, Lowery JC. Fostering implementation of health services research findings into practice: a consolidated framework for advancing implementation science. Implement Sci. 2009;4(1):50.

    Article  PubMed  PubMed Central  Google Scholar 

  14. Holt DT, Armenakis AA, Harris SG, Feild HS. Toward a comprehensive definition of readiness for change: A review of research and instrumentation. Res Organ Change Dev. 2007;16(2007):289–336.

    Article  Google Scholar 

  15. Moher D, Shamseer L, Clarke M, Ghersi D, Liberati A, Petticrew M, Shekelle P, Stewart LA. Preferred reporting items for systematic review and meta-analysis protocols (PRISMA-P) 2015 statement. Syst Rev. 2015;4(1):1.

  16. Hsieh H-F, Shannon SE. Three approaches to qualitative content analysis. Qual Health Res. 2005;15(9):1277–88.

    Article  PubMed  Google Scholar 

  17. Guerrero EG, Aarons GA, Grella CE, Garner BR, Cook B, Vega WA. Program capacity to eliminate outcome disparities in addiction health services. Administration and policy in mental health and mental health services research S2- administration and policy in mental health S2. Adm Ment Health. 2016;43(1):23–35.

    Article  Google Scholar 

  18. Lanham HJ, McDaniel RR, Crabtree BF, Miller WL, Stange KC, Tallia AF, et al. How improving practice relationships among clinicians and nonclinicians can improve quality in primary care. Jt Comm J Qual Patient Saf. 2009;35(9):457–AP2.

    Article  PubMed  PubMed Central  Google Scholar 

  19. Lanham HJ, Palmer RF, Leykum LK, McDaniel RR, Nutting PA, Stange KC, et al. Trust and reflection in primary care practice redesign. Health Serv Res. 2016;51(4):1489–514.

    Article  PubMed  Google Scholar 

  20. Finley EP, Pugh JA, Lanham HJ, Leykum LK, Cornell J, Veerapaneni P, et al. Relationship quality and patient-assessed quality of care in VA primary care clinics: development and validation of the work relationships scale. Ann Fam Med. 2013;11(6):543–9.

    Article  PubMed  PubMed Central  Google Scholar 

  21. Bohman TM, Kulkarni S, Waters V, Spence RT, Murphy-Smith M, McQueen K. Assessing health care organizations’ ability to implement screening, brief intervention, and referral to treatment. J Addict Med. 2008;2(3):151–7.

    Article  PubMed  Google Scholar 

  22. Claiborne N, Auerbach C, Lawrence C, Schudrich WZ. Organizational change: the role of climate and job satisfaction in child welfare workers’ perception of readiness for change. Child Youth Serv Rev. 2013;35(12):2013–9.

    Article  Google Scholar 

  23. Saldana L, Chapman JE, Henggeler SW, Rowland MD. The organizational readiness for change scale in adolescent programs: criterion validity. J Subst Abus Treat. 2007;33(2):159–69.

    Article  Google Scholar 

  24. Saleh S, Khodor R, Alameddine M, Baroud M. Readiness of healthcare providers for eHealth: the case from primary healthcare centers in Lebanon. BMC Health Serv Res. 2016;16(1):644.

  25. Melnyk BM, Fineout-Overholt E, Mays MZ. The evidence-based practice beliefs and implementation scales: psychometric properties of two new instruments. Worldviews Evid-Based Nurs. 2008;5(4):208–16.

    Article  PubMed  Google Scholar 

  26. Breckenridge-Sproat ST, Throop MD, Raju D, Murphy DA, Loan LA, Patrician PA. Building a unit-level mentored program to sustain a culture of inquiry for evidence-based practice. Clin Nurse Spec. 2015;29(6):329–37.

    Article  PubMed  Google Scholar 

  27. Warren JI, Montgomery KL, Friedmann E. Three-year pre–post analysis of EBP integration in a magnet-designated community hospital. Worldviews Evid-Based Nurs. 2016;13(1):50–8.

    Article  PubMed  Google Scholar 

  28. Aarons GA. Mental health provider attitudes toward adoption of evidence-based practice: the evidence-based practice attitude scale (EBPAS). Ment Health Serv Res. 2004;6(2):61–74.

    Article  PubMed  PubMed Central  Google Scholar 

  29. Anderson NR, West MA. Measuring climate for work group innovation: development and validation of the team climate inventory. J Organ Behav. 1998;19(3):235–58.

    Article  Google Scholar 

  30. Armenakis AA, Bernerth JB, Pitts JP, Walker HJ. Organizational change recipients’ beliefs scale: development of an assessment instrument. J Appl Behav Sci. 2007;43(4):481–505.

    Article  Google Scholar 

  31. Bobiak SN, Zyzanski SJ, Ruhe MC, Carter CA, Ragan B, Flocke SA, et al. Measuring practice capacity for change: a tool for guiding quality improvement in primary care settings. Qual Manag Health Care. 2009;18(4):278–84.

    Article  PubMed  Google Scholar 

  32. Cherry B. Assessing organizational readiness for electronic health record adoption in long-term care facilities. J Gerontol Nurs. 2011;37(10):14–9.

    Article  PubMed  Google Scholar 

  33. Demiris G, Courtney KL, Meyer W. Current status and perceived needs of information technology in critical access hospitals: a survey study. Inform Prim Care. 2007;15(1):45–51.

    PubMed  Google Scholar 

  34. Gibb H. An environmental scan of an aged care workplace using the PARiHS model: assessing preparedness for change. J Nurs Manag. 2013;21(2):293–303.

    Article  PubMed  Google Scholar 

  35. Helfrich CD, Li YF, Sharp ND, Sales AE. Organizational readiness to change assessment (ORCA): development of an instrument based on the promoting action on research in health services (PARIHS) framework. Implement Sci. 2009;4:38.

    Article  PubMed  PubMed Central  Google Scholar 

  36. Nelson JC, Raskind-Hood C, Galvin VG, Essien JD, Levine LM. Positioning for partnerships. Assessing public health agency readiness. Am J Prev Med. 1999;16(3 Suppl):103–17.

    Article  CAS  PubMed  Google Scholar 

  37. WP. Designing Effective Organizations: The Sociotechnical Systems Perspective. New York: Wiley; 1988.

    Google Scholar 

  38. Rubenstein LV, Danz MS, Crain AL, Glasgow RE, Whitebird RR, Solberg LI. Assessing organizational readiness for depression care quality improvement: relative commitment and implementation capability. Implement Sci. 2014;9:173.

    Article  PubMed  PubMed Central  Google Scholar 

  39. Shaw RJ, Kaufman MA, Bosworth HB, Weiner BJ, Zullig LL, Lee SY, et al. Organizational factors associated with readiness to implement and translate a primary care based telemedicine behavioral program to improve blood pressure control: the HTN-IMPROVE study. Implement Sci. 2013;8:106.

    Article  PubMed  PubMed Central  Google Scholar 

  40. Shea CM, Reiter KL, Weaver MA, McIntyre M, Mose J, Thornhill J, et al. Stage 1 of the meaningful use incentive program for electronic health records: a study of readiness for change in ambulatory practice settings in one integrated delivery system. BMC Med Inform Decis Mak. 2014;14:119.

    Article  PubMed  PubMed Central  Google Scholar 

  41. Shea CM, Jacobs SR, Esserman DA, Bruce K, Weiner BJ. Organizational readiness for implementing change: a psychometric assessment of a new measure. Implementation Sci. 2014;9:7.

    Article  Google Scholar 

  42. Toure M, Poissant L, Swaine BR. Assessment of organizational readiness for e-health in a rehabilitation Centre. Disabil Rehabil. 2012;34(2):167–73.

    Article  PubMed  Google Scholar 

  43. Zullig LL, Muiruri C, Abernethy A, Weiner BJ, Bartlett J, Oneko O, et al. Cancer registration needs assessment at a tertiary medical Centre in Kilimanjaro, Tanzania. World Health Popul. 2013;14(2):12–23.

    Article  PubMed  PubMed Central  Google Scholar 

  44. Aarons GA, Hurlburt M, Horwitz SM. Advancing a conceptual model of evidence-based practice implementation in public service sectors. Admin Pol Ment Health. 2011;38(1):4–23.

    Article  Google Scholar 

  45. Wandersman A, Duffy J, Flaspohler P, Noonan R, Lubell K, Stillman L, et al. Bridging the gap between prevention research and practice: the interactive systems framework for dissemination and implementation. Am J Community Psychol. 2008;41(3–4):171–81.

    Article  PubMed  Google Scholar 

  46. Helfrich CD, Hannon PA, Harris JR, Kohn MJ, Hammerback K, Parrish AT, et al. Does readiness-to-change change readiness to implement wellness practices among work sites participating in a site-randomized implementation trial? Poster presentation at the 10th annual conference on the science of dissemination and implementation in health. Arlington; 2017.

Download references


We would like to acknowledge Laura Damschroder for her work on the Consolidated Framework for Implementation Research (CFIR) and her consultation on the development of our codebook. We would also like to thank the peer reviewers for helpful comments on a previous version of the manuscript.


This work was completed through funding from VA’s Quality Enhancement Research Initiative, including funding to the Care Coordination QUERI program project (QUE 15–276). The opinions expressed represent those of the authors and do not necessarily represent the official views of the Department of Veterans Affairs or the United States government.

Author information

Authors and Affiliations



All authors made substantial contributions to the conception and design and were involved in the critical revisions for important intellectual content. Authors IMML and DMD were involved in the analysis and drafting of the manuscript. All authors contributed to the interpretation of the data. All authors read and approved the final manuscript.

Corresponding author

Correspondence to Isomi M. Miake-Lye.

Ethics declarations

Ethics approval and consent to participate

Not applicable.

Consent for publication

Not applicable.

Competing interests

The authors declare that they have no competing interests.

Additional information

Publisher’s Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Supplementary information

Additional file 1.

PRISMA Checklist.

Additional file 2.

Search Strategy.

Additional file 3.


Additional file 4.

Coding Form.

Rights and permissions

Open Access This article is distributed under the terms of the Creative Commons Attribution 4.0 International License (, which permits unrestricted use, distribution, and reproduction in any medium, provided you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made. The Creative Commons Public Domain Dedication waiver ( applies to the data made available in this article, unless otherwise stated.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Miake-Lye, I.M., Delevan, D.M., Ganz, D.A. et al. Unpacking organizational readiness for change: an updated systematic review and content analysis of assessments. BMC Health Serv Res 20, 106 (2020).

Download citation

  • Received:

  • Accepted:

  • Published:

  • DOI: