Skip to main content
  • Research article
  • Open access
  • Published:

A Rasch analysis of the Person-Centred Climate Questionnaire – staff version



Person-centred care is the bedrock of modern dementia services, yet the evidence-base to support its implementation is not firmly established. Research is hindered by a need for more robust measurement instruments. The 14-item Person-Centred Climate Questionnaire - Staff version (PCQ-S) is one of the most established scales and has promising measurement properties. However, its construction under classical test theory methods leaves question marks over its rigour and the need for evaluation under more modern testing procedures.


The PCQ-S was self-completed by nurses and other care staff working across nursing homes in 35 Swedish municipalities in 2013/14. A Rasch analysis was undertaken in RUMM2030 using a partial credit model suited to the Likert-type items. Three subscales of the PCQ-S were evaluated against common thresholds for overall fit to the Rasch model; ordering of category thresholds; unidimensionality; local dependency; targeting; and Differential Item Functioning. Three subscales were evaluated separately as unidimensional models and then combined as subtests into a single measure. Due to large number of respondents (n = 4381), two random sub-samples were drawn, with a satisfactory model established in the first (‘evaluation’) and confirmed in the second (‘validation’). Final item locations and a table converting raw scores to Rasch-transformed values were created using the full sample.


All three subscales had disordered thresholds for some items, which were resolved by collapsing categories. The three subscales fit the assumptions of the Rasch model after the removal of two items, except for subscale 3, where there was evidence of local dependence between two items. By forming subtests, the 3 subscales were combined into a single Rasch model which had satisfactory fit statistics. The Rasch form of the instrument (PCQ-S-R) had an adequate but modest Person Separation Index (< 0.80) and some evidence of mistargeting due to a low number of ‘difficult-to-endorse’ items.


The PCQ-S-R has 12 items and can be used as a unidimensional scale with interval level properties, using the nomogram presented within this paper. The scale is reliable but has some inefficiencies due to too few high-end thresholds inhibiting discrimination amongst populations who already perceive that person-centred care is very good in their environment.

Peer Review reports


Person-centredness is internationally regarded as an essential design principle underpinning modern dementia care services. Although tracing its roots to Rogerian psychotherapy [1], the seminal works of Tom Kitwood [2] are seen as defining the starting-point of person-centred dementia care, which seeks to address how perceptions of dementia can detrimentally affect a person’s standing in relation to those around them [3]. Gerontological nursing has since produced a healthy stock of related conceptual advances, particularly in developing enabling care relationships [4,5,6]. Each share social constructionist perspectives of ageing [7, 8] and are concerned with how identity, personhood and agency can be reinforced or compromised depending on the nature of the care environment [9]. Although there is no universally agreed definition, most articulations describe care based on a holistic understanding of a person’s lived experiences; providing a care environment congruent with their preferences and that encourages expression of self; promoting their place as valued members of social relationships and networks; and tailoring support to the individual [3, 5, 10].

The rise of person-centredness in dementia care finds support in an encouraging, if not definitive, evidence-base. Experimental designs have associated person-centred approaches with reduced behavioural symptoms (particularly agitation) and use of neuroleptics with care home residents [11,12,13]; with observational and qualitative studies also having linked person-centredness with improved wellbeing and physical health outcomes [14], and benefits for care workers [15]. However, evidence is inconsistent, and there is a dearth of research exploring the importance of how person-centredness is best implemented. A lack of high quality measurement instruments has been widely highlighted as one impediment to rigorous research [16, 17]. The most established measures, particularly Dementia Care Mapping [18], demands intensive recording of care interactions by specially-trained observers which is beyond the resources of many services and research groups. The need for robust questionnaire-based instruments has been highlighted [17].

The Person-Centred Climate Questionnaire

The Person-Centred Climate Questionnaire is one of the most well-documented and widely tested scales available for evaluating the person-centred quality of the care environment within institutional settings [17, 19]. It is based on an empirically-developed theory of how supportive care environments can protect personhood in the context of cognitive decline and beyond [19], and comprises 14 statements for respondents to assess their agreement with on a 6-point Likert scale (1 = No, I disagree completely, to 6 = Yes, I agree completely). The staff version (PCQ-S) is a proxy report version for use in care settings where an expected high prevalence of cognitive impairment/ dementia would inhibit self-report responses. Its 14 items are organized into three subscales - see Table 1 – spanning safety, homeliness and community. The original PCQ-S was developed in Swedish, but empirically-tested English, Norwegian, Slovenian, Chinese and Korean versions have been published [20,21,22,23]. A patient-completed version (PCQ-P) is also available, although that is not the subject of the present article [24].

Table 1 Items and factors of the PCQ-S

The strengths of the PCQ-S include its encouraging measurement properties, established across an array of empirical studies. Content validity has been supported by expert agreement methods [25] whilst repeated factor analytic studies in independent populations have supported a reasonably stable three factor structure [26]. Cronbach alpha for the subscales, and for the global score, are consistently estimated above 0.80 – even in other languages it has been translated into. Specific studies of reliability and cut-scores have been undertaken, providing greater utility for application in practice [27]. The PCQ-S is firmly established as a regular research tool in psychosocial studies of dementia care and its implications for patient welfare and staff satisfaction [28,29,30].

However, the PCQ-S was developed under classical test theory (CTT) which has been subject to increasing criticism as the appropriate framework for measurement instruments [31]. Four common limitations of CTT are highlighted here. First, the assumption that ordinal Likert-type items can be summed to form a measure with interval-level properties remains to be proven [32]. A significant leap of faith is necessary for the score of one point on an ordinally-constructed instrument to be assumed truly equal across the entire length of the scale. If this assumption is breached, parametric tests are not supported and even simple mathematical operations (e.g. mean scores) are then inappropriate [32, 33]. Second, reliability estimates are commonly thought to be artificially inflated in the presence of locally dependent items; whereby the pursuit of high Cronbach alpha statistics causes near-equivalent items to be combined as though they were statistically independent [34]. Third, error is assumed to be constant across the distribution of measurement whereas it is likely to vary (and so be less-suited, or demand larger samples, for some studies targeted at either end of the continuum). Finally, the procedures and justification for combining subscales into a single global score are not widely understood and so much research use multifactorial instruments as though they were unidimensional, despite having evidence to the contrary [35].

Rasch analysis has been proposed as a robust measurement paradigm [36]. Developed in education sciences as a means for measuring aptitude, Rasch analysis proposes that the likelihood of a person agreeing with a questionnaire statement is related to its ‘difficulty’. That is, some items are easier to agree with than others, and do not equally convey the same information about quality. Thus, a positive response to a statement that very few other people agree with would likely suggest that the respondent occupies a higher position on the latent continuum (whereas a CTT scale pays no regard to item difficulty). Where a consistent hierarchical structure exists within the questionnaire (a probabilistic form of the Guttman pattern) Rasch analysis forms estimates of each respondent’s location on the latent scale [34]. In the event that an instrument can demonstrate it satisfies a series of assumptions (see below), the resulting measure is assured interval-level properties suited for parametric hypothesis testing in research [34]. Furthermore, Rasch analysis permits detailed investigation of local dependence problems and the distribution of error across the continuum.

This paper aims to establish whether the PCQ-S conforms to Rasch assumptions and to provide a mechanism for researchers to convert raw scores to interval-level Rasch scores.


This study uses cross-sectional data collected as part of the Umeå ageing and health research program in Sweden (U-Age) [37]. The PCQ-S was administered through a self-administered questionnaire distributed to nursing home staff between November 2013 and September 2014. Further detail is as follows.


The U-Age data collection consisted of nationwide randomly selected nursing homes and their residents. A total of 60 (of 290) Swedish municipalities were randomly selected for the project and of those 35 agreed to participate. Within these municipalities, nursing home managers were contacted by telephone and 188 of 202 invited nursing homes agreed to participate. No further attempts were made to approach non-participating municipalities or units to find their reasons for not participating. This study was based on data from 172 nursing homes, since 16 did not return data although they had agreed to participate. The sample comprised staff working within 548 units. Further sampling details are available elsewhere [37].


Rasch analysis was implemented through a partial credit model [38] suitable for polytomous items, within RUMM2030 software. The objective of a Rasch analysis is to construct a scale from individual items and test its suitability for interval level measurement. Scale construction is possible by using a logistic function to relate a person’s probability of answering an item using a given response category to their underlying position on the latent continuum. Scores are thus measured in logits. RUMM2030 enables inspection of key assumptions to be satisfied [34], specifically:

  1. 1.

    Overall fit: A χ2 statistic assesses the overall fit to the Rasch model (against the null hypothesis of perfect fit). In addition, the distribution of standardised residuals for both persons and items should have a standard deviation no larger than 1.4.

  2. 2.

    Item and person fit: Individual person and item residuals should be as close to zero as possible. Residuals in excess of ±2.5 are considered potentially problematic. At the item level, large negative residuals indicate redundancy (akin to very high item-total correlations in CTT). These are evaluated within RUMM2030 using an F-test.

  3. 3.

    Threshold ordering: Each response category should be in the anticipated order and each should have the greatest likelihood of being chosen for some part of the latent continuum.

  4. 4.

    Local dependence: Items should not be correlated beyond that associated with the construct under measurement. Within RUMM2030, this is evaluated through the item residual correlation matrix, with those correlations larger than the mean + 0.20 regarded as problematic.

  5. 5.

    Unidimensionality: Rasch models require that only one construct is being measured. RUMM2030 conducts a principal components analysis of residuals, with negatively / positively loading items then separately used to estimate each respondent’s location on the logit scale. Paired t-tests then estimate the significance of these differences. The proportion of t-tests reaching significance should not exceed 5%.

  6. 6.

    Reliability: The internal consistency is estimated using the Person Separation Index and Cronbach alpha. A PSI in excess of 0.70 is generally viewed as a minimum for research purposes.

  7. 7.

    Differential Item Functioning: In this study, we use a set of general personal and job-related characteristics (gender, age, qualification, and type of care setting) to evaluate whether some groups of respondents have a different likelihood of answering an item / category despite being located at the same point on the latent continuum. DIF may be uniform (a constant differential across the scale) or non-uniform (where differential varies across the scale) and is evaluated through an ANOVA.

Against a null hypothesis of perfect fit, the Rasch tests are known to be over-powered for large n (e.g. n > 400); that is, even acceptable levels of deviation from the Rasch assumptions are found to be statistically significant. Because a large sample was achieved in this study (described below), two 10% random subsamples were drawn without replacement from the full dataset, following the approach of Gibbons et al. [39]. Statistical tests confirmed that there were no significant differences in the characteristics of the two samples. Rasch analysis was first conducted on subsample 1 (the ‘evaluation sample’), which was then reapplied in subsample 2 (the ‘validation sample’); with the stability of fit indicators assessed in both applications. For the purpose of describing the final item locations, standard errors, and a nomogram for converting raw to logit scores, the model was re-estimated in the full sample.

The PCQ-S has been found in previous testing to be best represented by a three-factor structure (see above). Three separate scales were therefore constructed and separately tested. To form a summary scale from the three factors, ‘subtests’ were formed within RUMM2030 (see [38] for a similar example of this process). Under subtests, the items within each subscale are combined and entered as ‘meta-items’ within the Rasch model. Since subtests parcel-out dependency between items, this necessarily reduces internal consistency estimates. In the event that a single summary score formed of these subtests satisfies the Rasch fit assumptions, then this ‘higher order’ Rasch scale is able to resolve the problems caused by dependency between subsets of items.


Of 6902 questionnaires distributed to staff, 4831 were returned representing a 70% response rate. The broad sample characteristics are presented in Table 2. Over 80% of the sample were registered nurses with others representing different grades of non-registered practitioners. Approximately a third were based in group living environments with the bulk of the sample drawn from nursing homes. The size of participating units ranged from 7 to 128 beds and included both general as well as special care units for dementia. As noted above, Rasch analysis was performed in ‘evaluation’ and ‘validation’ subsamples randomly drawn from this dataset.

Table 2 Demographics of the study group

Subscale 1: a climate of safety

Initial Rasch analysis of items 1–5 indicated a poor fit (p < 0.0001, see Table 3). There was evidence of disordered thresholds in four items, which were resolved by rescoring each by combining responses falling in the second and third categories. There appeared to be two additional causes of misfit. First, there were large residuals for items 1 and 5 and, second, evidence of local dependency between item 1 and 2 . Item 1 had a large negative residual indicating over-discrimination and redundancy and was the only item with a statistically significant F-statistic. Upon its removal, the Rasch assumptions were met (χ2(20) = 26.112, p = 0.162) with items fitting appropriately, albeit with a slightly larger than expected residual standard deviation. These four items formed a unidimensional subscale, with under 3% of paired t-tests reaching significance. The thresholds for each item category are shown in accompanying category characteristics curves in Additional file 1. Robustness of fit was tested by re-examining these four items in the validation sample, with similar results being achieved. No evidence of DIF was identified for gender, age group, qualification status, type of care setting or its size (full results from RUMM2030 for DIF analyses are provided as Additional file 2).

Table 3 Summary fit statistics (scales)

Subscale 2: a climate of everydayness

Analysis of items 6–10 indicated some misfit to the Rasch model as evidenced by a borderline significant χ2 value and a large item residual standard deviation. Items 7,8 and 10 all required rescoring in the same form as for items in subscale 1 to resolve disordered thresholds. A potential source of misfit was due to local dependence between items 8 and 9 indicating shared variation beyond the latent trait under measurement. Item 8 was on the threshold of significant misfit, and so was removed. The reduced scale had a non-significant χ2 value, was unidimensional and free from local dependency. Applying the same model to the validation sample gave satisfactory results. As for subscale 1, no evidence of (uniform or non-uniform) DIF was identified for any variables tested.

Subscale 3: a climate of community

Analysis of items 11–14 revealed good fit to the Rasch model including a non-significant χ2 value, with satisfactory distribution of residuals and was unidimensional. All thresholds were ordered without need for rescoring. Analysis of residual correlations indicated some evidence of local dependence between items 13 and 14 (a place where it is easy for patients to talk to staff/a place where patients have someone to talk to). The residual correlation was 0.29 greater than the mean correlation in the matrix. However, removing or combining the items caused other fit difficulties and so the two separate items were kept. No evidence of DIF was identified.

Summary scale

The 12 items forming the three subscales (referred to hereafter as the PCQ-S-R, denoting the Rasch version) were then combined in a single model, showing good fit to Rasch assumptions except for its (anticipated) multidimensionality. The three subscales were then formed as ‘testlets’ within RUMM2030 and re-estimated, showing generally good fit, as evidenced by non-significant χ2 value and non-significant test of unidimensionality once subscales were accounted for. The person separation index for the summary scale was 0.776, which was above the required minimum levels for research purposes (=0.70). Re-estimation within the validation sample supported these findings (see Table 3). Table 4 presents the item level results for the final model.

Table 4 Summary fit statistics (items)

The PCQ-S-R was then estimated for the whole sample. Figure 1 presents the distribution of respondents and item thresholds and indicates targeting problems. With item thresholds anchored with a mean of zero logits (lower panel) the mean of the person distribution was 1.10 logits (standard deviation of 0.86), with 7.6% of respondents at extreme values. A more efficient scale would have questions that were less often affirmed in the sample. Finally, Table 5 presents a nomogram enabling researchers to convert ordinal raw scores to metric logit scores (and rescaled logit scores matching the range of the raw score).

Fig. 1
figure 1

Person-Item Threshold Distribution

Table 5 Nomogram converting PCQ-S 12 item raw score to logit score and rescaled logit score


Amid clarion calls to improve measurement in person-centred care [40], this paper sought to bolster the quality and rigour of one such instrument through application of Rasch analysis to the PCQ-S. The PCQ-S is one of the most widely used questionnaire-based measures of person-centredness in dementia research, being translated into multiple languages and growing evidence of its measurement properties [17]. However, all current work has used classical test theory, leaving the PCQ-S open to concerns over how Likert-type items are simply summed to form the measure. This new research found that a 12-item version, labelled the PCQ-S-R, broadly satisfied the assumptions of the Rasch model by forming subscales from the three separate factors. A notable strength of the analysis is the large sample size on which it is drawn. By using the nomogram, researchers using the PCQ-S-R can be satisfied that any subsequent analysis would be of interval-level scores.

A further advance of the PCQ-S-R is that a single score can be used to accurately represent respondents’ perceptions of person-centredness, rather than relying on three distinct but correlated subscales. Unless a particular dimension is the subject of attention, combining the three scales into a global measure of person-centredness would be a researcher’s preference. Although it is commonplace to sum subscale scores under CTT into a global score, few applications have formally assessed (e.g. through bifactor modelling) how appropriate this would be for the construct of interest. Not all subscales comprise sufficient common variance to justify aggregation [41]. However, the PCQ-S-R, formed of three subscales, passes the Rasch assumptions.

An important feature of the PCQ-S lies in its spread of content across themes that resonate strongly with person-centred literature. However, the response patterns for two items did not accord with Rasch expectations, and so were removed in the PCQ-S-R. Item 1 (‘a place where I feel welcome’) was removed, which might be of some concern given this has been identified as an important aspect of service user and carer experience, at least in hospital settings [42]. Arguably, the notion of ‘feeling welcome’ is pertinent in joining new environments that is not one’s own, and may be less suited to long-term residential settings where some will have been resident for many months and years. It is therefore plausible that other items more accurately capture the essence of what was intended, as is indicated by the Rasch analysis. Similarly, the Rasch analysis found evidence of local dependency between item 8 (‘a place that is quiet and peaceful’) and item 9 (‘a place to get unpleasant thoughts out of your head’). Presumably respondents considered that these were tautological, and therefore the removal of item 8 would not be a considerable loss to the content validity of the scale. Additional research interviews with respondents to the scale would be useful to explore these redundancy issues further.

Rasch analysis has the added advantage of investigating the efficiency of a scale. The results suggest that the PCQ-S suffers from mistargeting. Many of the Likert thresholds at the lower end of the spectrum contribute little information, since so few individuals report care quality that is so poor. By contrast, at the higher quality end of the spectrum, too few thresholds mean that it is challenging to discriminate between respondents’ perceptions. The scale is therefore less-suited for monitoring change in services where person-centredness quality is already strong. Ideally, the PCQ would contain more items that, to be endorsed, would require even higher standards of person-centredness. Targeted qualitative work with those already perceiving that services are of a high quality could help to identify more ‘difficult’ standards for inclusion in the PCQ. In the future, further items could potentially form an item bank for use within computer adaptive testing (whereby questions are tailored during the test depending on earlier responses, to identify more accurate estimates of the phenomenon under measurement.)

The analysis is not without its limitations. First, the analysis relates to the Swedish language version of the PCQ-S, and it cannot be assumed that the same conclusions would apply to other versions. The challenges in creating semantically and culturally equivalent scales are well known [43], and formal analysis would require parallel application using other language versions. Second, the sample is restricted to residential settings and there can be no guarantees that the same findings would have been achieved from hospital-based respondents. That said, there is some reassurance from the absence of any Differential Item Functioning between the nursing homes and other settings within the sample. Finally, it is worth recalling that the PCQ-S relies on staff reports and these may differ from independent ratings of person-centredness in any given facility. Arguably, on average, staff will report more positive views of person-centredness within their facility than outside observers. Ideally these questions require improvement to clarify their distinction and this should be the focus of future research.


The PCQ-S-R is a 12-item scale that can be used to appraise the person-centredness in dementia care settings. The research represents a significant advance since the questionnaire can now be said to have been examined against the rigorous assumptions of the Rasch model, and can be more confidently analysed using parametric statistical procedures. Furthermore, the research offers a means for correctly calculating a single global score rather than three separate subscales. Yet some improvements are still required. Specifically, the scale is mistargeted, meaning that it may not be sensitive to change at higher levels of person-centred quality, and further research could explore and refine two items that may still be locally dependent.

Availability of data and materials

The datasets used and/or analysed during the current study are available on reasonable request.



Classical Test Theory


Differential Item Functioning


Person-centred Climate Questionnaire – Staff


Person-centred Climate Questionnaire – Staff – Rasch version


Person Separation Index


  1. Rogers CR. The attitude and orientation of the counselor in client-centered therapy. J Consult Psychol. 1949;13:82–94.

    Article  CAS  PubMed  Google Scholar 

  2. Kitwood T. Dementia reconsidered: the person comes first. Buckingham: Open University Press; 1997.

    Google Scholar 

  3. Brooker D. What is person-centred care in dementia? Rev Clin Gerontol. 2003;13:215–22.

    Article  Google Scholar 

  4. Dewing J. Concerns relating to the application of frameworks to promote person-centredness in nursing with older people. J Clin Nurs. 2004;13:39–44.

    Article  PubMed  Google Scholar 

  5. McCormack B, McCance TV. Development of a framework for person-centred nursing. J Adv Nurs. 2006;56:472–9.

    Article  PubMed  Google Scholar 

  6. Nolan MR, Davies S, Brown J, Keady J, Nolan J. Beyond person-centred care: a new vision for gerontological nursing. J Clin Nurs. 2004;13:45–53.

    Article  PubMed  Google Scholar 

  7. Bond J. The medicalization of dementia. J Aging Stud. 1992;6:397–403.

    Article  Google Scholar 

  8. Estes CL, Binney EA. The biomedicalization of aging: dangers and dilemmas. Gerontologist. 1989;29:587–96.

    Article  CAS  PubMed  Google Scholar 

  9. Sabat S, Harre R. The construction and deconstruction of self in Alzheimer’s disease. Ageing Soc. 1992;12:443–61.

    Article  Google Scholar 

  10. Wilberforce M, Challis D, Davies L, Kelly MP, Roberts C, Clarkson P. Person-centredness in the community care of older people: a literature-based concept synthesis. Int J Soc Welf. 2017;26:86–98.

    Article  Google Scholar 

  11. Livingston G, Kelly L, Lewis-Holmes E, Baio G, Morris S, Patel N, et al. Non-pharmacological interventions for agitation in dementia: systematic review of randomised controlled trials. Br J Psychiatry. 2014;205(6):436–42.

    Article  PubMed  Google Scholar 

  12. Chenoweth L, King MT, Jeon Y-H, Brodaty H, Stein-Parbury J, Norman R, et al. Caring for aged dementia care resident study (CADRES) of person-centred care, dementia-care mapping, and usual care in dementia: a cluster-randomised trial. Lancet Neurol. 2009;8:317–25.

    Article  PubMed  Google Scholar 

  13. Sloane PD, Hoeffer B, Mitchell CM, McKenzie DA, Barrick AL, Rader J, et al. Effect of person-centered showering and the towel bath on bathing-associated aggression, agitation, and discomfort in nursing home residents with dementia: a randomized, controlled trial. J Am Geriatr Soc. 2004;52:1795–804.

    Article  PubMed  Google Scholar 

  14. McKeown J, Clarke A, Ingleton C, Ryan T, Repper J. The use of life story work with people with dementia to enhance person-centred care. Int J Older People Nursing. 2010;5:148–58.

    Article  Google Scholar 

  15. McCormack B, Karlsson B, Dewing J, Lerdal A. Exploring person-centredness: a qualitative meta-synthesis of four studies. Scand J Caring Sci. 2010;24:620–34.

    Article  PubMed  Google Scholar 

  16. Edvardsson D, Innes A. Measuring person-centered care: a critical comparative review of published tools. Gerontologist. 2010;50:834–46.

    Article  PubMed  Google Scholar 

  17. Wilberforce M, Challis D, Davies L, Kelly MP, Roberts C, Loynes N. Person-centredness in the care of older adults: a systematic review of questionnaire-based scales and their measurement properties. BMC Geriatr. 2016;16:63.

    Article  PubMed  PubMed Central  Google Scholar 

  18. Brooker D. Dementia care mapping: a review of the research literature. Gerontologist. 2005;45:11–8.

    Article  PubMed  Google Scholar 

  19. Edvardsson D, Sjogren K, Lindkvist M, Taylor M, Edvardsson K, Sandman PO. Person-centred climate questionnaire (PCQ-S): establishing reliability and cut-off scores in residential aged care. J Nurs Manag. 2015;23:315–23.

    Article  PubMed  Google Scholar 

  20. Edvardsson JD, Sandman P-O, Rasmussen BH. Sensing an atmosphere of ease: a tentative theory of supportive care settings. Scand J Caring Sci. 2005;19:344–53.

    Article  PubMed  Google Scholar 

  21. Bergland A, Kirkevold M, Edvardsson D. Psychometric properties of the Norwegian person-centred climate questionnaire from a nursing home context. Scand J Caring Sci. 2012;26:820–8.

    Article  PubMed  Google Scholar 

  22. Sjogren K, Lindkvist M, Sandman PO, Zingmark K, Edvardsson D. Psychometric evaluation of the Swedish version of the person-centered care assessment tool (P-CAT). Int Psychogeriatrics. 2012;24:406–15.

    Article  Google Scholar 

  23. Sagong H, Kim DE, Bae S, Lee GE, Edvardsson D, Yoon JY. Testing reliability and validity of the person-centered climate questionnaire-staff version in Korean for long-term care facilities. J Korean Acad Community Health Nurs. 2018;29:11.

    Article  Google Scholar 

  24. Edvardsson D, Koch S, Nay R. Psychometric evaluation of the English language person-centred climate questionnaire - staff version. J Nurs Manag. 2010;18:54–60.

    Article  PubMed  Google Scholar 

  25. Vrbnjak D, Pahor D, Povalej Bržan P, Edvardsson D, Pajnkihar M. Psychometric testing of the Slovenian person-centred climate questionnaire - staff version. J Nurs Manag. 2017;25:421–9.

    Article  PubMed  Google Scholar 

  26. Edvardsson D, Sandman PO, Rasmussen B. Construction and psychometric evaluation of the Swedish language person-centred climate questionnaire – staff version. J Nurs Manag. 2009;17:790–5.

    Article  PubMed  Google Scholar 

  27. Lehuluante A, Nilsson A, Edvardsson D. The influence of a person-centred psychosocial unit climate on satisfaction with care and work. J Nurs Manag. 2012;20:319–25.

    Article  PubMed  Google Scholar 

  28. Sjögren K, Lindkvist M, Sandman P-O, Zingmark K, Edvardsson D. Person-centredness and its association with resident well-being in dementia care units. J Adv Nurs. 2013;69:2196–206.

    Article  PubMed  Google Scholar 

  29. Sjögren K, Lindkvist M, Sandman P-O, Zingmark K, Edvardsson D. To what extent is the work environment of staff related to person-centred care? A cross-sectional study of residential aged care. J Clin Nurs. 2015;24:1310–9.

    Article  PubMed  Google Scholar 

  30. Hobart JC, Cano SJ, Zajicek JP, Thompson AJ. Rating scales as outcome measures for clinical trials in neurology: problems, solutions, and recommendations. Lancet Neurol. 2007;6:1094–105.

    Article  PubMed  Google Scholar 

  31. Grimby G, Tennant A, Tesio L. The use of raw scores from ordinal scales: time to end malpractice? J Rehabil Med. 2012;44:97–8.

    Article  PubMed  Google Scholar 

  32. da Rocha NS, Chachamovich E, de Almeida Fleck MP, Tennant A. An introduction to Rasch analysis for psychiatric practice and research. J Psychiatr Res. 2013;47:141–8.

    Article  PubMed  Google Scholar 

  33. Pallant JF, Tennant A. An introduction to the Rasch measurement model: an example using the hospital anxiety and depression scale (HADS). Br J Clin Psychol. 2007;46:1–18.

    Article  PubMed  Google Scholar 

  34. Reise SP, Scheines R, Widaman KF, Haviland MG. Multidimensionality and structural coefficient Bias in structural equation modeling: a bifactor perspective. Educ Psychol Meas. 2013;73:5–26.

    Article  Google Scholar 

  35. Rasch G. Probabilistic models for some intelligence and attainment tests. Nielsen & Lydiche: Copenhagen; 1960.

    Google Scholar 

  36. Edvardsson D, Backman A, Bergland Å, Björk S, Bölenius K, Kirkevold M, et al. The Umeå ageing and health research programme (U-age): exploring person-centred care and health-promoting living conditions for an ageing population. Nord J Nurs Res. 2016;36:168–74.

    Article  Google Scholar 

  37. Andrich D. A rating formulation for ordered response categories. Psychometrika. 1978;43:561–73.

    Article  Google Scholar 

  38. Gibbons CJ, Kenning C, Coventry PA, Bee P, Bundy C, Fisher L, et al. Development of a multimorbidity illness perceptions scale (MULTIPleS). PLoS One. 2013;8:e81852.

    Article  PubMed  PubMed Central  CAS  Google Scholar 

  39. Harding E, Wait S, Scrutton J. The state of play in person-centred care: a pragmatic review of how person-centred care is defined, applied and measured. London: Health Policy Partnership; 2015.

    Google Scholar 

  40. Reise SP, Bonifay WE, Haviland MG. Scoring and modeling psychological measures in the presence of multidimensionality. J Pers Assess. 2013;95:129–40.

    Article  PubMed  Google Scholar 

  41. Tennant A, Conaghan P. The Rasch measurement model in rheumatology: what is it and why use it? When should it be applied, and what should one look for in a Rasch paper? Arthritis Care Res. 2007;57(8):1358–62.

    Article  Google Scholar 

  42. Bolton L, Loveard T, Brander P. Carer experiences of inpatient hospice care for people with dementia, delirium and related cognitive impairment. Int J Palliat Nurs. 2016;22:396–403.

    Article  PubMed  Google Scholar 

  43. Leplege A, Ecosse E. Methodological issues in using the Rasch model to select cross culturally equivalent items in order to develop a quality of life index: the analysis of four WHOQOL-100 data sets (Argentina, France, Hong Kong, United Kingdom). J Appl Meas. 2000;1(4):372–92.

    CAS  PubMed  Google Scholar 

Download references


This report is independent research arising from a Doctoral Research Fellowship (DRF-2013-06-038) supported by the National Institute for Health Research. The views expressed in this article are those of the author(s) and not necessarily those of the NHS, the National Institute for Health Research or the Department of Health.


MW was funded by a Doctoral Research Fellowship (DRF-2013-06-038) supported by the National Institute for Health Research. The U-Age programme was funded by the Swedish Research Council for Health, Working life and Welfare (2014–4016). This work was part-funded by the Wellcome Trust [ref: 204829] through the Centre for Future Health (CFH) at the University of York’.

Author information

Authors and Affiliations



MW conceived and executed the analysis, and led the manuscript drafting; AS contributed to interpretation of findings and drafting of the manuscript; DE contributed to interpretation of findings and drafting of the manuscript. All authors read and approved the final manuscript.

Corresponding author

Correspondence to Mark Wilberforce.

Ethics declarations

Ethics approval and consent to participate

The SWENIS study including its design was approved by the Umeå Regional Ethical Review Board (Dnr: 2013–269-31) Umeå, Sweden. This paper presents secondary analysis of data collected through the SWENIS study with second and third authors providing access to the anonymized data to the first author under a memorandum of understanding.

Consent for publication

Not applicable.

Competing interests

MW was an Associate Editor for the BMC Geriatrics at the time of the original submission to that journal, but after 18 months of review before transfer to this journal, he was no longer. AS and DE declare that they have no competing interests.

Additional information

Publisher’s Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Supplementary information

Additional file 1.

Category Characteristics Curves.

Additional file 2.

Differential Item Functioning.

Rights and permissions

Open Access This article is distributed under the terms of the Creative Commons Attribution 4.0 International License (, which permits unrestricted use, distribution, and reproduction in any medium, provided you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made. The Creative Commons Public Domain Dedication waiver ( applies to the data made available in this article, unless otherwise stated.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Wilberforce, M., Sköldunger, A. & Edvardsson, D. A Rasch analysis of the Person-Centred Climate Questionnaire – staff version. BMC Health Serv Res 19, 996 (2019).

Download citation

  • Received:

  • Accepted:

  • Published:

  • DOI: