Skip to main content

The Generic Short Patient Experiences Questionnaire (GS-PEQ): identification of core items from a survey in Norway



Questionnaires are commonly used to collect patient, or user, experiences with health care encounters; however, their adaption to specific target groups limits comparison between groups. We present the construction of a generic questionnaire (maximum of ten questions) for user evaluation across a range of health care services.


Based on previous testing of six group-specific questionnaires, we first constructed a generic questionnaire with 23 items related to user experiences. All questions included a "not applicable" response option, as well as a follow-up question about the item's importance. Nine user groups from one health trust were surveyed. Seven groups received questionnaires by mail and two by personal distribution. Selection of core questions was based on three criteria: applicability (proportion "not applicable"), importance (mean scores on follow-up questions), and comprehensiveness (content coverage, maximum two items per dimension).


1324 questionnaires were returned providing subsample sizes ranging from 52 to 323. Ten questions were excluded because the proportion of "not applicable" responses exceeded 20% in at least one user group. The number of remaining items was reduced to ten by applying the two other criteria. The final short questionnaire included items on outcome (2), clinician services (2), user involvement (2), incorrect treatment (1), information (1), organisation (1), and accessibility (1).


The Generic Short Patient Experiences Questionnaire (GS-PEQ) is a short, generic set of questions on user experiences with specialist health care that covers important topics for a range of groups. It can be used alone or with other instruments in quality assessment or in research. The psychometric properties and the relevance of the GS-PEQ in other health care settings and countries need further evaluation.

Peer Review reports


Collection of patient reported outcomes, including patient experiences, is an important component of health services evaluation. According to a recent review [1], several countries have programs for monitoring health care quality using surveys that inquire into patients' and other health care users' experiences. The surveys have different target populations such as the general population [2], users of specific services [3], or patients with specific conditions [4]. The recipients of the results, and hence the use of the results, also vary among the programs between national health authorities, health care managers at different levels, health insurers, providers, potential service users, and researchers. Depending on the survey design, the results can be used to monitor health system performance and/or inform quality improvement efforts at the level of service delivery.

As in many of the reviewed surveys [1], data are collected using lengthy, printed questionnaires in Norwegian national patient experiences surveys. Questions have been raised regarding the burden in collecting (both for respondents and collectors), analysing, and reporting the data versus the usefulness of the survey results [5]. Therefore, many instruments have been modified to short forms to reduce this burden, and to make it possible to use more than one instrument in survey questionnaires (e.g., tools to measure generic and condition-specific health-related quality of life [6, 7] and patient experiences in general practice and in hospitals [8, 9]). However, the methods used for shortening questionnaires vary. The Medical Outcomes Study Short Form (SF-36) was shortened to the SF-8 based on a combination of content coverage issues and an empirical assessment of the correlation between the short and long form [7]. The original instrument developed by the European Task Force on Patient Evaluations of General Practice (EUROPEP) was reduced to the EUROPEP 2006-SF10 version based on factor analysis and an assessment of item discrimination [8]. Factor analysis was used in the development of instruments such as the WHO responsiveness instrument [10] and the Quality from the Patient's Perspective questionnaire (QPP) [9]. The respondents' evaluation of importance was also considered when deciding which items to retain in the QPP.

In 2008, the Norwegian Knowledge Centre for the Health Services (NOKC) received a request from a specialist health care trust about constructing a generic short set of items to be based on the NOKC's previous work. The trust's objective was to use the constructed set of items in a practical, local routine for the continuous collection of patients' and other users' reports on quality. A generic tool would make it possible to approach different target populations with the same items, and a brief format would ease the response burden and was expected to facilitate a good response rate [11]. Limiting the number of items to a maximum of ten would allow testing other questionnaire formats in addition to the traditional printed booklet, such as a single sheet of paper or touch-screens.

The NOKC has developed and assessed a family of questionnaires for six target populations in different specialist health care services [1218]. These instruments have been used in a series of national surveys of patient experiences in Norway, and were the starting point for the present project.

As there are no standardised methods for shortening questionnaires, we built upon international literature to develop a set of criteria to identify core items including relevance, importance, and comprehensiveness. We conducted a survey among nine user groups who had had recent contact with the health trust, asking a number of user experiences questions and also asking, for each question, about the importance of the question's content.

The purpose of this paper is to describe the selection of ten generic core questions that cover the essential dimensions of users' experiences with the services provided across a range of specialist health care services. Question scores are intended to be used as single indicators for each of the ten specific content areas.



All data were collected in 2008 from patients admitted to Helse Bergen health trust. The trust included a university hospital that delivered local and regional specialist health care in the somatic and psychiatric sector and employed 7600 full-time equivalents. Alcohol and substance dependency care was delivered by agreement between the health trust and a private foundation.

Sample selection

Nine groups of care users were included and constituted subsamples: 1) outpatients undergoing rehabilitation; 2) out- and 3) inpatients in somatic care; 4) out- and 5) inpatients in psychiatric care; 6) out- and 7) inpatients in alcohol and substance dependency care, 8) next-of-kin to children in inpatient somatic care; and finally, 9) next-of-kin to children in outpatient mental care. Analyses performed by NOKC on data from previous similar surveys showed that a final sample of 200 to 400 respondents in each group would be good, and 50 to 100 would be acceptable [19]. Therefore, the objective was to obtain 200 respondents from each group.

Staff at the institutions delivered the questionnaires on site to the patients in alcohol and substance dependency care. The trust sent all the other patients and service users the questionnaire by mail at their home address.

Original questionnaires

Candidates for core items were chosen from the established family of instruments (see Table 1). The development and assessment of the six original questionnaires is documented in detail separately. The Patient Experience Questionnaire (PEQ) for adult, somatic inpatients was the first questionnaire to be developed [12] and PEQ was the basis for the later development of the Nordic Patient Experiences Questionnaire (NORPEQ) for inpatients (aged ≥ 16 years) discharged from somatic care [13]. The Outpatient Experiences Questionnaire (OPEQ) was developed for outpatients (aged ≥ 16 years) in somatic clinics and day units [14], the Psychiatric Inpatient Experiences Questionnaire (PIPEQ) for psychiatric inpatients (aged ≥ 18 years) [15], the Psychiatric Outpatient Experiences Questionnaire (POPEQ) for psychiatric outpatients (aged ≥ 18 years) [16], the Parents Experiences of Paediatric Care (PEPC) for reporting the experiences of next-of-kin of children under 16 years in paediatric inpatient care [17], and the Parent Assessment of Outpatient Child and Adolescent Mental Health Services (CAMHS) for reporting the experiences of next-of-kin of children under 16 years in outpatient psychiatric care [18].

Table 1 Origins of the generic questions (dimension in the source questionnaires) and their status in the ten-item version

Apart from the NORPEQ, all of the above instruments were developed separately and by following a similar development and validation process. (I) Preliminary questionnaires were constructed based on a literature review and open verbal interviews with users of the specific service. (II) Preliminary questionnaires were subjected to pre-testing by individual cognitive interviews with a small number of service users, who answered the questionnaire and then commented on the comprehensibility of the items and response options, as well as the relevance of the content and any missing topics. (III) A pilot test with subsequent assessment of the factor structure and psychometric properties completed the questionnaire development process. In 2008, NOKC had projects in progress to develop questionnaires for three more groups: cancer patients (including both out- and inpatient care), patients discharged from rehabilitation institutions, and patients in alcohol and substance dependency treatment.

Construction of the questionnaire about experiences and importance

The above questionnaires had many single items in common; however, the factor structure varied. To ensure that all essential aspects were represented among the candidate items, we included one item from each dimension that was previously identified in each instrument. See Additional file 1 for an outline of the original questionnaires. We chose items with high factor loadings to maximise the response variation. The final number of generic items relating to patient experiences was 23. Some rephrasing was necessary for items with similar topics and dissimilar wording in the established instruments. Adjusting the wording of the items to fit the various user groups and care levels was a significant challenge. For example, the professionals responsible for the care of somatic inpatients are physicians, but for psychiatric outpatients, this may be a psychologist or a specialist nurse. We solved this problem by inserting an introductory paragraph before the items pertaining to the personnel: 'by "the clinicians" we mean: those who have had main responsibility for examinations and treatment. Most often these are physicians, but some receive care from psychologists or other health or social workers'; and 'By "the other staff" we mean: In hospital wards: the nursing staff or the milieu therapists/staff. In outpatient clinics or day care units: the staff you had contact with other than the clinician'. All items were expressed in second person singular. Next-of-kin to children were instructed to base their answers on their own perception of the care that was provided to their children. An English translation of the generic questionnaire about experiences and importance is included as Additional file 2.

The first three questions of the eight-page generic questionnaire were posed to identify the health care contact the respondents were evaluating during completion. All eight items from the brief NORPEQ were included. One item pertaining to both physicians and nursing personnel was divided in two in order to collect separate scores for the personnel groups. Another fourteen items were included from the other five questionnaires to cover dimensions not present in NORPEQ. Twenty-two items about patient experiences had a five-point response scale, which has been shown to be suitable for assessing patient experiences [20]: 1 = 'Not at all', 2 = 'To a small extent', 3 = 'To a moderate extent', 4 = 'To a large extent', and 5 = 'To a very large extent'. In addition, 'Not applicable' was a response option for 22 items. The item relating to waiting times used a four-point scale. Each item was followed directly by the question: 'How important was this to you?'. Because all items were included in the original questionnaires because of their previously demonstrated importance, we used a five-point response scale that allowed discrimination between the higher scores for importance in the generic questionnaire: 1 = 'Not important', 2 = 'A little important', 3='Important', 4 = 'Very important', and 5 = 'Of utmost importance' [21]. Six questions asked for individual background data such as gender, age (in 10-year intervals), self-rated health status, and level of education. Finally, the respondents were asked for their comments about whether there were important topics missing in the questionnaire, and if so, what these were.

After reviewing the contents of the single items and comparing them to the dimensions to which they belonged in the various original questionnaires, we classified the 23 items into nine dimensions: five items were related to clinician services; five to other staff services; two to information; two to involvement in decision-making; three to organisation and co-operation; one to accessibility; two to facilities at the institution; two to general outcome; and one to incorrect treatment.

The first version of the questionnaire was presented in individual cognitive interviews to 16 informants representing all the different target populations. Special attention was directed to potential problems with the items that had been changed to a less precise wording due to the diverse groups of respondents. We also collected information about comprehensibility and whether important topics were missing. The results from the verbal interviews led to minor adjustments in the questionnaire. The interviewees confirmed that the generic wording of the items worked as intended, as the frame of reference, for example of the imprecise expression "other staff", was sufficiently limited within the concrete situations that were to be evaluated.

Postal distribution and responses

The patient administrative system (PAS) was used to include patients to receive postal questionnaires. The primary inclusion criterion was being discharged from inpatient care or having had a consultation in outpatient care during November 2008 or being next-of-kin to a child in the same situation. For outpatients in psychiatric care we used an additional secondary criterion, as they were not included unless they also had had at least three consultations between the beginning of August and the end of November. The secondary inclusion criterion was introduced to avoid including persons that had scarce experience with the services such as patients only recently admitted to care or patients who had discontinued treatment after very short time. Treatments of short duration are frequent in somatic outpatient clinics; therefore, the secondary criterion was not implemented for these patients.

From previous surveys, we had information about response rates in the different groups (ranging from 35% of psychiatric inpatients to 54% of next-of-kin in paediatric care), and this information was used to estimate sufficient subsample sizes. To achieve an adequate size for each subsample, we adjusted the number of included patients according to the expected response rate. Groups with high patient turnover had a surplus of eligible individuals, and from these, we included a random sample. From groups with low turnover, we included all eligible individuals, recognising that 200 respondents for all groups would be unattainable within the study time frame.

The health care contact the respondents chose to evaluate did not necessarily correspond to the contact recorded in the PAS, by which their inclusion was induced. The hospital trust offered care in many locations and departments, and for individuals who did not respond until after the second reminder, ten weeks could have passed between being included and receiving the last reminder. Hence, respondents may have had several health care contacts in the meantime. Therefore, we used 'inclusion groups' based on information from the PAS when calculating response rates and comparing respondents to non-respondents, and in the analyses we used 'analysis groups' based on the contact identified by the respondents. Respondents who did not provide information about the type of contact they had described were associated with their inclusion group.

The questionnaires, along with an information letter and a postage-paid return envelope, were mailed by the health trust and returns were received at the NOKC. Reminders were mailed after two weeks, and if necessary, again after four more weeks to non- respondents. After mailing the second reminder, anonymous data were collected from the trust's PAS to be used for comparing respondents to non-respondents. The subsamples included by means of PAS information and contacted by mail as well as sample sizes and response rates, are shown in Table 2.

Table 2 Inclusion groups

Subsamples and response rates

The postal contacts yielded 1,225 returned questionnaires. The response rate in the postal survey was assessed in the inclusion groups and was 43.6%, varying considerably among the groups. Adult outpatients from somatic care had the highest response rate (57.3%) and adult psychiatric inpatients had the lowest (23.3%) (Table 2). Mean age for respondents was 51 years and for non-respondents 43 years (t-test, p < 0.001). The tendency to respond did not differ significantly between women and men (χ2-test).

The sizes of the analysis groups were comparable to their corresponding inclusion group sizes, but some small differences were observed. The largest differences were between adult in- and outpatients in somatic and psychiatric services because some respondents that were included as outpatients chose instead to report their inpatient experiences.

Personal distribution and responses

A previous review indicated that a better response rate may be achieved with personal distribution compared to postal distribution of questionnaires among patients in alcohol and substance dependency care [22]. Therefore, these patients were included consecutively and in person by the staff at the institutions. Questionnaires were delivered to inpatients at their discharge and to outpatients in connection with a consultation, and returned directly on site in a closed envelope. The distribution was anonymous, and sending reminders to non-respondents was not possible for this subsample. Patients were included from November 2008 to March 2009.

A form on which to record the personal questionnaire distribution was provided to the institutions in order to calculate response rates and assess potential non-response bias.

According to the respondents' reports, the analysis groups contained 52 inpatient and 74 outpatient respondents in the alcohol and substance dependency care group. Among these individuals, 26 were included by mail, as they were registered in the PAS with other health care contacts, mainly as outpatients in psychiatry or rehabilitation. The recordings of personal questionnaire distribution in the institutions were insufficient for calculating response rates. To assess the representativeness of this subsample, we obtained statistics about patients in these institutions in 2008 concerning age (in 10-year intervals) and gender. We used the median of the 10-year intervals to compare the respondent group with the population. Compared to the institutions' population of inpatients in 2008, the respondent group was older (mean age was 41.1 versus 45.3 years (t-test, p = 0.016)) and included more women (34.3 versus 41.2%), but the gender difference was not statistically significant (χ2-test). Compared to the institutions' population of outpatients, the respondent group was of the same age (mean age was 40.1 versus 41.1 years (t-test, n.s.)), and it had fewer women (35.9 versus 45.2%), but the gender difference was not statistically significant (χ2-test).

Statistical analysis and criteria for identifying core items

The ten core items should be applicable for all user groups, address topics that are important in the eyes of the users, and cover different aspects of the health care experience.

The criteria used for selecting core items were as follows:

  1. 1.

    Applicability - the prevalence of responding "Not applicable" should be less than 20% in all the respondent groups. A proportion below 20% of item missing has been considered acceptable in previous studies [23]. This criterion implicated that items lacking relevance from the perspective of the respondents in any one of the groups should be excluded to avoid including redundant matter that many respondents would have no basis for evaluating. The remaining items could then be regarded as appropriate for the purpose of describing an aspect of patients' encounters with health care. Appropriateness is a central premise for an item to support good content validity [2428].

  2. 2.

    Importance - the item should be important in the respondents' perspective (high group mean average). Items may be relevant, but still of minor importance. This criterion implicated consulting the target population to evaluate the importance of suggested items [26, 28], and it followed other studies that have asked users to rate the importance of patient experience questions [2, 9].

  3. 3.

    Comprehensiveness - the items should together cover a wide range of topics. It is widely accepted that patient experience is a multidimensional phenomenon. The number of items per dimension was limited to two to allow a variety of topics to be represented. This criterion ensured that the set of items would describe a broad scope of experiences providing the best possible content coverage within the limit of ten questions. Content representativeness is a central premise for good content validity [25, 27, 28].

To assess the association between the core items and global satisfaction, we used item 24 about overall satisfaction as the dependent variable in a regression model and the remaining nine core items as explanatory variables. To determine whether the core items that were selected using the three criteria were advantageous with regard to their ability to explain the variance in global satisfaction scores, we conducted ten multiple regressions with overall satisfaction as the dependent variable and nine explanatory variables that were randomly chosen among all candidate items in the generic questionnaire.

Total sample

For the analyses, we used a sample of 1,324 respondents, with subsample sizes from 52 (inpatients in dependency care) to 323 (adult somatic inpatients).


According to the joint body of the Norwegian Regional Committees for Medical and Health Research Ethics, approval is not required for quality assurance and evaluation projects that are part of the health services activities, even if the projects are carried out with scientific methods and aim to generate knowledge for potential publication, provided that the projects do not include changes in ordinary clinical practice. Also, ethical approval is not required if the data are anonymous [29]. The present study included no changes in clinical practice and the collected data were never linked to person-identifying information.

The health care trust notified The Data Inspectorate about the project (Notification no. 34106) and the responsibility for privacy protection in accordance with regulations was executed by a specific position at the trust.

The included individuals were informed that participation was voluntary and assured of confidentiality.


Generic core items

The prevalence of "Not applicable" responses varied between the analysis groups and the different items—from zero to 58.1% (see Additional files 3 and 4). In general, outpatients judged the 'other staff' and the institutions' co-operation with external service providers to be less relevant. The first criterion left 13 items for further consideration. This conclusion was the same irrespective of whether the per cent "Not applicable" was calculated using the subsample n or the number of valid responses in the subsample as the denominator.

Importance scores were calculated as the average of group means. We tried alternative methods for producing the importance ranking (for example, the percentage of respondents ticking the highest possible score), and although the items' position in the ranking varied depending on the method, it did not vary to an extent that it altered the final set of items.

The 13 remaining items included all five items about the clinicians, and in accordance with the comprehensiveness criterion, only the two with the highest importance scores were selected as core items. Hence, the application of the stated criteria produced a set of ten generic core items (Table 3).

Table 3 Core items

The proportion of missing answers to the core items among the 1,324 respondents varied from 0.8% (item 4) to 4.2% (item 25) (see Additional file 4). The ceiling effect in the scores from the 1,324 respondents varied between the core items—from 13.4% (item 17) to 67.4% (item 26), and the mean ceiling effect between the ten items was 29.5% (data not shown). There were differences between the groups' scores for all the items regarding both health care experiences and importance (see Additional file 5).

Core items and general satisfaction

The adjusted R2 for the model with nine core items regressed on item 24 showed that 70% of the variation in the general satisfaction score could be explained by the scores on the other nine core items in the total sample, ranging from 43% (alcohol and substance dependency care, inpatients) to 81% (psychiatric care, inpatients) in the subsamples. The direct effect of item 17 about involvement in decision-making was not statistically significant, and the item about general outcome had the strongest effect. In the ten models with randomly chosen explanatory variables, the explained variance in the general satisfaction score varied from 57% to 68% (mean 64%) in the total sample.


The study identified ten generic core items covering major dimensions of experiences that patients across a range of specialist health care services report to be important for user groups. To our knowledge, this is the first generic, short questionnaire for collecting data about user experiences across different types of services (inpatient, outpatient), patient groups (somatic, psychiatry, addiction, rehabilitation), and user groups (patients, next-of-kin to children).

The low proportion of missing answers indicates good acceptability of the generic questionnaire, and means that the survey findings are relevant to large proportions of the target populations. The ceiling effect varied among the items. A high ceiling effect is a common feature in measures of patient experiences [30], and reduces the instruments' ability to describe differences through time or between units or institutions. This is a small concern when the instrument is used for large-scale performance monitoring, but may be a weakness if it is to be included in small-scale quality improvement documentation. However, the ceiling effect in this study is a direct consequence of the ceiling effects in the questionnaires we built upon, so it is not a specific concern for the Generic Short Patient Experiences Questionnaire (GS-PEQ).

The present project differed from traditional psychometric development and assessment, in which phenomena that are central for specific groups are identified and then operationalized with multi-item scales for reliable measurement. The present short, generic questionnaire cannot provide information with comparable specificity [31]. The purpose of conducting patient evaluation varies and the methods must be adapted to the context in question [32]. The GS-PEQ is a useful alternative to a single global item or lengthier questionnaires when used in large samples to meet the strategic level management's need for quality monitoring Caution has been raised regarding the use of single global items to collect patients' evaluation [33], and lengthier questionnaires increase the use of resources.

If the purpose of patient surveys is to inform decision-making or evaluate the services at the operational management level, the core items alone will be lacking in specificity, and hence, be less useful [34]. For such purposes, the core items should be used along with tools that specifically address the context in question.

To obtain a high response rate, it is recommended to ensure that the questionnaire content is relevant to potential respondents [11]. Therefore, one argument against a generic format would be that it could have a negative impact on the response rate. Peytremann-Bridevaux et al. [35] compared one generic and two psychiatric—specific patient satisfaction questionnaires in a sample of psychiatric patients. The results indicated that none of the instruments was superior to the others regarding the measurement properties-including response rates—and the patients' evaluation of the questionnaires.

The ten items in the GS-PEQ were rated as having high importance and relevance by respondents in this study. The agreement between importance and relevance suggests that the selected core items reflect general wishes pertaining to all types of health care; a positive outcome from health care enacted by clinicians that are competent and take their patients seriously within a well organised service. However; this implies a limitation regarding more specific perspectives. For example, the question regarding preparation for the time after the treatment was finished was excluded because the proportion "not applicable" score was more than 20 percent in three groups. Among respondents in inpatient substance and alcohol dependency care only four percent found this question "not applicable" and also ranked it fourth in importance. Less than one percent of the respondents who reported experiences from inpatient pediatric care found the question regarding other staff's skills "not applicable" and ranked it ninth in importance. Hence; in surveys of specific groups, one should consider including items that are rated as important by that specific group, but are not included among the core items due to low applicability for any of the other groups (see Additional files 4 and 5).

The low response rate—despite using two reminders-is a potential limitation in the study. It is important to assess the influence of non-response bias [36]; however, previous studies suggest that the non-response impact is relatively small [37]. Potential non-response bias has been studied in three of the present target populations in Norway: adult somatic inpatients [38], next-of-kin to children in outpatient psychiatric care [39], and adult psychiatric outpatients [40]. These studies showed that there are only minor differences between the postal respondents and the postal non-respondents who provided answers through follow-up telephone interviews. It is not known whether these findings can be generalised to populations where such studies have not been conducted. For these populations, the present results should be interpreted with caution.

The simultaneous inclusion of nine user groups is a strength of the present study. However, even if the health trust in question serves both local and regional functions, it is still a one-site study and we do not know to what degree the samples represent users in Norwegian hospitals in general. In addition, we cannot exclude the possibility that further subsample division, e.g., into condition specific groups, might have shown differences between respondents with different conditions. A recent study has shown that patients' priorities varied between five patient groups that were all are categorised as somatic in the current study [33].

The criteria for selecting the ten core items were used to secure content coverage. As a complement to this method, regression analysis was conducted in which the single item about general satisfaction (item 24) was used as the dependent variable and the remaining items as predictors for this variable [33, 41]. The fact that the selected items could account for a large proportion of the variance in global satisfaction gives additional evidence for the claim that the GS-PEQ covers relevant aspects of health care experiences across the included health services. Moreover, this regression model was compared to ten other models with random selection of predictor variables across the 23 piloted items. The differences between the random models and the model with the selected items were small, but consistently in favour of the model with the selected core items. Of particular concern is the low proportion of explained variance for inpatients in dependency care compared to other groups, which may indicate that important variables are missing for this group. However, our instruction to the respondents to report missing topics in both the verbal interviews and in the present survey did not provide new topics, leaving a substantiated assumption that all topics of importance were covered.

Test-retest reliability and response rates for the GS-PEQ could not be calculated for the data collected in the present survey. The effect of different distribution modes should be studied such as when or where the responses are given (i.e., at home on paper or on site on touch-screens).

Single items are normally less reliable than scales [42], meaning that larger sample sizes are needed for single items to achieve a similar level of score precision. Further research should assess sample sizes needed to achieve satisfactory reliability levels for items in the GS-PEQ, in addition to data quality issues and psychometric properties in studies using GS-PEQ as a standalone tool or alongside other measures. Further research is also needed to examine the usefulness of comparing results between different health care services. For example, the relatively low scores for many of the experiences for psychiatric inpatients may be a true reflection of service quality at this specific hospital trust. The low scores from psychiatric patients may also be explained by their health condition, as was shown in a previous study in which there were similar differences when comparing this group to others [43].


The GS-PEQ is a comprehensive and efficient set of items covering important aspects of user experiences across a range of specialist health care services. The GS-PEQ could be used alone or with other instruments in quality measurement activities at different levels or in research, but the relevance of GS-PEQ for other health care settings and countries requires further evaluation.


  1. 1.

    Garratt AM, Solheim E, Danielsen K: National and cross-national surveys of patient experiences: a structured review. 2008, Oslo, The Norwegian Knowledge Centre for the Health Services, Report from the Knowledge Centre no 07-2008

    Google Scholar 

  2. 2.

    Valentine N, Darby C, Bonsel GJ: Which aspects of non-clinical quality of care are most important? Results from WHO's general population surveys of "health systems responsiveness" in 41 countries. Soc Sci Med. 2008, 66: 1939-1950. 10.1016/j.socscimed.2007.12.002.

    Article  PubMed  Google Scholar 

  3. 3.

    Giordano LA, Elliott MN, Goldstein E, Lehrman WG, Spencer PA: Development, implementation, and public reporting of the HCAHPS survey. Med Care Res Rev. 2010, 67: 27-37. 10.1177/1077558709341065.

    Article  PubMed  Google Scholar 

  4. 4.

    Paddock LE, Veloski J, Chatterton ML, Gevirtz FO, Nash DB: Development and validation of a questionnaire to evaluate patient satisfaction with diabetes disease management. Diabetes Care. 2000, 23: 951-956. 10.2337/diacare.23.7.951.

    CAS  Article  PubMed  Google Scholar 

  5. 5.

    Draper M, Cohen P, Buchan H: Seeking consumer views: what use are results of hospital patient satisfaction surveys?. Int J Qual Health Care. 2001, 13: 463-468. 10.1093/intqhc/13.6.463.

    CAS  Article  PubMed  Google Scholar 

  6. 6.

    Kosinski M, Bayliss MS, Bjorner JB, Ware JE, Garber WH, Batenhorst A, et al: A six-item short-form survey for measuring headache impact: the HIT-6. Qual Life Res. 2003, 12: 963-974. 10.1023/A:1026119331193.

    CAS  Article  PubMed  Google Scholar 

  7. 7.

    Ware J, Kosinski M, Dewey JE, Gandek B: How to Score and Interpret Single-Item Health Status Measures: A Manual for Users of the SF-8 ™ Health Survey. 2001, Lincoln, RI: QualityMetric Incorporated

    Google Scholar 

  8. 8.

    Wensing M, Baker R, Vedsted P, Heje H, Klingenberg A, Broge B, et al: EUROPEP 2006 Revised Europep instrument and user manual. 2006, TOPAS Europe Association. Accessed 13-8-2010, []

    Google Scholar 

  9. 9.

    Wilde LB, Larsson G: Development of a short form of the Quality from the Patient's Perspective (QPP) questionnaire. J Clin Nurs. 2002, 11: 681-687. 10.1046/j.1365-2702.2002.00640.x.

    Article  Google Scholar 

  10. 10.

    Darby C, Valentine N, Murray CJL, de Silva A: World Health Organisation (WHO): strategy for the measurement of responsiveness. 2000, Geneva, World Health Organisation. Global Programme on Evidence for Health Policy Discussion Paper Series, No. 23

    Google Scholar 

  11. 11.

    Edwards P, Roberts I, Clarke M, DiGuiseppi C, Pratap S, Wentz R, et al: Increasing response rates to postal questionnaires: systematic review. BMJ. 2002, 324: 1183-10.1136/bmj.324.7347.1183.

    Article  PubMed  PubMed Central  Google Scholar 

  12. 12.

    Pettersen KI, Veenstra M, Guldvog B, Kolstad A: The Patient Experiences Questionnaire: development, validity and reliability. Int J Qual Health Care. 2004, 16: 453-463. 10.1093/intqhc/mzh074.

    Article  PubMed  Google Scholar 

  13. 13.

    Oltedal S, Garratt A, Bjertnaes O, Bjornsdottir M, Freil M, Sachs M: The NORPEQ patient experiences questionnaire: data quality, internal consistency and validity following a Norwegian inpatient survey. Scand J Public Health. 2007, 35: 540-547. 10.1080/14034940701291724.

    Article  PubMed  Google Scholar 

  14. 14.

    Garratt A, Bjertnaes O, Krogstad U, Gulbrandsen P: The OutPatient Experiences Questionnaire (OPEQ): data quality, reliability, and validity in patients attending 52 Norwegian hospitals. Qual Saf Health Care. 2005, 14: 433-437. 10.1136/qshc.2005.014423.

    CAS  Article  PubMed  PubMed Central  Google Scholar 

  15. 15.

    Garratt A, Danielsen K, Bjertnaes O, Ruud T: PasOpp - en metode for å måle brukererfaringer i psykisk helsevern [PIPEQ--a method for measurement of user satisfaction in mental health services]. Tidsskr Nor Laegeforen. 2006, 126: 1478-1480.

    PubMed  Google Scholar 

  16. 16.

    Garratt A, Bjorngaard JH, Dahle KA, Bjertnaes OA, Saunes IS, Ruud T: The Psychiatric Out-Patient Experiences Questionnaire (POPEQ): data quality, reliability and validity in patients attending 90 Norwegian clinics. Nord J Psychiatry. 2006, 60: 89-96. 10.1080/08039480600583464.

    Article  PubMed  Google Scholar 

  17. 17.

    Garratt A, Bjertnaes O, Barlinn J: Parent experiences of paediatric care (PEPC) questionnaire: reliability and validity following a national survey. Acta Paediatr. 2007, 96: 246-252.

    Article  PubMed  Google Scholar 

  18. 18.

    Bjertnaes O, Garratt A, Helgeland J, Holmboe O, Dahle KA, Hanssen-Bauer K, et al: Foresattes vurdering av barne- og ungdomspsykiatriske poliklinikker [Parent assessment of outpatient child and adolescent mental health services for children and adolescents). Tidsskr Nor Laegeforen. 2008, 128: 1041-1045.

    PubMed  Google Scholar 

  19. 19.

    Oltedal S, Helgeland J, Garratt A: Pasienters erfaringer med døgnenheter ved somatiske sykehus. Metodedokumentasjon for nasjonal undersøkelse i 2006 [Inpatient experiences in somatic hospitals. Documentation of methods used in the national survey in 2006]. 2007, Oslo, The Norwegian Knowledge Centre for the Health Services, PasOpp-rapport Nr.2

    Google Scholar 

  20. 20.

    Garratt AM, Helgeland J, Gulbrandsen P: Five-point scales outperform 10-point scales in a randomized comparison of item scaling for the Patient Experiences Questionnaire. J Clin Epidemiol. 2010

    Google Scholar 

  21. 21.

    Nathorst-Boos J, Munck IM, Eckerlund I, Ekfeldt-Sandberg C: An evaluation of the QSP and the QPP: two methods for measuring patient satisfaction. Int J Qual Health Care. 2001, 13: 257-264. 10.1093/intqhc/13.3.257.

    CAS  Article  PubMed  Google Scholar 

  22. 22.

    Danielsen K, Garratt A, Kornør H: Måling av brukererfaringer med avhengighetsbehandling: En litteraturgjennomgang av validerte måleinstrumenter [The measurement of patient experiences in alchohol and substance dependency treatment. A literature review of validated measurement tools]. 2007, Oslo, The Norwegian Knowledge Centre for the Health Services

    Google Scholar 

  23. 23.

    Valentine N, Prasad A, Rice N, Robone S, Chatterji S: Health systems responsiveness: a measure of the acceptability of health-care processes and systems from the user's perspective. Performance Measurement for Health System Improvement. Edited by: Smith, PC., Mossialos E, Papanicolas I, Leatherman Se. 2009, Cambridge: Cambridge University Press, 138-186.

    Google Scholar 

  24. 24.

    Fitzpatrick R, Davey C, Buxton MJ, Jones DR: Evaluating patient-based outcome measures for use in clinical trials. Health Technol Assess. 1998, 2: i-74.

    CAS  PubMed  Google Scholar 

  25. 25.

    Fitzpatrick R: Measurement issues in health-related quality of life: Challenges for health psychology. Psychol Health. 2000, 15: 99-108. 10.1080/08870440008400291.

    Article  Google Scholar 

  26. 26.

    McColl E, Jacoby A, Thomas L, Soutter J, Bamford C, Steen N, et al: Design and use of questionnaires: a review of best practice applicable to surveys of health service staff and patients. Health Technol Assess. 2001, 5: 1-256.

    CAS  Article  PubMed  Google Scholar 

  27. 27.

    Nunnally JC, Bernstein IH: Psychometric theory. 1994, New York: Mc-Graw Hill, 3

    Google Scholar 

  28. 28.

    Vogt DS, King DW, King LA: Focus groups in psychological assessment: enhancing content validity by consulting members of the target population. Psychol Assess. 2004, 16: 231-243.

    Article  PubMed  Google Scholar 

  29. 29.

    Regionale komiteer for medisinsk og helsefaglig forskningsetikk [Regional Committees for Medical and Health Research Ethics]: Eksempler på virksomhet som ikke skal søke REK [Examples of activities that do not require approval]. Accessed 11-2-2011, []

  30. 30.

    Sitzia J, Wood N: Patient satisfaction: a review of issues and concepts. Soc Sci Med. 1997, 45: 1829-1843. 10.1016/S0277-9536(97)00128-7.

    CAS  Article  PubMed  Google Scholar 

  31. 31.

    Smith GT, McCarthy DM, Anderson KG: On the sins of short-form development. Psychol Assess. 2000, 12: 102-111.

    CAS  Article  PubMed  Google Scholar 

  32. 32.

    Carr-Hill RA: The measurement of patient satisfaction. J Public Health Med. 1992, 14: 236-249.

    CAS  PubMed  Google Scholar 

  33. 33.

    de Boer D, Delnoij D, Rademakers J: Do patient experiences on priority aspects of health care predict their global rating of quality of care? A study in five patient groups. Health Expect. 2010

    Google Scholar 

  34. 34.

    Reeves R, Seccombe I: Do patient surveys work? The influence of a national survey programme on local quality-improvement initiatives. Qual Saf Health Care. 2008, 17: 437-441. 10.1136/qshc.2007.022749.

    CAS  Article  PubMed  PubMed Central  Google Scholar 

  35. 35.

    Peytremann-Bridevaux I, Scherer F, Peer L, Cathieni F, Bonsack C, Clθopas A, et al: Satisfaction of patients hospitalised in psychiatric hospitals: a randomised comparison of two psychiatric-specific and one generic satisfaction questionnaires. BMC Health Serv Res. 2006, 6: 108-10.1186/1472-6963-6-108.

    Article  PubMed  PubMed Central  Google Scholar 

  36. 36.

    Asch DA, Jedrziewski MK, Christakis NA: Response rates to mail surveys published in medical journals. J Clin Epidemiol. 1997, 50: 1129-1136. 10.1016/S0895-4356(97)00126-1.

    CAS  Article  PubMed  Google Scholar 

  37. 37.

    Lasek RJ, Barkley W, Harper DL, Rosenthal GE: An evaluation of the impact of nonresponse bias on patient satisfaction surveys. Med Care. 1997, 35: 646-652. 10.1097/00005650-199706000-00009.

    CAS  Article  PubMed  Google Scholar 

  38. 38.

    Guldvog B, Hofoss D, Pettersen KI, Ebbesen J, Ronning OM: PS-RESKVA -pasienttilfredshet i sykehus [PS-RESKVA -patient satisfaction in hospitals]. Tidsskr Nor Laegeforen. 1998, 118: 386-391.

    CAS  PubMed  Google Scholar 

  39. 39.

    Dahle KA, Iversen H, Sjetne IS, Oltedal S: Do non-respondents' characteristics and experiences differ from those of respondents? Poster at the the 25th International Conference by the Society for Quality in Health Care (ISQua). 2008, Copenhagen

    Google Scholar 

  40. 40.

    Iversen H, Dahle KA, Holmboe O: Using telephone interviews to explore potential non-response bias in a national postal survey. Poster at the the 26th International Conference by the Society for Quality in Health Care (ISQua). 2009, Dublin

    Google Scholar 

  41. 41.

    Crow R, Gage H, Hampson S, Hart J, Kimber A, Storey L, et al: The measurement of satisfaction with healthcare: implications for practice from a systematic review of the literature. Health Technol Assess. 2002, 6: 1-244. 10.1007/978-94-010-0481-7_1.

    CAS  Article  PubMed  Google Scholar 

  42. 42.

    Streiner DL, Norman GR: Health measurement scales: a practical guide to their development and use. 2003, Oxford: Oxford University Press, 3

    Google Scholar 

  43. 43.

    Rahmqvist M: Patient satisfaction in relation to age, health status and other background factors: a model for comparisons of care units. Int J Qual Health Care. 2001, 13: 385-390. 10.1093/intqhc/13.5.385.

    CAS  Article  PubMed  Google Scholar 

Pre-publication history

  1. The pre-publication history for this paper can be accessed here:

Download references


We appreciate the co-operation of the personnel at Helse Bergen HF—Anita Lyssand, Alf Marton Aksland, and Odd Rune Hegrenes in particular.

The software used for patient inclusion and data collection was developed by Tomislav Dimoski.

We thank Saga Høgheim for careful data entry.

Author information



Corresponding author

Correspondence to Ingeborg Strømseng Sjetne.

Additional information

Competing interests

The Norwegian Knowledge Centre for the Health Services is a non-commercial public organisation that regularly conducts national surveys for varying target populations receiving or delivering health care.

Authors' contributions

ISS planned the paper with OAB, RVO, HII, and GB, co-operated with the health trust in data collection, carried out the statistical analysis, and drafted the paper. All authors contributed with critical revisions of the draft and all authors read and approved the final manuscript.

Electronic supplementary material

Rights and permissions

This article is published under license to BioMed Central Ltd. This is an Open Access article distributed under the terms of the Creative Commons Attribution License (, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.

Reprints and Permissions

About this article

Cite this article

Sjetne, I.S., Bjertnaes, O.A., Olsen, R.V. et al. The Generic Short Patient Experiences Questionnaire (GS-PEQ): identification of core items from a survey in Norway. BMC Health Serv Res 11, 88 (2011).

Download citation


  • Generic Questionnaire
  • Dependency Care
  • Health Care Experience
  • Core Item
  • Health Trust