Open Access
Open Peer Review

This article has Open Peer Review reports available.

How does Open Peer Review work?

Rapid assessment of infrastructure of primary health care facilities – a relevant instrument for health care systems management

Contributed equally
BMC Health Services Research201515:183

https://doi.org/10.1186/s12913-015-0838-8

Received: 26 March 2014

Accepted: 8 April 2015

Published: 1 May 2015

Abstract

Background

Health care infrastructure constitutes a major component of the structural quality of a health system. Infrastructural deficiencies of health services are reported in literature and research. A number of instruments exist for the assessment of infrastructure. However, no easy-to-use instruments to assess health facility infrastructure in developing countries are available. Present tools are not applicable for a rapid assessment by health facility staff. Therefore, health information systems lack data on facility infrastructure.

Methods

A rapid assessment tool for the infrastructure of primary health care facilities was developed by the authors and pilot-tested in Tanzania. The tool measures the quality of all infrastructural components comprehensively and with high standardization. Ratings use a 2-1-0 scheme which is frequently used in Tanzanian health care services. Infrastructural indicators and indices are obtained from the assessment and serve for reporting and tracing of interventions. The tool was pilot-tested in Tanga Region (Tanzania).

Results

The pilot test covered seven primary care facilities in the range between dispensary and district hospital. The assessment encompassed the facilities as entities as well as 42 facility buildings and 80 pieces of technical medical equipment. A full assessment of facility infrastructure was undertaken by health care professionals while the rapid assessment was performed by facility staff. Serious infrastructural deficiencies were revealed. The rapid assessment tool proved a reliable instrument of routine data collection by health facility staff.

Conclusions

The authors recommend integrating the rapid assessment tool in the health information systems of developing countries. Health authorities in a decentralized health system are thus enabled to detect infrastructural deficiencies and trace the effects of interventions. The tool can lay the data foundation for district facility infrastructure management.

Keywords

Health care facility infrastructure Rapid assessment tool Tanzania Health care facility management Maintenance

Background

Health services research and health care management have frequently focussed on the role of supplies and personnel for the quantity and quality of health care services [1]. It is obvious that qualified and motivated personnel are crucial for preventive and curative medical services, and vaccines as well as drugs are of utmost importance for the health of people. However, the role of health care facility infrastructure as a major component of a health care system must not be underestimated. For instance, the WHO Alliance for Health Policy and Systems Research defines six building blocks of health care systems, the infrastructure constituting one component of the building block “service delivery” [2].

The term ‘infrastructure’ is used in manifold ways to describe the structural elements of systems. In the context of a health care system and in reference to health care facilities, we defined “facility infrastructure” as the total of all physical, technical and organizational components or assets that are prerequisites for the delivery of health care services. It can be seen as a major component of the structural quality of a health care system [3,4]. Same applies to health care facilities, i.e., functionality, quality and extent of such components and assets determine the accessibility, availability, quality and acceptability of health care services as well as the working conditions of facility staff [5-10].

Figure 1 displays the seven major components of the infrastructure of a health care facility: (1) the facility and its management, (2) the physical infrastructure, (3) the supply facility system, (4) the disposal system, (5) technical medical equipment, (6) information and communication technology, and (7) the outreach services. Most people associate with infrastructure only (2) and (5), but all components are prerequisites of a good structural quality.
Figure 1

Health care facility infrastructure – major components. Figure 1 displays the seven major components that comprehensively describe all aspects of health care facility infrastructure. All these components have to be considered in the collection of data on facility infrastructure. Assessment specific data like date, name of data collector etc. complete the data collection.

Professional management is required to safeguard the functionality of all components. For instance, maintenance of infrastructure frequently constitutes a problem in resource poor countries [11]. It is frequently neglected due to lack of funds, availability of spare parts, poor training or little availability of maintenance personnel and a culture disregarding maintenance. Consequently, the condition of assets is often rather poor and contributes to the low structural quality of health care services [12]. This calls for a higher managerial awareness of infrastructure.

Although the relevance of health facility infrastructure for the health care quality is obvious, the literature on health care facility infrastructure is limited. In particular, there are no easy-to-use instruments to assess the quality of health care infrastructure. This article presents a rapid assessment tool which was developed for primary health care facilities in Tanzania. The main purpose is to demonstrate how this tool can provide an evidence-base for health policy decision-making by providing fast and reliable knowledge on the condition of health care infrastructure.

The next section provides an overview of existing health care infrastructure assessment instruments. Afterwards, we present the rapid assessment tool which was produced for Tanzania and an appraisal of this tool. The paper closes with a discussion on the integration of the tool in the health system.

Assessment of health care infrastructure

A recent and rather comprehensive study on health care infrastructure was presented by Hsia et al. [13]. They evaluated six infrastructural key components of hospitals and health centres in African sub-Saharan countries under the aspect of access to surgical and emergency care. The study describes “…dramatic deficiencies in infrastructure […] in all countries studied” [13]. An older evaluation of the health care infrastructure of 16 Lutheran hospitals in Tanzania done by Flessa [14] came to the same results.

However, health policy decision-making cannot be based on snap-shot like research. Instead, routine data must be gathered within the existing Health Information Systems (HIS) [15-18]. Many developing countries currently implement the roll-out of a contemporary software-based HIS which includes epidemiological and demographic data. The majority of health data are gathered at the level of health care facilities. Data collection is performed by medical or administrative staff of the health care facilities in a monthly, quarterly or yearly routine. However, health information systems are reported to function insufficiently [16,19]. Furthermore, infrastructural components are only partly assessed even in recently developed tools like ‘Service Availability and Readiness Assessment’ (SARA) [20]. Assessments consider such components only to a minor extent in existing health information systems like MTUHA in Tanzania.

A reliable and sustainable health care facility infrastructure assessment tool must fulfil the following seven criteria: Firstly, a thorough collection of infrastructural data must cover all seven components of infrastructure (comprehensiveness). Secondly, the procedure must be a routine, i.e., data collection, assessment and recording must be simple enough that administrative or medical staff of (primary) health care facilities can perform it without professional support (easiness of use). Thirdly, data collection should not take too much time (rapidness) and, fourthly, it should fit into the existing HIS (adaptability). Fifthly, the tool and data collection process should be standardized so that the results are comparable (standardization). Sixthly, results have to constitute a reliable base for health policy decision-making including allocation decisions of central funds (usefulness). Finally, it must be easily adaptable to varying circumstances (flexibility).

These criteria which were derived from the desired function of the tool are in line with Rogers’ research on the diffusion of innovations [21]. Rogers emphasizes the importance of the perception of an innovation by the members of the respective social system. According to his research, the compatibility of an innovation is positively related to its rate of adoption, while complexity has the contrary effect. While various internal stakeholders of a health system [14] are involved in the use of an innovative assessment tool on facility infrastructure, the authors’ reflection of compatibility and complexity criteria primarily focussed on the first-line users i.e. the health care facility staff.

A number of assessment tools are available in developing countries that consider facility infrastructure, but they seem to fall short to fulfil the criteria stated above. Prominent examples for such tools are the ‘Service Availability and Readiness Assessment’ (SARA), the ‘Health Center Assessment Handbook’ and the methods of Kielmann et al. and of Halbwachs:
  1. 1.

    SARA, first published by WHO in 2012 [22], represents a very profound survey of manifold aspects of health care services. The tool is designed to be applied regularly as a systematic survey of facility service delivery using a standard questionnaire for the assessment [23]. It considers a number of aspects of infrastructure. It produces infrastructural indicators that describe aspects of service availability and service readiness (e.g. supply facilities, medical equipment and waste management). WHO published reports about assessments that used the SARA tool in Sierra Leone (2011) and Zambia (2010) [24]. The reports show the potential of SARA to display assessment results in a very distinguished way (e.g. general service readiness, differentiated as per facility type and in categories like basic amenities, basic equipment, and laboratory means etc.). This kind of data analysis and presentation of results is supportive for health policy decisions regarding interventions into the improvement of specific health services. However, this tool will usually require external professional staff to perform it [25]. Thus, it cannot be used as routine instrument for HIS in many developing countries.

     
  2. 2.

    The ‘Health Center Assessment Handbook’ of the Ethiopian Federal Ministry of Health, Planning and Programming Department and USAID [26] thoroughly describes a tool for a very detailed assessment of all facets of facility infrastructure. This method requires more professional human resources than all other tools.

     
  3. 3.

    Kielmann et al. [27] developed a comprehensive tool focusing on health needs, services and health care systems at district level. Among other aspects, the assessment provides a physical inventory of health facilities. Various components of physical infrastructure are enquired, but the tool does not cover all aspects of facility infrastructure (failing comprehensiveness). The advantage of this method is that it can be used by health care professionals without engineers.

     
  4. 4.

    Halbwachs [28] focuses on the appraisal of management of physical infrastructure of health services. His assessment tool can be applied at national as well as district and at facility level and encompasses “…a semi-quantitative and quick method of appraising the management of physical assets in health care”. The tool uses elements of the protocols for data collection developed by Kielmann et al. [27]. It is more detailed than the latter one, but it requires professional data collectors.

     

Consequently, none of these tools meets the above described requirements of rapid assessment and comprehensiveness. Tools differ in terms of extent of assessment, consideration of infrastructural data categories and use of ratings. All available tools presuppose professional expertise to a major extent. Only the ‘Health Center Assessment Handbook’ considers all components of facility infrastructure. All tools are more time-consuming than acceptable for a rapid assessment and require human or financial resources that exceed the possibilities of routine data collection in a HIS. There is no doubt that these tools are highly appropriate for scientific work or snap-shot like assessments for particular interventions (e.g. a major renovation program financed by development aid), but they cannot be applied for regular assessment within the HIS routine. In our opinion, even SARA which is supposed to be a practical tool for routine falls short of user friendliness and easiness to use.

To our knowledge no assessment tool is available that meets the requirements of rapid facility infrastructure assessment sufficiently. This results in insufficient quality of infrastructural data. Although a yearly professional assessment of facility infrastructure is desirable because it ensures better data quality the authors assumed that this will not be an option for the yearly data collection routine in resource-poor developing countries due to limited financial resources. Consequently, the object of our research was the development of a rapid assessment tool for health care facility infrastructure that can be applied by in-charge staff members of primary health care facilities. The tool shall contribute to the strengthening of primary health care services.

The rapid assessment tool was tested in a health district in Tanzania. At the same time, a civil engineer and a biomedical engineer performed a professional assessment. It included all data queries of the rapid assessment. The results of rapid and full assessment were compared.

The rapid assessment tool will be presented in the next section.

Rapid assessment tool for primary health care facility infrastructure

In this section we describe the rapid assessment tool with its dimensions of data structure, rating schemes, indicators and indices. Afterwards the scope of a pilot test of the tool is outlined.

Data structure of rapid assessment tool

The assessment comprises data that are collected for the facility as an entity, for specific buildings (building-wise) and for technical medical equipment (asset-wise). Types of data queries are “information” (e.g. date, time or name), “predefined text” (e.g. for rating the condition of roofing material) and “predefined ratings” of condition and of availability or reliability. Figure 2 shows the components of health care facility infrastructure for which data are collected.
Figure 2

Aspects of facility infrastructure assessment. Figure 2 shows details of facility infrastructure assessment which describe aspects under the seven major components (compare figure 1).

Data are organized in a hierarchy of 4 levels. Figure 3 depicts this exemplarily for a part of infrastructural component 2 (Physical infrastructure).
Figure 3

Data structure of rapid assessment tool – example. The data structure of the rapid assessment tool is explained in this figure, displaying an excerpt of infrastructural component number 2 (physical infrastructure), with the elements number 0 (compound) and 2 (buildings) and a selection of related sub-elements and data queries. The data code number 20100 is exemplarily indicated bottom left for ‘rating of security of compound’.

Data queries are numbered using five digits (e.g. “20100” stands for “rating of security of compound”, see data query bottom left in Figure 3).

The data structure is flexible for adaption to local circumstances. While a set of standard data queries must remain unchanged, specific locally required data queries can be added. However, any adaption to specific local circumstances reduces comparability of data in a wider context.

A maximum extent of standardization in the form of predefined answers and simple and consistent rating schemes grants the required easiness of use and the comparability of results. Thus, the adaption into a software-based HIS is facilitated.

The current version of the rapid assessment tool comprises a total of 101 data queries in 7 infrastructural components as shown in Table 1.
Table 1

Rapid assessment tool – components, elements and number of data queries

Number

Infrastructural component (number of data queries)

Infrastructural elements in component (number of data queries)

Focus / Remarks

1

The facility and its management (30)

Statistics (18)

The facility as an entity

Maintenance (1)

Accessibility (1)

Information management (1)

Services (9)

2

Physical infrastructure (26)

Compound (2)

Facility (entity)

Buildings - basic data (5)

Buildings are

Buildings – construction (7)

assessed discretely

Buildings – interior works (6)

Buildings – installations (6)

3

Supply facilities systems (16)

Electrical supply (5)

Facility (entity)

Water supply (5)

Rain water harvesting (6)

4

Disposal systems (14)

Waste (9)

Facility (entity)

Effluent discharge (5)

5

Technical medical equipment (7)

Asset location data (1)

Assets are assessed discretely

Asset statistical data (2)

Asset functionality (2)

Maintenance resources (2)

6

Information Communication Technology (4)

Telephone (2)

Facility (entity)

Internet (2)

7

Outreach services (4)

Transport (2)

Facility (entity)

Referral (2)

Table 1 refers to the seven major infrastructural components (compare Figure 1) and describes all infrastructural elements that are considered under rapid assessment, together with the related number of data queries.

An inventory of facility buildings needs to be compiled prior to the first rapid assessment of facility infrastructure. Same applies to technical medical equipment. Both inventories should be synchronized with standard lists (if available) prior to the assessment. If such standards are not defined on a national level, standard lists (as per facility level) can be developed by applying internationally accepted instruments, such as SARA [22] or Standards-Based Management and Recognition (SBM-R) of JHPIEGO [10,29].

Rating schemes

A number of data elements are assessed by using ratings. With reference to Halbwachs [28] ‘conditions’ are rated using the unified scheme ‘2-1-0’. The implications of ratings are:
  • 2: very good or good condition; therefore no need of action;

  • 1: minor problems; therefore need of action, but not immediate;

  • 0: major problems or hazards; therefore need of immediate action.

These implications need to be explicitly explained to data collectors.

Explanatory text is added where necessary in order to adopt the wording of the ‘2-1-0’ scheme to the specific data requirements of the question (e.g. for ‘roof leakages’: ‘2 = no or minor leakages’, ‘1 = leakages require roof repair’, ‘0 = leakages cause major problems’).

Wherever applicable, data of the ‘information-type’ are transformed to ratings in the analysis of the assessment data according to defined standards. To give an example for such standards, dispensaries are expected to offer services 5 days per week (the datum ‘5 days’ is transformed to the rating ‘2’ for the calculation of the ‘accessibility indicator’).

Indicators, indices and standards

All ratings are grouped under infrastructural aspects and are used for the calculation of indicators that describe the different infrastructural aspects (e.g. ‘accessibility’). The calculation procedure takes into account that the requirements differ as per facility type.

The infrastructural indicators are merged in indices. General facility indicators are combined in a ‘General Facility Index’. Building indices are calculated for all buildings separately and then combined in one ‘Buildings Index’. Asset indicators are combined in the calculation of the ‘Asset Index’. Finally the three major indices form a ‘Facility Infrastructure Index’. All calculations use arithmetic averages. Figure 4 depicts the systematic approach.
Figure 4

Indicators and indices of rapid assessment of health care facility infrastructure. Figure 4 depicts the systematic approach how infrastructural indicators and indices are calculated: data of the ‘rating-type’ are used for the calculation of indicators (for details, refer to Table 2). Indicators are combined under superior aspects (general facility aspects, buildings and medical technical equipment) forming infrastructural indices which finally enter into the ‘Facility Infrastructure Index’.

Infrastructural standards were not available for the field test and thus defined by the authors. Table 2 displays the infrastructural indicators obtainable from rapid assessment of facility infrastructure and the related number of data queries. Scores are calculated in accordance with the defined standards. For instance, the indicator “staff working conditions” achieves the maximum score of 100 if
  • the facility offers a sufficient number of staff quarters AND

  • staff quarters are in good condition and security is rated as good AND

  • private washing facilities and toilets in a good and clean condition are available for staff members AND

  • service and treatment rooms are equipped with separate hand washing facilities for staff members.

Field assessment in Tanga region

Questionnaires for full and rapid assessment were pilot-tested in a field assessment in Tanga Region (Tanzania) in August and September 2012. The full assessment was facilitated by Gesellschaft fuer Internationale Zusammenarbeit (GIZ) and executed by a team of professional engineers with a long working experience in the health care field. The questionnaires for rapid assessment were filled in by the In Charges of the facilities or by other responsible staff members. Data collectors were asked to skip questions in case the wording exceeded their English language skills (for an excerpt of the questionnaire, refer to Additional file 1).
Table 2

Infrastructural indicators in rapid assessment

Object of rapid assessment

Infrastructural indicator

Number of data queries

Facility

Accessibility

2

 

Disposal safety

7

 

ICT reliability

1

 

Maintenance readiness

2

 

Outreach possibilities

3

 

Safety of compound

2

 

Service readiness

4

 

Supply capacity (water, rain water, electricity)

17

Buildings

Physical infrastructure condition

2

 

Safety of rooms

6

 

Staff working condition

9

 

Storage safety

3

Technical medical equipment

Asset status and utilization

2

Table 2 lists all infrastructural indicators that are calculated from data collected by using the rapid assessment tool and the related number of data queries.

The scope of full assessment was a total number of 7 facilities with 42 buildings and 80 pieces of technical medical equipment. The rapid assessment covered 7 facilities and 42 buildings. Equipment was not the object of rapid assessment because the full assessment was done together with responsible staff members. If this had been followed instantly by a rapid assessment a lack of comparability of data would have been anticipated.

In the study the average time required for the assessment of dispensaries was 3 hours for full assessment / 1.5 for rapid assessment respectively 4.5/2.5 hours at health centres and 2 days /1 day at the district hospital.

Results

The results have two dimensions. Firstly, we want to present briefly some results of the assessment itself. Secondly, we will present the results of the applicability of the rapid assessment tool, in particular in comparison to the full assessment undergone.

Health care infrastructure in Tanga region

The gross result of the assessment was not really surprising, i.e., the assessment revealed infrastructural deficiencies. Infrastructural accessibility standards were achieved mainly, security standards partly. Service readiness was limited. Contrary to power supply, standards for water supply were not met at the majority of the facilities. The potential for rain water harvesting was only partly exploited.

Health care services were performed in buildings and rooms of insufficient infrastructural quality. Roof leakages and lack of plinth protection, including an insufficient rain water drainage, were major areas of concern. Facilities were not prepared for regular maintenance of technical medical equipment. They lacked functioning incinerators and separate disposal systems for infectious medical waste. The minority of dispensaries was equipped with placenta pits. All facilities reported problems with effluent discharge systems. Mobile phone reception was good at all facilities but only one facility provided office phones for communication. Web access was very limited.

The number of staff houses was insufficient. There was a lack of separate and clean facilities for hygiene of staff members. The need of maintenance of buildings and equipment was obvious but the facilities lacked capacities. This referred to qualified personnel, to space for workshops and to maintenance equipment.

User-friendliness of the assessment tool

First of all, the interviews revealed a need to translate parts of the English questionnaires to Swahili in all dispensaries and in one health centre. Besides that, the reporting using indicators and indices in diagrams proved useful to point at specific infrastructural areas of concern. The field assessment showed that the tool’s data queries covered the components of facility infrastructure comprehensively. At none of the facilities and in both types of assessment, the data collectors pointed at elements of infrastructure that had not been considered in the development of the tool.

The pilot test showed that the building-specific approach of data collection was challenging to staff members but without alternative. Different buildings in the same facility showed very uneven condition of infrastructure. Without a separate assessment of every single building, infrastructural indicators and indices would give a false image of the actual infrastructural condition of a facility. Same applies to the asset-wise data collection on medical devices.

Using indicators and indices for verbal, numeric or graphic presentation of assessment results allows the easy highlighting of infrastructural components that require improvement. In relation with general health goals of a district, concretisation and prioritization of infrastructural measures is facilitated (for example, ‘improve rain water harvesting potential and its utilization at health centres in region X’ or ‘improve availability and functioning of basic sterilization equipment in dispensaries of Y district’).

The field assessment showed that staff members of all facilities were able to fill in the questionnaires for rapid assessment and to do this in a timely manner. The time required (between two hours and one day) was acceptable. Thus, the utilization of the tools – either in the form of questionnaires within the traditional reporting system or by using software – requires an adequate input of labour.

Comparison of rapid and professional assessment

Data collected by facility staff using the rapid assessment tool (RAT) and by the engineers applying the full assessment tool (FAT) were compared in reference to general facility data, data on buildings and on supply and disposal facilities. Data entries (texts and ratings) of FAT and RAT were examined. Data were excluded from further appraisal if the RAT-data proved to be obviously wrong.

A total number of 284 ratings represented comparable data sets. Using the full professional assessment as reference, the detection of infrastructural deficiencies by use of the rapid assessment showed specificity of 82% and sensitivity of 71%.

The degree of urgency for interventions was rated higher in full assessment than in rapid assessment. This is assumed to have two reasons: firstly, the influence of the professional background of data collectors – an engineer’s expectations regarding the functioning and condition of infrastructure are likely to lead to a more critical rating; secondly, the habituation of staff members to deficiencies they have to cope with in their daily work is presumed to have a mitigating effect on the rating, while the professionals saw the infrastructure for the first time. It can be concluded that areas of infrastructure that require intervention are detected by both tools (rapid and full assessment), across all infrastructural components.

Summary of field test

The survey showed that both assessment tools – full and rapid assessment –covered relevant aspects of the facilities’ infrastructure comprehensively. Standardization of data queries and the predefined and recurring rating schemes supported user friendliness and comparability of results. The field assessment revealed deficiencies of infrastructure and its management, resulting in an inefficient use of resources at the facilities.

In a nut-shell, the rapid assessment tool of health care facility infrastructure proved a reliable instrument of routine data collection by health facility staff that leads to results rather similar to the full assessment of professionals. That calls for its application as an integral instrument of the health information system.

Discussion

Infrastructural data must become a natural component of the yearly routine data flow in a HIS as the melioration of facility infrastructure has the potential of improving health services [5,9,13] thus contributing to the final goal of better health of the population [17]. This unexploited potential should no longer be neglected.

The integration of data on health facility infrastructure into national health information systems in developing countries is overdue. Based on the findings of this case-study it is recommendable to integrate data on facility infrastructure in all health information systems of developing countries.

The rapid assessment tool explained in this article is ready to serve for this purpose. It is available in an excel spreadsheet version that was improved in evaluation of the field survey. The tool is transformable to software that is used for data collection in a HIS (e.g. by smart phones) and considers the capability of the facility staff acting as data collectors [15]. It provides a data structure that specifically measures all components and elements of facility infrastructure and is designed in accordance with the needs of a yearly data collection in existing health information systems.

The tool lays ground for data analysis in a structured reporting format and for the visualization of all relevant facets of facility infrastructure and possible malfunctions for reporting and management purposes. District health authorities are equipped with a tool which they own, which they apply and which they utilize for their decisions.

The regular rapid assessment of facility infrastructure has the potential to serve as an effective management tool for district facility infrastructure management.

Limitations

To determine specificity and sensitivity of the rapid assessment test in this article, the full professional assessment was used as reference test. This assessment procedure requires further practice and evaluation to gain the status of a standard reference test.

The scope and design of the rapid assessment tool is at present limited to primary care facilities in the range between rural health posts and district hospitals. While the general assessment approach remains unchanged, the more complex infrastructure of referral facilities requires the definition of standards that specifically consider the hospitals’ functions.

The classification of medical devices in a rapid assessment requires a simple but effective scheme which has still to be developed. Existing systems like Global Medical Device Nomenclature [30] do not meet the simplicity requirements of rapid assessment.

The current lack of internationally agreed standards for health care facility infrastructure leads to a lack of comparability of results of infrastructure assessment above the national level.

Conclusion

Applying WHO criteria [17], the collection and analysis of data on facility infrastructure by using the rapid assessment tool will enable health authorities (1) to improve the performance of a health care system by detecting and eliminating infrastructural deficiencies; this leads to better services, for example by assuring the availability and functioning of the required technical medical equipment; (2) to respond to threats by improving the reporting on facility infrastructure and related problems; (3) to improve health of the population which is, as above cited research proves, the consequence of improving accessibility, availability and quality of health services by providing good facility infrastructure. The instrument is ready and available for application.

Notes

Declarations

Acknowledgements

Appreciation goes to Gesellschaft fuer Internationale Zusammenarbeit (GIZ) for facilitating a field assessment of the infrastructure of Primary Health Care facilities in Tanga Region / Tanzania and for the generous contribution to the costs of publication.

Godfrey Katabaro (Biomedical Engineer at Bombo Hospital, Tanga, Tanzania) deserves thanks for supporting the design of the assessment tool with regard to technical medical equipment. He was responsible for this part of the field test.

Authors’ Affiliations

(1)
University of Greifswald, Department of Health Care Management
(2)
GIZ Office Tanzania, Tanzanian German Programme to Support Health

References

  1. WHO. World Health Report 2006: Working Together for Health. Geneva: World Health Organization; 2006.Google Scholar
  2. Savigny Dd, Adam T. Systems thinking for health systems strengthening. Geneva: Alliance for Health Policy and Systems Research; World Health Organization; 2009 [Alliance Flagship report series].Google Scholar
  3. Donabedian A. Explorations in quality assessment and monitoring: The Definition of Quality and Approaches to Its Assessment. Volume I. Ann Arbor, Mich: Health Administration Press; 1980. p. 1980–2.Google Scholar
  4. Donabedian A. Explorations in quality assessment and monitoring: The Criteria and Standards of Quality. Volume II. Ann Arbor, Mich: Health Administration Press; 1982. p. 1980–2.Google Scholar
  5. Le Bacq F, Rietsema A. High maternal mortality levels and additional risk from poor accessibility in two districts of northern province, Zambia. Int J Epidemiol. 1997;26:357–63.View ArticlePubMedGoogle Scholar
  6. Urassa DP, Carlstedt A, Nystrom L, Massawe SN, Lindmark G. Quality assessment of the antenatal program for anaemia in rural Tanzania. International J Qual Health Care. 2002;14:441–8.View ArticleGoogle Scholar
  7. Haynes AB, Weiser TG, Berry WR, Lipsitz SR, Breizat AS, Dellinger EP, et al. A Surgical Safety Checklist to Reduce Morbidity and Mortality in a Global Population. N Engl J Med. 2009;360:491–9.View ArticlePubMedGoogle Scholar
  8. Choo S, Perry H, Hesse AAJ, Abantanga F, Sory E, Osen H, et al. Assessment of capacity for surgery, obstetrics and anaesthesia in 17 Ghanaian hospitals using a WHO assessment tool. Trop Med Int Health. 2010;15:1109–15.PubMedGoogle Scholar
  9. Fotso JC, Mukiira C. Perceived quality of and access to care among poor urban women in Kenya and their utilization of delivery care: harnessing the potential of private clinics? Health Policy Plan. 2012;27:505–15.View ArticlePubMedGoogle Scholar
  10. Rawlins BJ, Kim Y, Rozario AM, Bazant E, Rashidi T, Bandazi SN, et al. Reproductive health services in Malawi: an evaluation of a quality improvement intervention // Reproductive health services in Malawi: An evaluation of a quality improvement intervention. Midwifery. 2013;29:53–9.View ArticlePubMedGoogle Scholar
  11. Görgen H, Kirsch-Woik T, Schmidt-Ehry B. The District Health System: Experiences and Prospects in Africa. Manual for public health practitioners. 2nd ed. iesbaden: Universum Verlag: Deutsche Gesellschaft für Technische Zusammenarbeit; 2004.Google Scholar
  12. Flessa S. Internationales Gesundheitsmanagement: Effizienz im Dienst für das Leben. Oldenbourg R: München; 2012.View ArticleGoogle Scholar
  13. Hsia RY, Mbembati NA, Macfarlane S, Kruk ME. Access to emergency and surgical care in sub-Saharan Africa: the infrastructure gap. Health Policy Plan. 2012;27:234–44.View ArticlePubMedGoogle Scholar
  14. Flessa S. Gesundheitsreformen in Entwicklungsländern: Eine kritische Analyse aus Sicht der kirchlichen Entwicklungshilfe. Frankfurt am Main: Lembeck; 2002.Google Scholar
  15. WHO. Developing health management information systems: A practical guide for developing countries. Manila, Philippines: World Health Organization, Regional Office for the Western Pacific; 2004.Google Scholar
  16. Kimaro HC, Nhampossa JL. Analyzing the problem of unsustainable health information systems in less-developed economies: Case studies from Tanzania and Mozambique. Information Techno Dev. 2005;11:273–98.View ArticleGoogle Scholar
  17. Health Metrics Network WHO. Framework and Standards for Country Health Information. Syst Inf Technol Dev. 2005;11(3):278.Google Scholar
  18. Flessa S, Moeller M, Ensor T, Hornetz K. Basing care reforms on evidence: The Kenya health sector costing model. BMC Health Serv Res. 2011;11:128.View ArticlePubMedPubMed CentralGoogle Scholar
  19. Nyamtema AS. Bridging the gaps in the Health Management Information System in the context of a changing health sector. BMC Med Inform Decis Mak. 2010;10:1–6 // 36.View ArticleGoogle Scholar
  20. WHO: Service Availability and Readiness Assessment (SARA). An annual monitoring system for service delivery. Reference Manual [http://www.who.int/healthinfo/systems/SARA_Reference_Manual_Chapter2.pdf?ua=1].
  21. Rogers EM. Diffusion of innovations. 5th ed. New York: Free Press; 2003.Google Scholar
  22. WHO. WHO | Service Availability and Readiness Assessment (SARA) [http://www.who.int/healthinfo/systems/sara_introduction/en/]
  23. WHO. WHO | Service Availability and Readiness Assessment (SARA): an annual monitoring system for service delivery. Reference Manual. Version 2.1 [http://www.who.int/healthinfo/systems/sara_reference_manual/en/]
  24. WHO. WHO | Service Availability and Readiness Assessment (SARA). Reports [http://www.who.int/healthinfo/systems/sara_reports/en/index.html]
  25. O’Neill K, Takane M, Sheffel A, Abou-Zahr C, Boerma T. Monitoring service delivery for universal health coverage: the Service Availability and Readiness Assessment. Bull World Health Organ. 2013;91:923–31.View ArticlePubMedPubMed CentralGoogle Scholar
  26. Federal Ministry of Health, Ethiopia. Health Center Assessment Handbook. … a field guide for Architects and Engineers [http://pdf.usaid.gov/pdf_docs/pnads014.pdf]
  27. Kielmann AA, Janovsky K, Annett H. Assessing district health needs, services, and systems: Protocols for rapid data collection and analysis. Nairobi, Kenya, London: African Medical and Research Foundation; Macmillan Education Ltd; 1991.Google Scholar
  28. Halbwachs H. PAD. Protocols for the Appraisal of Physical Assets Management in Health Services in Developing Economies [http://www.frankshospitalworkshop.com/organisation/management_documents/gtz_physical_assets_management_in_health_services.pdf]
  29. Necochea E, Bossemeyer D. Standards-Based Management and Recognition: A Field Guide. Baltimore: JHPIEGO; 2005.Google Scholar
  30. GMDN Agency Ltd.: GMDN User Guide. A comprehensive guide to the Global Medical Device Nomenclature. Version 2010 [http://www.who.int/medical_devices/innovation/GMDN_Agency_User_Guide_v120810.pdf?ua=1]

Copyright

© Scholz et al.; licensee BioMed Central. 2015

This article is published under license to BioMed Central Ltd. This is an Open Access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/4.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly credited. The Creative Commons Public Domain Dedication waiver (http://creativecommons.org/publicdomain/zero/1.0/) applies to the data made available in this article, unless otherwise stated.