Skip to main content

Association between antenatal care facility readiness and provision of care at the client level and facility level in five low- and middle-income countries

Abstract

Background

Despite growing interest in monitoring improvements in quality of care, data on service quality in low-income and middle-income countries (LMICs) is limited. While health systems researchers have hypothesized the relationship between facility readiness and provision of care, there have been few attempts to quantify this relationship in LMICs. This study assesses the association between facility readiness and provision of care for antenatal care at the client level and facility level.

Methods

To assess the association between provision of care and various facility readiness indices for antenatal care, we used multilevel, multivariable random-effects linear regression models. We tested an inflection point on readiness scores by fitting linear spline models. To compare the coefficients between models, we used a bootstrapping approach and calculated the mean difference between all pairwise comparisons. Analyses were conducted at client and facility levels.

Results

Our results showed a small, but significant association between facility readiness and provision of care across countries and most index constructions. The association was most evident in the client-level analyses that had a larger sample size and were adjusted for factors at the facility, health worker, and individual levels. In addition, spline models at a facility readiness score of 50 better fit the data, indicating a plausible threshold effect.

Conclusions

The results of this study suggest that facility readiness is not a proxy for provision of care, but that there is an important association between facility readiness and provision of care. Data on facility readiness is necessary for understanding the foundations of health systems particularly in countries with the lowest levels of service quality. However, a comprehensive view of quality of care should include both facility readiness and provision of care measures.

Peer Review reports

Background

As the Sustainable Development Goals (SDGs) recognize the importance of both health-care coverage and the quality of health services [1, 2], there has been a global effort to emphasize the critical role of delivering high quality health services [3, 4]. Low-income and middle-income countries (LMICs) are tasked with monitoring progress toward improved coverage and improved service quality. However, measuring service quality is complicated by the lack of a consistent definition of quality of care and the lack of well-defined indicators for measuring quality of care for health services. As quality of care is complex, countries remain unclear about what indicators are most important for them to measure and monitor and what data sources should be utilized to do so. Global priorities are focused on reducing maternal and newborn morbidity and mortality, making the measurement of the quality of health services for pregnant women especially critical. For example, high quality antenatal care (ANC) services have the potential to reduce maternal morbidity and improve newborn survival since these services ensure women maintain a healthy pregnancy and also promote safe delivery and postnatal attendance [5,6,7].

Many quality of care frameworks have been proposed, most building from the seminal quality of care framework introduced by Donabedian in 1988 which characterizes quality at three levels: (1) structure — commonly called facility readiness, (the setting in which care occurs including material resources, human resources, and organizational structure), (2) process — commonly referred to as provision of care, (the quality of medical advice delivered by providers to clients, as well as interpersonal relationships between the provider and the client), and (3) outcome — the effects of care on the health status of patients, including both changes in health status and health behavior, as well as improvements in patient knowledge and the degree of the patient’s satisfaction with care [8]. Donabedian theorized that there is a relationship between facility readiness, provision of care, and health outcomes; high facility readiness increases the likelihood of good provision of care, and good provision of care increases the likelihood of a good outcome [8, 9]. Having the essential facility readiness components is necessary to provide a quality health care service, but not sufficient to guarantee quality care. It is the provision of care, or the set of activities, that transforms facility inputs into improved health outcomes for patients.

Despite the increasing focus on quality of care, data on service quality in LMICs is limited [10, 11]. Increasingly, health facility assessments (HFAs) are being conducted in LMICs to try to fill this data gap [12]. HFAs are a rich source of information about the performance of health systems yet are variable in terms of the information collected on quality of care. Two of the most widely implemented HFAs are the Service Provision Assessment (SPA) and the Service Availability and Readiness Assessment (SARA) [13, 14]. SPA and SARA surveys both capture data on facility readiness across a wide range of health services through implementation of a facility audit. However, only the SPA collects data on provision of care for a limited set of health services through implementation of direct observation of client visits and patient exit interviews [13, 14]. In addition to the SPA, there are several service-specific surveys implemented in LMICs that collect provision of care data such as the Quick Investigation of Quality (QIQ) tool, which assesses family planning, and the Needs Assessments for Emergency Obstetric and Newborn Care (EmONC), which assess delivery and newborn care [15, 16]. The more limited availability of provision of care data can be attributed to the challenges in collecting this data as it is both resource-intensive and logistically challenging to implement [17, 18]. There has been interest in using facility readiness as a potential proxy for provision of care since it is easier to collect. A recent study found that facility-level correlation between readiness and provision of care is low [19]. As a result, the recent The Lancet Global Health Commission on High Quality Health Systems in the SDG Era recommended focusing on provision of care measures when assessing quality of care [3].

While health systems researchers have hypothesized the relationship between facility readiness and provision of care, there have been few attempts to quantify this relationship in LMICs. A number of studies have attempted to identify determinants of service quality in LMICs. Studies have identified associations between service quality and service volume [20, 21], and between various domains of quality [22], but not between structural and process quality. Understanding this association is important for both measurement of quality of care and quality improvement processes. This study builds on the work by Leslie et al. (2017) [19] and incorporates a client-level analysis which allows for adjustment for facility and client level characteristics as well as the use of multiple approaches to generate readiness indices. This study aims to assess the association between facility readiness and provision of care for ANC at the client level and facility level utilizing multiple methods for generating facility readiness indices and adjusting for facility-level and client-level characteristics.

Methods

Data

Service provision assessment (SPA)

This analysis used data from the SPA to generate nationally representative data on health service delivery [13, 23]. The SPA includes a standard set of survey instruments: a facility inventory questionnaire, health worker interviews, observation of ANC consultations, and exit interviews with ANC clients.

We examined all SPA surveys for inclusion in the analysis (total of 31). We included all SPA surveys that were available for public use as of early 2018 which used the Demographic and Health Surveys (DHS)-VI or DHS-VII questionnaire (14 surveys excluded), that included observations of ANC consultations (six surveys excluded), were conducted in the last 10 years (between 2012 and 2022; two surveys excluded), and were the most recent survey for the country meeting inclusion criteria (1 survey excluded). The included surveys are from Haiti (2013), Malawi (2013/2014), Nepal (2015), Senegal (2016), and Tanzania (2014/2015). Comprehensive information on the survey methodology and questionnaires is detailed in the SPA final country reports [24,25,26,27,28]. In Nepal, Senegal, and Tanzania, the survey was a nationally representative sample of health facilities selected using stratified systematic probability sampling with stratification by administrative area (geo-ecological region and development-ecological zone in Nepal, and region in Tanzania) and facility type (with oversampling of some facility types such as hospitals). In Haiti and Malawi, the survey was comprised of a national census of all health facilities. The facility inventory module was completed by all surveyed facilities in the five countries. Additionally, up to eight health workers were interviewed within each facility. Selected health workers include those whose consultations were observed and those who provided information for any section of the inventory questionnaire. Sampling of clients for observation was done using systematic sampling and was dependent on the number of clients present at each service site on the day of the visit. For facilities where the number of ANC clients could not be anticipated, opportunistic sampling was used when clients arrived. At a minimum, five client observations were completed per service provider, with a maximum of 15 observations in any given facility for each service. Client exit interviews were conducted following each client-provider observation.

Analysis

In order to standardize expected clinical actions, we limited this analysis to facilities offering ANC services with at least one first ANC client observation, and to women attending the health facility for a first ANC visit. Of the total number of facilities offering ANC services, the percentage which also included any ANC client observations ranged from 52% in Nepal to 90% in Senegal. After additionally restricting to facilities with first visit ANC client observations and complete cases, the percentage of facilities offering ANC included in the analysis ranged from 32% in Nepal to 55% in Senegal. We did not include observations containing incomplete data. Supplementary Tables 1, Additional File 1 provides information on the full sample size and analytical sample size for each country.

To assess the effect of excluding incomplete cases, we compared facilities, health workers, and ANC clients with and without complete data across background characteristics. For continuous variables, we calculated means and used t-tests to assess differences between groups. For categorical variables, we calculated proportions and used chi-square tests to assess differences between groups. There were no statistically significant differences between the groups. Supplementary Tables 2, Additional File 1 provides the details of the all cases versus complete cases analysis.

Facility readiness and provision of care indices

As described previously by Sheffel et al., we created nine indices for facility readiness using three methods for selecting items (core set of items, expert survey set of items, and maximum set of items) and three methods for combining items (simple additive, weighted additive, and principal component analysis (PCA)) [29]. These methods are commonly utilized across service areas to create summary indices of service quality, yet there is no consensus on a single best approach [12, 29,30,31]. In addition, we created a provision of care index using the expert survey set of items for selecting items and a weighted approach to combine items (Fig. 1). A detailed description of item selection, item combination, and index creation are described in Sheffel et al. [29]. Briefly described, the core set of items (21 readiness items) was identified by reviewing the provision of care items required for an ANC visit based on WHO FANC guidelines and WHO recommendations on ANC for a positive pregnancy experience, and by determining the human resources, equipment and supplies, medicines, and diagnostics required to deliver each specific item. The expert set of items (19 readiness items, 49 provision of care items) was identified using results from an expert survey whereby items rated by the expert group as essential were included. The maximum set of items (38 readiness items) included all items identified in the SPA related to ANC readiness across the following domains: human resources, equipment and supplies, medicines, diagnostics, and basic amenities. For each index, the possible range of scores was from 0 to 100%. The distribution of readiness scores and provision of care scores for each country are presented in Supplementary Fig. 1, Additional File 1 and Supplementary Fig. 2, Additional File 1.

Fig. 1
figure 1

Process of Index Creation

Standardization of facility and health worker level variables

Facility-level variables such as facility type and managing authority were standardized across the five surveys. For facility type, three categories were constructed — hospital, health center/clinic, and dispensary. Since facility size is often associated with facility type, we conducted descriptive analyses of the number of total staff and the number of inpatient beds per facility type in order to determine which facility types were most similar in size. Supplementary Tables 3, Additional File 1 provides details of this analysis. For managing authority, we created three categories — government/public, private (non-faith based), and private faith based. At the health worker level, qualification was standardized across the surveys into three categories — physicians, clinical officers, and nurses/midwives (Supplementary Text, Additional File 1 provides cadre definitions).

Additional covariates

We included facility, health worker, and individual characteristics as covariates in our regression models. Facility-level covariates included facility type, managing authority, urbanicity, and average number of staff. Health worker-level covariates included qualification and gender. Individual-level covariates included the number of weeks pregnant, if the client had a previous pregnancy, age, and the highest level of education attained by the client. For Nepal, the covariate urbanicity was excluded as this data was not collected in the Nepal survey. For Haiti and Senegal, the covariate number of weeks pregnant was excluded because more than 50% of the data was missing for this item.

Regression analysis

To assess the association between provision of care and facility readiness, we first conducted a bivariate regression of provision of care on each facility readiness index described above. Next, we used a multilevel, multivariable linear regression model with random effects for facility and health worker, controlling for facility, health worker, and individual characteristics. We ran nine separate models, one for each facility readiness index. The outcome in each case was the same provision of care index. Analyses were conducted separately for each country. Visual examination of the data suggested a possible inflection point around a readiness score of 50 (see Supplementary Figs. 3–12, Additional File 1); we tested this by fitting linear spline models with a single knot at 50 for all countries except for Senegal where there were few facilities with readiness scores below 50. For each model, we divided the coefficient for the readiness index by the standard error to obtain a measure of the strength of the association that accounts for both the estimate and the standard error.

We compare the readiness coefficients obtained from the client-level analyses to quantify the difference in the strength of association between readiness and provision of care using the different approaches for item selection and item aggregation. We use bootstrapping to non-parametrically estimate the joint distribution of these two coefficients and to estimate their mean difference and its standard error [32, 33]. For each dataset, we generated 500 independent survey samples by resampling facilities with replacement, selecting the same number of facilities as in the original survey. Then, we merged the sampled facilities with client data to get sampled clients for the client-level analysis. For each of the 500 samples, we ran the nine linear models and nine linear spline models. We then calculated the mean difference of the coefficient for facility readiness and the standard error of the mean difference for each of the 36 pairwise comparisons. Statistical significance was determined using a Bonferroni adjusted p-value to account for the multiple pairwise comparisons.

Finally, we collapsed the client-level dataset into a facility-level dataset by taking the mean provision of care index score across clients within a facility and conducted a multivariable linear regression adjusting for facility characteristics. We compare the readiness coefficient obtained from the common facility-level regression with its analogue from the client-level analysis to quantify the bias resulting from failure to control for client-level confounders and other model misspecifications. We use facility-level bootstrapping to non-parametrically estimate the joint distribution of these two coefficients and to estimate their mean difference and its standard error [32, 33].

All statistical analyses were carried out using R version 4.1.3 [34]. We did not include survey weights to account for the complex survey design in the regression analyses. There is considerable debate among survey experts regarding the best approach for regression analysis with complex survey data [35,36,37,38]. Given that this paper aimed to explore the client-level relationship between readiness and provision of care and not to estimate a population mean, we followed the common practice of opting not to weight the data by the complex survey design weights to avoid unnecessarily increasing the variance of regression coefficients and thereby decreasing the study power. By virtue of not using weights, primary level facilities that are underrepresented in the sample are not upweighted in the regression analyses. However, adjusting for facility type and managing authority in the regression models mitigates this potential bias.

Results

Background characteristics

The final analytic sample consisted of 358 health facilities, 406 health workers and 779 ANC clients in Haiti; 253 health facilities, 283 health workers, and 815 ANC clients in Malawi; 282 health facilities, 337 health workers, and 520 ANC clients in Nepal; 179 health facilities, 184 health workers, and 297 ANC clients in Senegal; and 632 health facilities, 799 health workers, and 1,681 ANC clients in Tanzania (Supplementary Table 1, Additional File 1). In all countries, most health facilities were government-run health centers, and in all countries except Nepal the majority of health facilities were rural. In addition, health workers were predominantly female nurses and midwives, with the minority of providers being physicians and clinical officers. More than half of ANC clients were primigravida clients, and the majority had at least a primary education except for in Senegal where just under half of clients had at least a primary education (Table 1).

Table 1 Background Characteristics (Haiti, Malawi, Nepal, Senegal, and Tanzania)

Association between facility readiness and provision of care at the client level

The bivariate analysis results are presented in Supplementary Table 4, Supplementary Table 5, and Supplementary Figs. 3–12; Additional File 1. The adjusted association between facility readiness and provision of care at the client level is shown in Table 2. In Haiti, Malawi, and Tanzania the majority of readiness indices were statistically significantly associated with provision of care while in Nepal and Senegal three out of nine readiness indices were statistically significantly associated with provision of care. However, across all countries the effect sizes were relatively small. The effect size for the measure of association between facility readiness and provision of care at the client level across all countries ranged from 0.05 (core PCA, Nepal) to 0.39 (maximum PCA, Senegal). This range of effect sizes indicates that for every 10-point increase in readiness, provision of care increased by 0.5 points (core PCA, Nepal) to 3.9 points (maximum PCA, Senegal) on a 100-point scale. There was no discernable pattern in terms of which types of readiness indices were more strongly associated with provision of care.

Table 2 Multilevel, Random Effects Model of the Association between Readiness and Provision of Care, Haiti, Malawi, Nepal, Senegal, and Tanzania

The comparison of each of the client-level index estimates of association is shown in Supplementary Table 6, Additional File 1. There were few significant differences between the coefficients from different models, indicating that the approaches to item selection and item combination utilized for readiness index construction ultimately had little impact on the relationship between readiness and provision of care. Across all countries, there was a total of 19-pairwise comparisons with statistically significant differences. Of the 36-pairwise comparisons in each country, Tanzania had the most statistically significant differences (9) while Nepal had the fewest statistically significant differences (1). The index comparisons that were significantly different most often used different approaches for combining the same set of items (9/19 significantly different comparisons).

The adjusted associations between facility readiness and provision of care with a linear spline at 50 at the client level are shown in Table 3; Fig. 2. Adding the spline resulted in larger coefficients, relative to the models without a spline, when the facility readiness score was less than 50, and in smaller coefficients when the readiness score was greater than 50 for almost all models. For example, in Tanzania, when the facility readiness score was less than 50, the measure of association between facility readiness and provision of care at the client level ranged from 0.362 (core weighted) to 0.976 (expert simple) and was statistically significant for all nine indices. When the facility readiness score was more than 50, the measure of association between facility readiness and provision of care at the client level ranged from 0.058 (expert PCA) to 0.16 (core simple) and was statistically significant for eight out of the nine indices.

Table 3 Multilevel, Random Effects Model of the Association between Readiness and Provision of Care with a Spline at 50, Haiti, Malawi, Nepal, and Tanzania
Fig. 2
figure 2

Multilevel, Random Effects Model: Association between Readiness and Provision of Care with Spline at 50

The comparison of each of the client-level spline model estimates of association (using a bootstrapping approach) is shown in Supplementary Table 7, Additional File 1. There were few statistically significant differences between coefficients. Across all countries (except Senegal), there was a total of 4-pairwise comparisons with statistically significant differences, one occurring when facility readiness was less than 50 and three occurring when facility readiness was greater than 50. Of the 72-pairwise comparisons in each country, only Tanzania and Nepal had any statistically significant differences, three and one respectively. The index comparisons that were significantly different most often used different approaches for combining the same set of items (1/1 significantly different comparison when facility readiness was less than 50; 2/3 significantly different comparisons when facility readiness was greater than 50).

Association between facility readiness and provision of care at the facility level

The adjusted association between facility readiness and provision of care at the facility level is shown in Table 4. At the facility level, the effect size for the association between facility readiness and provision of care was similar to that seen in the client-level analysis. However, fewer facility readiness indices showed a statistically significant association with provision of care. Notably, collapsing to a facility-level analysis reduced the sample size in Haiti from 779 clients to 358 facilities, in Malawi from 815 clients to 253 facilities, and in Nepal from 520 clients to 282 facilities. In Tanzania, the facility-level sample size remained fairly large (632 facilities) and the measure of association between facility readiness and provision of care at the facility level was statistically significant for all nine indices. In Senegal, the facility-level sample size remained fairly small (179 facilities, 297 clients).

Table 4 Facility Level Linear Model of the Association between Readiness and Provision of Care, Haiti, Malawi, Nepal, Senegal, and Tanzania

The comparison of each of the facility-level index estimates of association to the client-level index estimates of association are shown in Supplementary Table 8, Additional File 1. Across all countries, the nine comparisons showed that none of the coefficients were statistically significantly different between the client-level models and the facility-level models. However, some models showed a loss of information when collapsing from a client-level to a facility-level dataset, particularly in Haiti, where the facility-level analysis was 1.1 to 1.8 times less efficient than the client-level analysis. All other countries had less information loss in the facility-level model.

The adjusted association between facility readiness and provision of care with a linear spline at 50 at the facility-level is shown in Supplementary Table 9, Additional File 1. Similar to the client-level models, adding the spline resulted in larger coefficients, relative to the models without a spline, when facility readiness was less than 50, and in smaller coefficients when readiness was greater than 50. However, fewer of the associations were statistically significant. This may be partially due to the smaller sample size in some indices, especially in the facility readiness less than 50 group. For example, in Malawi, six of the nine indices had less than 40 observations when readiness was less than 50.

The comparison of each of the facility level spline model index estimates of association to the client-level spline model index estimates of association is shown in Supplementary Table 10, Additional File 1. Across all countries, none of the estimates of association were statistically significantly different between the client-level models and the facility-level models. However, as with the non-spline models, some models showed a loss of information when collapsing from a client-level to a facility-level dataset, particularly in Haiti, where the facility-level spline analysis would require up to 1.4 times the number of observations in order to do as well as the client-level analysis. All other countries had less information loss in the facility-level spline models.

Discussion

This study assessed the association between facility readiness and provision of care for antenatal care at the client level and facility level in five LMICs. We found a small but significant association between facility readiness and provision of care across countries and most index constructions. Across the five countries, a ten-point increase in facility readiness was associated with a one- to two-point increase in provision of care score for most indices. This is consistent with the findings by Leslie et al. (2017) which found a positive but weak correlation between readiness and provision of ANC [19]. In our analysis, this association was strongest in the client-level analyses that had a larger sample size and adjusted for factors at the facility, health worker, and individual levels. In addition, we found that the models with a spline at a facility readiness score of 50 fit the data better, suggesting a threshold effect that is conceptually plausible. Across four of the five countries (Haiti, Malawi, Nepal, and Tanzania), the client-level and facility-level analyses found that when facility readiness was less than 50, a ten-point increase in facility readiness was associated with a one- to ten-point increase in provision of care score. This finding diverges from the findings by Leslie et al. (2017) which did not identify a significant inflection point [19].

We found that the majority of the facility readiness indices were associated with provision of care and there were several facility readiness indices that were strongly associated with provision of care in all countries, including the core simple facility readiness index. These findings may be helpful for countries in selecting an approach for creating quality of care summary indices. Based on these findings, a pragmatic approach that uses simple metrics that are easily calculated, interpreted, and adapted at country level may be warranted.

We also found that the association between readiness and provision of care held at the facility level, but there was a small loss of information when analyzing provision of care data at facility level compared to the client level. However, in some cases this tradeoff may be worthwhile in order to facilitate interpretation and use of data for decision-making. For both global and national monitoring of quality improvements, it may be more meaningful to report the overall quality of a facility as opposed to the aggregate quality of patient-provider interactions across facilities. Similarly, effective coverage is being used increasingly as a metric for monitoring progress towards universal health coverage (UHC) globally [39,40,41,42,43,44]. Many efforts to estimate effective coverage have linked household survey data on coverage of interventions with facility-level data on quality of care [45,46,47,48,49,50,51]. This analysis contributes to the methodological evidence-base for how to measure quality of care for use in estimating effective coverage.

While recent efforts have championed a shift towards measuring provision of care, which continues to be a significant data gap, our results highlight the continued importance of facility readiness, especially in countries that are believed to have poor quality health systems. We found that there was a strong association between readiness and provision of care when facility readiness scores were below 50. This finding is consistent with conceptual frameworks for quality of care which hypothesize that facility readiness is important and to some extent explains process quality, but that process quality is also affected by many other factors including client volume, provider motivation, provider competence, and health system management practices [17, 52,53,54,55,56,57,58]. Our results suggest that, while facility readiness is not a proxy for provision of care, there is a minimum threshold of facility inputs required for health care workers to deliver high quality ANC services. Below the threshold of 50, increases in readiness are associated with increases in provision of care as health care workers are gaining the necessary inputs to deliver a service. However, beyond the threshold, facility inputs are no longer driving the quality of care received. It is likely that once the minimum threshold is reached, it is what the health care workers do with facility inputs that is a larger determinant of the quality of care.

There are significant challenges to collecting provision of care data in LMICs using direct observation of clinical care. Direct observation of client visits requires highly clinically trained data collectors such as doctors or nurses to observe the consultations. In addition, it requires that patients come to the health facility seeking care for the services of interest on the day of the survey. This can be difficult in small, low-volume facilities where patient volume is insufficient to meet survey sample sizes in a timely manner [59]. It is perhaps not surprising that these smaller facilities also often deliver lower quality care, and their exclusion is problematic for obtaining accurate estimates of quality of care. For example, a study by Kruk et al. found that the quality of maternal care was substantially lower in primary care facilities as compared to secondary care facilities and that low delivery volume was consistently associated with poor quality [20]. This may have biased our results towards the null if in fact these smaller facilities have lower levels of readiness and quality as we have hypothesized here. Having data on more lower level facilities in future data collection and analysis efforts may strengthen the findings of this association. Finally, direct observation requires data collectors to be present for the entire length of the service. For some services, such as antenatal care, this may be a relatively short consultation. However, for other services such as labor and delivery, the service can extend over the course of more than 24 hours which in many cases is impractical for data collectors. Record review of patient charts is an alternative to direct observation that has been used in high-income countries and has been proposed for use in LMICs. However, in many LMICs, individual patient charts are not readily available for most services, and both patient charts and registers in LMICs have been found to often be incomplete and of questionable quality [60,61,62]. While it is not impossible to collect provision of care data in LMICs, the resource-intensiveness and logistical challenges have limited the availability of provision of care data. Of particular concern, the countries with the poorest health indicators often do not have provision of care data and may be left behind if they are excluded from quality of care measurement and ultimately quality improvement efforts. However, many LMICs do have data on facility readiness. Our results suggest that these data can provide useful insights into the foundations of a health system which are required for delivering high quality care.

To our knowledge, this is the first study that has examined the client-level association between facility readiness and provision of care for ANC in LMICs. Despite our important findings, there are several limitations to this study. First, the analysis is limited in scope as it is comprised of one country in Latin America, one country in South-East Asia, two countries in East/Southern Africa, and one country in West Africa. Consequently, our findings may not be generalizable to LMICs globally but are likely generalizable to similar countries. Second, because we limited this analysis to first ANC consultations, it restricted the sample size for this study. However, the surveys in Haiti and Malawi were a census of facilities and the sample size for Tanzania was the largest of the surveys, thus resulting in a sufficient number of observations. Third, in Malawi and Nepal, the number of facilities with readiness less than 50 was small for certain index constructions (i.e., Nepal and Malawi expert simple n = 15). Our analysis in these countries and for these index constructions may be susceptible to outliers, potentially influencing our identification of an inflection point. Fourth, this analysis was limited to ANC services and may not be representative of primary health care services more broadly. Future research to investigate this association across the continuum of care for women and children would help to build evidence across the health system. Finally, while we controlled for individual, health worker, and facility-level characteristics, there is potential for residual uncontrolled confounding due to unmeasured covariates. However, the SPA collects a range of facility, health worker, and individual characteristics and the covariates selected are likely representative of the most influential factors.

Conclusions

The results of this study suggest that facility readiness is not a proxy for provision of care, but that there is a statistically significant and meaningful association between facility readiness and provision of care, consistent with the relationship postulated in quality of care frameworks. Our results also suggest a minimum level of facility readiness is needed in order to deliver high quality ANC services in five LMICs. Collecting data on facility readiness is necessary for understanding the foundations of health systems and for quality improvement efforts particularly in countries with the lowest levels of service quality. However, a comprehensive view of quality of care should include both facility readiness and provision of care measures.

Data Availability

The SPA datasets analyzed during the current study are available in the DHS repository, [https://dhsprogram.com/data/available-datasets.cfm] [23].

Abbreviations

ANC:

antenatal care

DHS:

Demographic and Health Surveys

EmONC:

Emergency Obstetric and Newborn Care

HFA:

health facility assessment

LMICs:

low-income and middle-income countries

PCA:

principal component analysis

QIQ:

Quick Investigation of Quality

SARA:

Service Availability and Readiness Assessment

SDG:

Sustainable Development Goal

SPA:

Service Provision Assessment

UHC:

universal health coverage

References

  1. Millennium Development Goals http://www.un.org/millenniumgoals/.

  2. Sustainable Development Goals. https://www.un.org/sustainabledevelopment/sustainable-development-goals/.

  3. Kruk ME, Gage AD, Arsenault C, Jordan K, Leslie HH, Roder-DeWan S, Adeyi O, Barker P, Daelmans B, Doubova SV, et al. High-quality health systems in the Sustainable Development Goals era: time for a revolution. Lancet Glob Health. 2018;6(11):e1196–252.

    Article  PubMed  PubMed Central  Google Scholar 

  4. World Health Organization, Organisation for Economic Co-operation and Development, The World Bank. Delivering quality health services: a global imperative for universal health coverage. Geneva, Switzerland: 2018. Available: http://documents.worldbank.org/curated/en/482771530290792652/pdf/127816-REVISED-quality-joint-publication-July2018-Complete-vignettes-ebook-L.pdf.

  5. Callaghan-Koru JA, McMahon SA, Chebet JJ, Kilewo C, Frumence G, Gupta S, Stevenson R, Lipingu C, Baqui AH, Winch PJ. A qualitative exploration of health workers’ and clients’ perceptions of barriers to completing four antenatal care visits in Morogoro Region, Tanzania. Health Policy Plann. 2016;31(8):1039–49.

    Article  Google Scholar 

  6. Gross K, Schellenberg Armstrong J, Kessy F, Pfeiffer C, Obrist B. Antenatal care in practice: an exploratory study in antenatal care clinics in the Kilombero Valley, south-eastern Tanzania. BMC Pregnancy Childbirth. 2011;11:36:1–11.

    Article  Google Scholar 

  7. Gupta S, Yamada G, Mpembeni R, Frumence G, Callaghan-Koru JA, Stevenson R, Brandes N, Baqui AH. Factors associated with four or more antenatal care visits and its decline among pregnant women in Tanzania between 1999 and 2010. PLoS ONE 2014, 9(7).

  8. Donabedian A. The quality of care: how can it be assessed? JAMA. 1988;260(12):1743–3.

    Article  PubMed  CAS  Google Scholar 

  9. Donabedian A. Evaluating the quality of medical care. Milbank Q. 2005;83(4):691–729.

    Article  PubMed  PubMed Central  Google Scholar 

  10. Akachi Y, Kruk ME. Quality of care: measuring a neglected driver of improved health. Bull World Health Organ. 2017;95(6):465–72.

    Article  PubMed  Google Scholar 

  11. O’Neill K, Takane M, Sheffel A, Abou-Zahr C, Boerma T. Monitoring service delivery for universal health coverage: the service availability and readiness Assessment. Bull World Health Organ. 2013;91(12):923–31.

    Article  PubMed  PubMed Central  Google Scholar 

  12. Sheffel A, Karp C, Creanga AA. Use of Service Provision assessments and service availability and readiness assessments for monitoring quality of maternal and newborn health services in low-income and middle-income countries. BMJ Global Health 2018, 3(6).

  13. The DHS, Program SPA, Overview. http://dhsprogram.com/What-We-Do/Survey-Types/SPA.cfm.

  14. Service availability and readiness assessment (SARA). https://www.who.int/data/data-collection-tools/service-availability-and-readiness-assessment-(sara)?ua=1.

  15. MEASURE Evaluation. Quick Investigation of Quality: A User’s Guide for Monitoring Quality of Care in Family Planning. Publication. Chapel Hill, North Carolina: MEASURE Evaluation, University of North Carolina, 2016. Available: https://www.measureevaluation.org/resources/publications/ms-15-104.

  16. Averting Maternal Death and Disability (AMDD). Needs Assessment of Emergency Obstetric and Newborn Care: Data Collection Modules. Columbia University Mailman School of Public Health: 2010. Available: https://www.mailman.columbia.edu/research/averting-maternal-death-and-disability-amdd/toolkit#toolkit.

  17. Kak N, Burkhalter B, Cooper M-A. Measuring the Competence of Healthcare Providers. Bethesda, MD: Published for the U.S. Agency for International Development (USAID) by the Quality Assurance (QA) Project, 2001. Available: https://www.hciproject.org/sites/default/files/Measuring%20the%20Competence%20of%20HC%20Providers_QAP_2001.pdf.

  18. Holla A, The World Bank Group. Measuring the quality of health care in clinics., 2013. Available: https://www.globalhealthlearning.org/sites/default/files/page-files/Measuring%20Quality%20of%20Health%20Care_020313.pdf.

  19. Leslie HH, Sun Z, Kruk ME. Association between infrastructure and observed quality of care in 4 healthcare services: a cross-sectional study of 4,300 facilities in 8 countries. PLoS Med. 2017;14(12):e1002464.

    Article  PubMed  PubMed Central  Google Scholar 

  20. Kruk ME, Leslie HH, Verguet S, Mbaruku GM, Adanu RMK, Langer A. Quality of basic maternal care functions in health facilities of five african countries: an analysis of national health system surveys. The Lancet Global Health. 2016;4(11):e845–55.

    Article  PubMed  Google Scholar 

  21. Arsenault C, Yakob B, Tilahun T, Nigatu TG, Dinsa G, Woldie M, Kassa M, Berman P, Kruk ME. Patient volume and quality of primary care in Ethiopia: findings from the routine health information system and the 2014 Service Provision Assessment survey. BMC Health Serv Res. 2021;21(1):485.

    Article  PubMed  PubMed Central  Google Scholar 

  22. Do M, Wang W, Hembling J, Ametepi P. Quality of antenatal care and client satisfaction in Kenya and Namibia. Int J Qual Health Care. 2017;29(2):183–93.

    Article  PubMed  Google Scholar 

  23. The DHS Program. - Available Datasets https://dhsprogram.com/data/available-datasets.cfm.

  24. Institut Haïtien de l’Enfance, ICF International. Évaluation de Prestation des Services de Soins de Santé, Haïti, 2013. 2014. Available: https://dhsprogram.com/pubs/pdf/SPA19/SPA19.pdf.

  25. Malawi Ministry of Health, ICF International. Malawi Service Provision Assessment (MSPA) 2013–2014. MoH and, International ICF. 2014. Available: https://www.dhsprogram.com/pubs/pdf/SPA20/SPA20.pdf.

  26. Nepal Ministry of Health, New, Nepal ERA. Nepal Health Sector Support Program (NHSSP), ICF. Nepal Health Facility Survey 2015. Kathmandu, Nepal: Ministry of Health, Nepal, 2015. Available: https://dhsprogram.com/pubs/pdf/SPA24/SPA24.pdf.

  27. Agence Nationale de la Statistique et de la Démographie (ANSD). Sénégal, ICF. Sénégal: Enquête Continue sur la Prestation des Services de Soins de Santé (ECPSS) 2016. ANSD and ICF: Rockville, Maryland, USA, 2016. Available: https://dhsprogram.com/pubs/pdf/SPA26/SPA26.pdf.

  28. Tanzania Ministry of Health & Social Welfare, Tanzania National Bureau of Statistics, ICF International. Tanzania Service Provision Assessment Survey 2014–2015. 2016. Available: https://dhsprogram.com/pubs/pdf/spa22/spa22.pdf.

  29. Sheffel A, Zeger S, Heidkamp R, Munos MK. Development of summary indices of antenatal care service quality in Haiti, Malawi and Tanzania. BMJ Open 2019, 9(12).

  30. Mallick L, Temsah G, Wang W. Comparing summary measures of quality of care for family planning in Haiti, Malawi, and Tanzania. PLoS ONE. 2019;14(6):e0217547.

    Article  PubMed  PubMed Central  CAS  Google Scholar 

  31. Stierman EK, Ahmed S, Shiferaw S, Zimmerman LA, Creanga AA. Measuring facility readiness to provide childbirth care: a comparison of indices using data from a health facility survey in Ethiopia. BMJ Global Health. 2021;6(10):e006698.

    Article  PubMed  PubMed Central  Google Scholar 

  32. Efron B, Tibshirani RJ. An introduction to the bootstrap. CRC press; 1994.

  33. Field CA, Welsh AH. Bootstrapping clustered data. J Royal Stat Soc Ser B: Stat Methodol. 2007;69(3):369–90.

    Article  Google Scholar 

  34. R Core Team. In: Vienna, editor. R: a Language and Environment for Statistical Computing. Austria: R Foundation for Statistical Computing; 2022.

    Google Scholar 

  35. Gelman A. Struggles with Survey Weighting and Regression modeling. Stat Sci. 2007;22(2):153–64.

    Google Scholar 

  36. Kish L, Frankel MR. Inference from Complex samples. J Royal Stat Soc Ser B (Methodological). 1974;36(1):1–37.

    Google Scholar 

  37. Korn EL, Graubard BI. Analysis of large health surveys: accounting for the Sampling Design. J Royal Stat Soc Ser (Statistics Society). 1995;158(2):263–95.

    Article  Google Scholar 

  38. Rao JNK, Bellhouse DR. History and development of the theoretical foundations of survey based estimation and analysis. Surv Methodol. 1990;16(1):3–29.

    Google Scholar 

  39. Ng M, Fullman N, Dieleman J, Flaxman A, Murray C, Lim S. Effective coverage: a metric for monitoring Universal Health Coverage. PLoS Med 2014, 11(9).

  40. Leslie HH, Malata A, Ndiaye Y, Kruk ME. Effective coverage of primary care services in eight high-mortality countries. BMJ Glob Health. 2017;2(3):e000424.

    Article  PubMed  PubMed Central  Google Scholar 

  41. Kim Mk, Kim SA, Oh J, Kim CE, Arsenault C. Measuring effective coverage of maternal and child health services in Cambodia: a retrospective analysis of demographic and health surveys from 2005 to 2014. BMJ Open. 2022;12(9):e062028.

    Article  PubMed  Google Scholar 

  42. Hategeka C, Arsenault C, Kruk ME. Temporal trends in coverage, quality and equity of maternal and child health services in Rwanda, 2000–2015. BMJ Global Health. 2020;5(11):e002768.

    Article  PubMed  PubMed Central  Google Scholar 

  43. Nguyen PH, KhÆ°Æ¡ng LQ, Pramanik P, Billah SM, Menon P, Piwoz E, Leslie HH. Effective coverage of nutrition interventions across the continuum of care in Bangladesh: insights from nationwide cross-sectional household and health facility surveys. BMJ Open. 2021;11(1):e040109.

    Article  PubMed  Google Scholar 

  44. Wang W, Mallick L, Allen C, Pullum T. Effective coverage of facility delivery in Bangladesh, Haiti, Malawi, Nepal, Senegal, and Tanzania. PLoS ONE. 2019;14(6):e0217853.

    Article  PubMed  PubMed Central  CAS  Google Scholar 

  45. Kanyangarara M, Chou VB. Linking household surveys and health facility assessments to estimate intervention coverage for the lives Saved Tool (LiST). BMC Public Health. 2017;17(Suppl 4):780.

    Article  PubMed  PubMed Central  Google Scholar 

  46. Carter ED, Ndhlovu M, Eisele TP, Nkhama E, Katz J, Munos M. Evaluation of methods for linking household and health care provider data to estimate effective coverage of management of child illness: results of a pilot study in Southern Province, Zambia. J Global Health 2018, 8(1).

  47. Willey B, Waiswa P, Kajjo D, Munos M, Akuze J, Allen E, Marchant T. Linking data sources for measurement of effective coverage in maternal and newborn health: what do we learn from individual- vs ecological-linking methods? J Global Health. 2018;8(1):010601.

    Article  Google Scholar 

  48. Exley J, Gupta PA, Schellenberg J, Strong KL, Requejo JH, Moller AB, Moran AC, Marchant T. A rapid systematic review and evidence synthesis of effective coverage measures and cascades for childbirth, newborn and child health in low- and middle-income countries. J Global Health. 2022;12:04001.

    Article  Google Scholar 

  49. Marsh AD, Muzigaba M, Diaz T, Requejo J, Jackson D, Chou D, Cresswell JA, Guthold R, Moran AC, Strong KL, et al. Effective coverage measurement in maternal, newborn, child, and adolescent health and nutrition: progress, future prospects, and implications for quality health systems. Lancet Glob Health. 2020;8(5):e730–6.

    Article  PubMed  PubMed Central  Google Scholar 

  50. Amouzou A, Leslie HH, Ram M, Fox M, Jiwani SS, Requejo J, Marchant T, Munos MK, Vaz LME, Weiss W, et al. Advances in the measurement of coverage for RMNCH and nutrition: from contact to effective coverage. BMJ Glob Health. 2019;4(Suppl 4):e001297.

    Article  PubMed  PubMed Central  Google Scholar 

  51. Do M, Micah A, Brondi L, Campbell H, Marchant T, Eisele T, Munos M. Linking household and facility data for better coverage measures in reproductive, maternal, newborn, and child health care: systematic review. J Global Health. 2016;6(2):020501.

    Article  Google Scholar 

  52. Donabedian A. The definition of Quality and Approaches to its Assessment vol 1. Explorations in Quality Assessment and Monitoring. Ann Arbor, Michigan: Health Administration Press; 1980.

    Google Scholar 

  53. Lagarde M, Huicho L, Papanicolas I. Motivating provision of high quality care: it is not all about the money. BMJ. 2019;366:l5210.

    Article  PubMed  PubMed Central  Google Scholar 

  54. Das J, Woskie L, Rajbhandari R, Abbasi K, Jha A. Rethinking assumptions about delivery of healthcare: implications for universal health coverage. BMJ 2018, 361.

  55. Das J, Hammer J, Leonard K. The quality of medical advice in low-income countries. J Economic Perspect. 2008;22(2):93–114.

    Article  Google Scholar 

  56. Das J, Hammer J. Quality of Primary Care in Low-Income Countries: facts and economics. Annual Rev Econ. 2014;6(1):525–53.

    Article  Google Scholar 

  57. Lembani M, de Pinho H, Delobelle P, Zarowsky C, Mathole T, Ager A. Understanding key drivers of performance in the provision of maternal health services in eastern cape, South Africa: a systems analysis using group model building. BMC Health Serv Res. 2018;18(1):912.

    Article  PubMed  PubMed Central  Google Scholar 

  58. Rowe AK, de Savigny D, Lanata CF, Victora CG. How can we achieve and maintain high-quality performance of health workers in low-resource settings? The Lancet. 2005;366(9490):1026–35.

    Article  Google Scholar 

  59. Leonard KL, Masatu MC. The use of direct clinician observation and vignettes for health services quality evaluation in developing countries. Soc Sci Med. 2005;61(9):1944–51.

    Article  PubMed  Google Scholar 

  60. Mace KE, Gueye AS, Lynch MF, Tassiba EM, Rowe AK. An evaluation of methods for assessing the quality of case management for inpatients with malaria in Benin. Am J Trop Med Hyg. 2014;91(2):354–60.

    Article  PubMed  PubMed Central  Google Scholar 

  61. Hermida J, Nicholas DD, Blumenfeld SN. Comparative validity of three methods for assessment of the quality of primary health care. Int J Qual Health Care. 1999;11(5):429–33.

    Article  PubMed  CAS  Google Scholar 

  62. Norman G, Neufeld V, Walsh A, Woodward C, McConvey G. Measuring physicians’ performances by using simulated patients. J Med Educ. 1985;60(12):925–34.

    PubMed  CAS  Google Scholar 

Download references

Acknowledgements

The authors wish to acknowledge the Bill & Melinda Gates Foundation for their support of this project.

Funding

This work was supported by the Improving Measurement and Program Design grant (OPP1172551) from the Bill & Melinda Gates Foundation. The funding agency had no role in the design of the study, analysis, interpretation of data, or writing of the manuscript.

Author information

Authors and Affiliations

Authors

Contributions

All authors contributed to conceptualizing the paper and analysis. AS prepared the manuscript. MKM, EC, and SZ critically reviewed and revised the manuscript.

Corresponding author

Correspondence to Ashley Sheffel.

Ethics declarations

Ethics approval and consent to participate

The Institutional Review Board at the Johns Hopkins Bloomberg School of Public Health determined this analysis to be exempt from human subjects review.

Consent for publication

Not applicable.

Competing interests

The authors declare that they have no competing interests.

Additional information

Publisher’s Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Electronic supplementary material

Below is the link to the electronic supplementary material.

12913_2023_10106_MOESM1_ESM.pdf

Supplementary Material 1: Additional File 1: This additional file contains supplementary text, tables, and figures referenced in the main manuscript.

Rights and permissions

Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/. The Creative Commons Public Domain Dedication waiver (http://creativecommons.org/publicdomain/zero/1.0/) applies to the data made available in this article, unless otherwise stated in a credit line to the data.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Sheffel, A., Carter, E., Zeger, S. et al. Association between antenatal care facility readiness and provision of care at the client level and facility level in five low- and middle-income countries. BMC Health Serv Res 23, 1109 (2023). https://doi.org/10.1186/s12913-023-10106-5

Download citation

  • Received:

  • Accepted:

  • Published:

  • DOI: https://doi.org/10.1186/s12913-023-10106-5

Keywords