Article Text
Abstract
Objectives The extent to which care quality influenced outcomes for patients hospitalised with COVID-19 is unknown. Our objective was to determine if prepandemic hospital quality is associated with mortality among Medicare patients hospitalised with COVID-19.
Design This is a retrospective observational study. We calculated hospital-level risk-standardised in-hospital and 30-day mortality rates (risk-standardised mortality rates, RSMRs) for patients hospitalised with COVID-19, and correlation coefficients between RSMRs and pre-COVID-19 hospital quality, overall and stratified by hospital characteristics.
Setting Short-term acute care hospitals and critical access hospitals in the USA.
Participants Hospitalised Medicare beneficiaries (Fee-For-Service and Medicare Advantage) age 65 and older hospitalised with COVID-19, discharged between 1 April 2020 and 30 September 2021.
Intervention/exposure Pre-COVID-19 hospital quality.
Outcomes Risk-standardised COVID-19 in-hospital and 30-day mortality rates (RSMRs).
Results In-hospital (n=4256) RSMRs for Medicare patients hospitalised with COVID-19 (April 2020–September 2021) ranged from 4.5% to 59.9% (median 18.2%; IQR 14.7%–23.7%); 30-day RSMRs ranged from 12.9% to 56.2% (IQR 24.6%–30.6%). COVID-19 RSMRs were negatively correlated with star rating summary scores (in-hospital correlation coefficient −0.41, p<0.0001; 30 days −0.38, p<0.0001). Correlations with in-hospital RSMRs were strongest for patient experience (−0.39, p<0.0001) and timely and effective care (−0.30, p<0.0001) group scores; 30-day RSMRs were strongest for patient experience (−0.34, p<0.0001) and mortality (−0.33, p<0.0001) groups. Patients admitted to 1-star hospitals had higher odds of mortality (in-hospital OR 1.87, 95% CI 1.83 to 1.91; 30-day OR 1.46, 95% CI 1.43 to 1.48) compared with 5-star hospitals. If all hospitals performed like an average 5-star hospital, we estimate 38 000 fewer COVID-19-related deaths would have occurred between April 2020 and September 2021.
Conclusions Hospitals with better prepandemic quality may have care structures and processes that allowed for better care delivery and outcomes during the COVID-19 pandemic. Understanding the relationship between pre-COVID-19 hospital quality and COVID-19 outcomes will allow policy-makers and hospitals better prepare for future public health emergencies.
- COVID-19
- Hospitals
- Quality in health care
- Health policy
- INFECTIOUS DISEASES
Data availability statement
Star rating data and COVID-19 utilisation data are available to the public through CMS. CMS claims data to calculate the COVID-19 outcome can be obtained through a third party.
This is an open access article distributed in accordance with the Creative Commons Attribution Non Commercial (CC BY-NC 4.0) license, which permits others to distribute, remix, adapt, build upon this work non-commercially, and license their derivative works on different terms, provided the original work is properly cited, appropriate credit is given, any changes made indicated, and the use is non-commercial. See: http://creativecommons.org/licenses/by-nc/4.0/.
Statistics from Altmetric.com
STRENGTHS AND LIMITATIONS OF THIS STUDY
Our study includes data for more than a million patients and four thousand hospitals.
Our study compared hospital quality before the pandemic with risk-standardised COVID-19 outcomes.
Sensitivity analyses did not refute the results of our study.
Claims data are limited by proper coding practices.
Claims data could not be used to assess the impact of vaccination.
Introduction
Prior to the COVID-19 pandemic, and despite struggles to respond to an earlier influenza epidemic, hospitals likely did not prioritise preparation for a future pandemic.1 This lack of adequate preparation may have contributed in part to the death of more than a million people in the USA alone. As the WHO and the USA, and Europe consider the risk of future pandemics,2 it is important to understand how to identify hospitals in need of better preparedness for future public health emergencies.
During normal operations, high-quality hospitals can deliver evidence-based, timely, patient-centred and equitable care when adequately staffed with high-quality workers who can support good communication.3 High-quality hospitals have better patient outcomes, including lower risk-standardised mortality rates (RSMRs), for specific conditions (such as pneumonia and heart failure) and specific procedures (such as heart surgery), and evidence shows that care quality for one condition is associated with care quality for other conditions.4 Therefore, during normal operations, structures and processes of care may transfer across teams and patients, however, we do not know if this is true during a major stressor such as a pandemic. During a pandemic, resilient hospitals may be able to continue to deliver high quality care despite the stressor. Research suggests that some of the same characteristics associated with high quality during normal operations, such as communication and adherence to evidence-based processes, are also associated with readiness/resilience.5–7 We, therefore, hypothesised that prepandemic hospital quality could be a marker of hospital readiness/resilience, and that hospitals with higher quality prior to the pandemic would be more likely to be able to respond to the pandemic and translate the same structures and processes across care teams and patients, resulting in better patient outcomes. To test this hypothesis, we first developed a measure of hospital response to the pandemic (ability to deliver high-quality care as measured by patient outcomes), by calculating hospital-level risk-standardised COVID-19 mortality rates among patients hospitalised with COVID-19 (COVID-19 RSMRs). We then explored the relationship between a marker of pre-COVID-19 hospital quality (the hospital summary score used to calculate the Centers for Medicare and Medicaid Services (CMS) Overall Hospital Quality Star Rating—hereafter ‘star rating summary score’—and its components) and COVID-19 RSMRs. We stratified the association between pre-COVID-19 hospital quality and COVID-19 outcomes by hospital characteristics and explored the relationship between COVID-19 outcomes and hospital COVID-19 burden.
We acknowledge that there is no gold standard for what defines a ‘high-quality’ hospital in quantitative terms. While there are existing and accepted quality frameworks, such as the US Institute of Medicine’s six pillars of hospital quality (safety, effectiveness, patient-centredness, timeliness, efficiency and equity),3 quality measures within these domains can differ, and there is no one accepted ‘gold standard’ on which individual measures to include or how to combine them into an overall quantitative assessment of hospital quality.8 We, therefore, used a publicly available and publicly vetted definition of overall hospital quality (CMS Overall Hospital Quality Star Rating) as the basis for our study.9
Methods
Data sources
We used administrative claims data from the CMS that included patients diagnosed with COVID-19 who were admitted to hospitals in the USA and its territories between 1 April 2020 and 30 September 2021. We used the CMS Provider of Services files to obtain hospital characteristics,10 including the urban/rural definition that is based on the US Office of Budget and Management (OMB) definition that designates urban counties as metropolitan (a county containing a core urban area of 50 000 or more population) and micropolitan (a county containing a core urban core of at least 10 000 (but less than 50 000) population. ‘Rural’ encompasses all population, housing and territory not included within an urban area.11 12 To examine the impact of the level of hospital ‘COVID-19 burden’ on these results, we used hospital-reported data provided to the public by the US Department of Health and Human Services.13
Study cohort
We examined all Medicare Fee-for-Service (FFS) and Medicare Advantage (MA) hospital-submitted inpatient admission claims with a principal or secondary (present on admission) discharge diagnosis of COVID-19 (International Classification of Diseases, ICD-10 code U07.1) for patients discharged from an acute care or critical access hospital (CAH) between 1 April 2020 and 30 September 2021.
Measures of pre-COVID-19 hospital quality and COVID-19 burden
To characterise pre-COVID-19 hospital quality, we used the star rating summary score (April 2021 release) used to calculate CMS’ Overall Hospital Quality Star Rating, which summarises existing hospital quality information by assigning hospitals 1–5 stars based on their performance in measures within groups (online supplemental appendix, figure 1A). Version 4.0 categorises measures into 5 groups—mortality (7 measures), readmission (11), safety of care (8), patient experience (8) and timely and effective care.14 Summary scores are calculated using a weighted average of group scores and star rating categories are assigned based on hospitals’ summary scores. All quality measures included in this analysis used performance data prior to 2020 (online supplemental appendix, table 1A). Hospitals with insufficient data for a star rating are not included in the analyses of associations. We calculated hospital COVID-19 burden as the weekly average number of laboratory-confirmed COVID-19 hospitalisations for all adult patients (not limited to Medicare patients) divided by the number of hospital beds.
Supplemental material
Outcome assessment
We calculated hospital-level risk-standardised mortality rates (RSMRs) for COVID-19 patients (in-hospital and within 30 days from the date of admission) using hierarchical logistic regression models.15–18 The models adjust for components of the Charlson Comorbidity Index, including age (online supplemental appendix, table 2A).19 The commonly used Charlson Comorbidity Index calculates a risk score for each patient using 19 ICD diagnosis codes from administrative data. Hospital RSMRs are the ratio of a hospital’s ‘predicted’ to ‘expected’ mortality, multiplied by the national observed mortality rate. The approach simultaneously models data at patient and hospital levels to account for the variation in mortality within and between hospitals.20
Statistical analyses
We calculated volume-weighted Pearson correlations to evaluate associations between hospital-level RSMRs and pre-COVID-19 quality (star rating summary and group scores), and stratified correlations by hospital characteristics. We calculated these associations for each hospital with a star rating summary score, and for each hospital with a group score (see the Results section for the number of hospitals in each category). For sensitivity analyses, we limited our sample to hospitals with ≥25 COVID-19 patients and recalculated results after removing hospitals with the 20 highest and lowest 30-day RSMRs (based on the distribution of outliers) to explore the impact of COVID-19 on RSMR outliers. In addition, we repeated the analyses limiting the data period to the early pandemic (from March 2020 to September 2020) to assess if associations between star rating summary scores and COVID-19 mortality rates differed earlier in the pandemic. We then examined correlations between COVID-19 RSMRs and the star rating summary score and each of its components (group scores), as well as between COVID-19 RSMRs and COVID-19 burden, calculated as the weekly average number of laboratory-confirmed COVID-19 hospitalisations for all patients (not limited to Medicare patients) divided by the number of hospital beds. To estimate the number of deaths that might be attributable to care in a lower-quality hospital, we applied the mean COVID-19 RSMR for hospitals within the 5-star ratings category to the total number of patients admitted to the hospital with COVID-19 between 1 April 2020 and 30 September 2021, and subtracted that value from the total (observed) number of patients admitted with COVID-19 who died within 30 days.
CMS assigns a star rating to hospitals that report 3, 4 or 5 measure groups (hospitals are peer-grouped prior to k-means clustering and then are assigned a star rating; see online supplemental appendix).21 To examine the impact of the number of group scores hospitals reported to CMS and our observations, we recalculated correlation coefficients after stratifying hospitals by their number of reported group scores (3, 4 or 5 measure groups). To examine the impact of COVID-19 hospitalisation volume we examined results for hospitals with at least 25 COVID-19 patients. As a sensitivity analysis to determine the impact of outliers on the observed associations, we recalculated correlation coefficients (among all hospitals) after removing hospitals with the 20 highest and 20 lowest 30-day COVID-19 RSMRs, and the 20 highest and 20 lowest star rating summary scores. To examine the adequacy of risk adjustment using the CCI, we calculated the c-statistic for both in-hospital and 30-day mortality models.
All analyses used SAS Enterprise Guide and SAS V.9.4 and were performed by two authors (S-XL and YW).
Results
Variation in hospital-level COVID-19 RSMRs and stratification by hospital characteristics
Between 1 April 2020 and 30 September 2021, 1 229 071 Medicare beneficiaries with a diagnosis of COVID-19 were admitted to 4343 US hospitals. Among those admitted patients, 230 358 (18.7%) died in the hospital, and 338 358 patients (27.5%) died within 30 days of admission. Patient characteristics are shown in table 1.
Patient characteristics (number and frequency of Charlson Comorbidity Index variables)
At the hospital level, we found striking variation in COVID-19 RSMRs. Among the 4343 hospitals with at least 1 COVID-19 patient, in-hospital RSMRs ranged from 4.5% to 59.9%; the median in-hospital RSMR was 18.2% (IQR 14.7%–23.7%). A 30-day RSMRs also varied widely, from 12.9% to 56.2% (IQR 24.6%–30.6%). Results were similar for hospitals with at least 25 cases.
In-hospital RSMRs differed by hospital characteristics. Mean in-hospital RSMRs were significantly (p<0.0001) higher in the following: urban hospitals (vs rural), hospitals with more (vs fewer) beds, teaching hospitals (vs non-teaching hospitals), hospitals not designated as CAHs (vs CAHs) and for-profit (vs non-profit or government owned) hospitals (table 2). Differences in mean in-hospital mortality rates between hospitals in different nurse-to-bed ratios were small. Differences in 30-day RSMRs by hospital characteristic were also small but statistically significant except for urban versus rural where the difference was not significant (table 2). Results were similar for hospitals with at least 25 cases.
Mean risk-adjusted COVID-19 RSMRs by hospital characteristics (for hospitals with ≥1 COVID-19 admission) between 1 April 2020 and 30 September 2021*
To determine if the level of COVID-19 burden might explain these results, we examined the association between hospital-level COVID-19 burden calculated using weekly hospital-reported COVID-19 utilisation data (see the Methods section) and COVID-19 RSMRs. We found, however, only a weak relationship: the Pearson correlation coefficient was −0.04, (p=0.01) for in-hospital RSMRs and −0.03 (p=0.03) for 30-day RSMRs.
Association between pre-COVID-19 hospital quality and hospital-level COVID-19 RSMRs
We examined how COVID-19 RSMRs differed by pre-COVID-19 hospital quality as defined by star rating category (1–5 stars). When hospitals were stratified by star rating category we found that in-hospital and 30-day RSMRs were systematically lower (better) with each increase in star rating category: For example, mean in-hospital RSMRs were 29.1% for 1-star hospitals (n=201) vs 18.0% for 5-star hospitals (n=449); mean 30-day RSMRs were 32.1% for 1-star hospitals vs 24.5% for 5-star hospitals (table 2). Patients admitted to 1-star hospitals had higher odds of in-hospital (OR 1.87, 95% CI 1.54 to 1.62) and 30-day mortality (OR 1.46, 95% CI 1.31 to 1.39), compared with patients admitted to 5-star hospitals, after adjusting for clinical characteristics using the Charlson Comorbidity Index which includes age (see online supplemental appendix table, 2A).
We then calculated Pearson correlation coefficients between pre-COVID-19 star rating performance (summary scores and the five individual measure group scores) and COVID-19 RSMRs with April 2020–September 2021 data, near the peak of the Delta variant wave in the USA. Star rating summary scores among 4256 hospitals in our analysis were moderately inversely correlated with in-hospital (−0.41, p<0.0001) and 30-day (−0.38, p<0.0001) RSMRs (figure 1). For in-hospital RSMRs, the patient experience and timely and effective care group scores showed the strongest associations (−0.39, p<0.001; −30, p<0.0001, respectively). For 30-day RSMRs, patient experience and mortality group scores showed the strongest associations (−0.34, p<0.0001; −0.33, p<0.001, respectively). When we limited our analyses of the associations between star rating summary scores and RSMRs in the early pandemic period (March 2020–September 2020), the relationship between 30-day RSMRs and star rating mortality group scores was weaker (Pearson correlation coefficient, −0.12, p<0.0001) compared with the 18-month period of this study (Pearson correlation coefficient −0.34, p<0.0001) (data not shown).
Volume-weighted correlations between pre-COVID-19 hospital quality star rating summary scores and in-hospital and 30-day risk-standardised COVID-19 mortality rates (RSMRs). Blue bars represent correlation coefficients for in-hospital COVID-19 RSMRs and orange bars represent correlations for 30-day RSMRs. All p values are <0.0001 except where indicated *not significant (p=0.07). A number of hospitals that qualified for analysis in each category are as follows: summary score (“overall group score”), n=4256; mortality group score n=3934; readmission group score, n=4182; safety of care group score, n=3401; patient experience group score, n=3198; timely and effective care group score, n=4202. RSMRs, risk-standardised mortality rates.
In stratified analyses by hospital characteristics, stronger correlations were seen between 30-day hospital COVID-19 RSMRs and the star rating summary score and its component group scores for: larger versus smaller bed-size hospitals (−0.43 for hospitals with 400+ beds vs −0.22 for hospitals with 1–99 beds), hospitals with academic affiliation versus without (−0.46 vs −0.32), hospitals in urban versus rural locations (−0.41 vs −0.21, respectively), government and not-for-profit hospitals versus for-profit (−0.48 and −0.38 vs −0.12, respectively) and non-CAH versus CAH (−0.39 vs −0.13, respectively); differences by nurse-to-bed ratio categories were small (table 3). Differences by hospital characteristic for in-hospital RSMRs were generally smaller compared with observations for 30-day RSMRs (table 4).
Hospital characteristics and associations between star rating summary scores and 30-day COVID-19 RSMRs (values represent Pearson correlation coefficients between star rating summary and group scores and hospital-level 30-day COVID-19 RSMRs).
Hospital characteristics and associations between star rating summary scores and in-hospital COVID-19 RSMRs (values represent Pearson correlation coefficients between star rating summary and group scores and in-hospital COVID-19 RSMRs).
To address concerns that hospitals reporting 3 vs 4 vs 5 measure groups may have differing hospital quality, and possibly confound the results of this study, we recalculated correlation coefficients after stratifying hospitals by their number of reported group scores. We found that while the strength of the association between the summary score and the 3-group strata (−0.27) is somewhat weaker compared with the 5-group strata (−0.39), the relationships are statistically significant for all three strata, and we also found the same pattern of associations between all three strata and all of the individual group scores (see table 3A, online supplemental appendix). We also note that about three fourths of hospitals that receive a star rating report five measure groups. To examine the impact of COVID-19 hospitalisation volume, we examined results for hospitals with at least 25 COVID-19 patients; we found that among the 3405 hospitals that met these criteria, results were similar (data not shown) to results with hospitals with at least one COVID-19 hospitalisation. As a sensitivity analysis to determine the impact of outliers on the observed associations, we recalculated correlation coefficients (among all hospitals) after removing hospitals with the 20 highest and 20 lowest 30-day COVID-19 RSMRs, and the 20 highest and 20 lowest star rating summary scores and found the correlation was similar (−0.40, p<0.0001, n=4196 hospitals). Finally, the c-statistics for the in-hospital and 30-day mortality models were 0.609 and 0.663, respectively, demonstrating adequate risk adjustment for the purposes of this study.
Discussion
Using data from a representative sample of more than 1.2 million COVID-19-associated hospitalisations of Medicare beneficiaries across more than 4300 hospitals, we found that risk-standardised 30-day mortality rates were significantly associated with pre-COVID-19 hospital quality. Associations were stronger in quality domains associated with communication and the use of processes. A potential explanation for the observed association between pre-COVID-19 hospital quality and COVID-19 outcomes is that hospitals may have been able to transfer effective care structures and processes used during normal operations to the care of patients with COVID-19 during the pandemic. Pre-COVID-19 hospital quality may also reflect, at least in part, a hospital’s readiness/resilience to respond to stressors and provide high-quality care under stress. In our study, differences in hospital readiness, as measured by pre-COVID-19 hospital quality, had serious consequences; on average, a patient admitted to a lower-quality (1-star hospital) was 87% and 46% more likely to die in the hospital and within 30 days, respectively, compared with a patient admitted to a higher quality (5-star) hospital (absolute differences of 11 percentage points for in-hospital and 7.6 percentage points for 30-day mortality).
This study has some important strengths and limitations. The strengths of this study include that it represents COVID-19 outcomes from more than a million Medicare beneficiaries and hospital quality for more than four thousand hospitals across the USA. In addition, we calculated RSMRs to assess patient outcomes. Our study also used a comprehensive and publicly reported measure of hospital quality to assess pre-COVID-19 hospital readiness/resilience. We also examined, as a potential confounder, hospital-level COVID-19 burden.
This study has the limitations of any observational study, including that no direct causal relationship can be attributed to the associations between hospital quality and mortality rates for patients hospitalised with COVID-19. In addition, while RSMRs were adjusted for age and comorbidities, we did not include a time variable in the risk model, although we did examine associations during the early and later part of the pandemic and did not see marked differences except for the association with the pre-COVID-19 mortality group score and COVID-19 RSMRs, which was stronger in the later part of the pandemic. Because hospital-level COVID-19 burden became available starting in August 2020, we were not able to include it in the risk model. Therefore, while the results do not directly assess the confounding effect of COVID-19 burden on the associations between pre-COVID-19 hospital quality and COVID-19 RSMRs, we did examine the associations between hospital-level COVID-19 burden with both the outcome (COVID-19 RSMRs) and the exposure (pre-COVID-19 star rating). Because burden was not substantially related to either the exposure or outcome, we expect this variable would not be an important confounder in the associations. In addition, we were not able to explore the relationship between these observations and a patient’s vaccination status, due to lack of reliable patient-level data within claims; the ICD-10 vaccination status code became effective 1 April 2022. Furthermore, COVID-19 mortality rates were calculated with MA and Medicare FFS claims for patients aged 65 and older; most of the measures in star rating are based on Medicare FFS patients. Finally, while measures within star rating use data from 2016 to 2019, some measures are based on different time periods (some are 1-year measures, others are 3-year measures).21 However, within measure groups, measures have similar reporting timelines and most hospitals (74%) report all five measure groups, suggesting that comparisons are based on information that reflects the same quality signal. Furthermore, we have found associations between individual components (group scores) of pre-COVID-19 star rating and COVID-19 mortality.
Our results are, in part, consistent with and extend on the findings of other work examining drivers of mortality rates in patients admitted to ICUs at 70 hospitals between March and June 2020.22 Study authors found that at the patient level, while most of the variation in mortality (70%) was explained by the physiology of the patient at ICU admission, demographics (primarily age) and comorbidities, hospital quality (among other hospital factors) was also a contributing factor. The findings from our work expand this observation by examining hospital-level associations with quality not limited to the ICU, to all patients diagnosed with COVID-19 over an 18-month period for more than a million patients at over four thousand hospitals.
Our findings suggest that quality domains such as communication (represented by the patient experience group score), and quality domains tied more closely to processes and checklists (reflected within the timely and effective care and mortality group scores) are associated with better outcomes in patients hospitalised with COVID-19. During regular operations, the development of, and adherence to, evidence-based care processes that are tied to better outcomes is a hallmark of high-quality hospitals,23–26 and it is possible that hospitals that were able to rapidly translate those capabilities were better positioned to care for patients hospitalised with COVID-19.
The COVID-19 outcome variation and association found in this work cannot, however, likely be tied to any single care process (nor could this be determined in our study using the national data available). One study, however, found wide variation in adherence to Acute Respiratory Distress Syndrome (ARDS) protocols for patients with COVID-19 and while not statistically significant, hospitals with better protocol adherence had lower mortality rates.27 Concentrating expertise and processes in a single setting may have been an effective protocol; patients admitted to hospitals dedicated to the care of COVID-19 patients had better outcomes compared with hospitals that did not specialise.28
There are many other hospital-level factors that may have influenced even a prepared hospital’s ability to respond to the pandemic. For example, one study found that after controlling for other factors, ICU patients in hospitals with a higher proportion of patients with social risk factors had worse outcomes.22 In our study, we found that urban location, larger bed size, teaching affiliation and government or non-profit ownership had a stronger association between worse performance on star rating summary scores and higher 30-day COVID-19 RSMRs. Several of these characteristics are also associated with a larger proportion of patients with social risk factors but could also reflect differences in the geographical impact of COVID-19 over time. In addition, urban location, larger bed size and teaching affiliation are often overlapping characteristics, and urban areas were early pandemic hotspots. Another study, however, did not find an association between academic status, profit status or urban/non-urban setting and hospital COVID-19 outcomes during the first 6 months of the pandemic.29
Another potential explanation for our findings could be that hospitals with better quality during normal operations also have more care quality-independent resources (eg, financial resources) and those hospitals may have been able to pivot those resources to provide better care for patients with COVID-19 or to better care for their staff through purchase of supplies such as personal protective eqipment (PPE). If this were true, one might predict that if resources were limited, hospital performance would decline as the level of COVID-19 burden increased. However, there is mixed evidence (from this work and others) for the relationship between hospital-level COVID-19 mortality and measures of hospital and/or community level COVID-19 burden and differences between the association early versus later in the pandemic.22 29–33 In our study we found only a weak association between hospital-level mortality in patients hospitalised with COVID-19 and hospital-level COVID-19 burden, defined by the total number of hospitalised patients with COVID-19 divided by the number of hospital beds. Taken together, the evidence suggests that the capacities of hospitals to manage large patient loads may not have been a defining characteristic or may have been important mainly in the very early months of the pandemic. Future studies using additional measures (such as processes of care), additional data sources, including data from electronic health records and financial records, and data from multiple time points during and before the pandemic, may help tease out the underlying drivers of the associations between prepandemic quality and outcomes for patients hospitalised with COVID-19.
On a broader scale, the COVID-19 pandemic has exposed disparities not just between hospitals, states or regions, but in outcomes across the world. These disparities are driven by several different factors, including prepandemic healthcare system resilience/preparedness. For example, Haldane et al34 examined outcomes across 28 countries and characterised performance within a resilience/preparedness framework that includes healthcare service delivery and healthcare workforce (including the quality and quantity of the workforce), connected by two communication domains; communication across sectors (eg, government and healthcare) and engagement with the community. While overall, the authors did not identify a ‘silver bullet’ that characterised better outcomes (lower mortality per capita), they were able to identify core capabilities of higher-performing countries which parallel our findings in this study. For example, in parallel with the concept that higher quality hospitals may have had more resources that could be deployed to address COVID-19, higher performing countries (those with lower per-capita COVID-19 deaths) were found to have been well funded and could pivot their resources towards obtaining supplies, reallocating and training healthcare workers, and communicating with the public. Those well-funded and higher-performing systems were also better resourced to be able to continue to deliver primary care while addressing the surge of COVID-19 patients.
Conclusion
Across a national sample of hospitals, we found that prepandemic hospital quality is associated with COVID-19 hospitalisation outcomes suggesting that hospital quality for common care may be a marker of hospital readiness/resilience to respond to a stress/shock such as COVID-19. Hospitals with better prepandemic quality may have been able to better translate care structures or processes used during normal operations into better care for patients during the COVID-19 pandemic. These results can help policy-makers at local, national, and international levels plan for future challenges and can help hospital leadership assess their readiness/resilience for a future pandemic.
Data availability statement
Star rating data and COVID-19 utilisation data are available to the public through CMS. CMS claims data to calculate the COVID-19 outcome can be obtained through a third party.
Ethics statements
Patient consent for publication
Ethics approval
This study involves human participants and was approved by Yale School of Medicine IRB: Research on the Centers for Medicare and Medicaid Services (CMS) Data; Number: 0903004927. This is a retrospective study based on claims data (billing data) submitted by hospitals to the Centers for Medicare and Medicaid Services, as part of the billing/payment process.
Acknowledgments
The analyses on which this publication is based were performed under the Measure and Instrument Development and Support (MIDS) contract #HHSM-75FCMC18D0042, Task Order HHSM-75FCMC19F0001–Development, Reevaluation and Implementation of Outcome/Efficiency Measures for Hospitals and Eligible Clinicians, funded by the Centers for Medicare & Medicaid Services, an agency of the US Department of Health and Human Services. The authors assume full responsibility for the accuracy and completeness of the ideas presented. We acknowledge Steven Spivack, Sydnie Stackland and Demetri Goutos for their assistance in this work.
References
Supplementary materials
Supplementary Data
This web only file has been produced by the BMJ Publishing Group from an electronic file supplied by the author(s) and has not been edited for content.
Footnotes
Twitter @arjunvenkatesh
Contributors MS, S-XL, EWT, YW and DP contributed to the conception and design of the study; JZ provided assistance with data access and management; YW and S-XL performed all analyses. DP drafted the manuscript. EWT, AKV, JG and ZL contributed critical additions and conceptual revisions to the manuscript; LAF and LS assisted with interpretation of results; SB, EN and AKV were involved in the development of star ratings. DP serves as guarantor of the study.
Funding This work was supported by The Centers for Medicare and Medicaid Services (CMS), contract number HHSM-75FCMC18D0042.
Disclaimer The content of this publication does not necessarily reflect the views or policies of the US Department of Health and Human Services nor does the mention of trade names, commercial products or organisations imply endorsement by the US government.
Competing interests S-XL, YW, JZ, JG, KM, EN, ZL, AV, LS and EWT receive salary support from the Centers for Medicare and Medicaid Services to develop, implement and maintain hospital performance outcome measures, including the methodology for the Overall Hospital Star Ratings, that are publicly reported. SB and MS are employed by CMS. LAF is employed by the Perelman School of Medicine at the University of Pennsylvania. DP is a subcontractor to Yale/CORE.
Patient and public involvement Patients and/or the public were not involved in the design, or conduct, or reporting, or dissemination plans of this research.
Provenance and peer review Not commissioned; externally peer reviewed.
Supplemental material This content has been supplied by the author(s). It has not been vetted by BMJ Publishing Group Limited (BMJ) and may not have been peer-reviewed. Any opinions or recommendations discussed are solely those of the author(s) and are not endorsed by BMJ. BMJ disclaims all liability and responsibility arising from any reliance placed on the content. Where the content includes any translated material, BMJ does not warrant the accuracy and reliability of the translations (including but not limited to local regulations, clinical guidelines, terminology, drug names and drug dosages), and is not responsible for any error and/or omissions arising from translation and adaptation or otherwise.