Context In response to concerns about the quality of care in US hospitals, the Centers for Medicare & Medicaid Services began measuring hospital performance and reporting this performance on their Web site, Hospital Compare. It is unknown whether these process performance measures are related to hospital-level outcomes.
Objective To determine whether quality measured with the process measures used in Hospital Compare are correlated with and predictive of hospitals' risk-adjusted mortality rates.
Design, Setting, and Participants Cross-sectional study of hospital care between January 1 and December 31, 2004, for acute myocardial infarction, heart failure, and pneumonia at acute care hospitals in the United States included on the Hospital Compare Web site. Ten process performance measures included in Hospital Compare were compared with hospital risk-adjusted mortality rates, which were measured using Medicare Part A claims data.
Main Outcome Measures Condition-specific inpatient, 30-day, and 1-year risk-adjusted mortality rates.
Results A total of 3657 acute care hospitals were included in the study based on their performance as reported in Hospital Compare. Across all acute myocardial infarction performance measures, the absolute reduction in risk-adjusted mortality rates between hospitals performing in the 25th percentile vs those performing in the 75th percentile was 0.005 for inpatient mortality, 0.006 for 30-day mortality, and 0.012 for 1-year mortality (P<.001 for each comparison). For the heart failure performance measures, the absolute mortality reduction was smaller, ranging from 0.001 for inpatient mortality (P = .03) to 0.002 for 1-year mortality (P = .08). For the pneumonia performance measures, the absolute reduction in mortality ranged from 0.001 for 30-day mortality (P = .05) to 0.005 for inpatient mortality (P<.001). Differences in mortality rates for hospitals performing in the 75th percentile on all measures within a condition vs those performing lower than the 25th percentile on all reported measures for acute myocardial infarction ranged between 0.008 (P = .06) and 0.018 (P = .008). For pneumonia, the effects ranged between 0.003 (P = .09) and 0.014 (P<.001); for heart failure, the effects ranged between −0.013 (P = .06) and −0.038 (P = .45).
Conclusions Hospital performance measures predict small differences in hospital risk-adjusted mortality rates. Efforts should be made to develop performance measures that are tightly linked to patient outcomes.
It is widely recognized that the quality of health care in the United States is uneven and often inadequate.1,2 In the outpatient setting, quality of care varies across individuals depending on age, sex, race, and socioeconomic status.3 Overall, only half of US individuals receive recommended care.1 In hospitals, quality of care is also variable. Compliance with hospital performance measures varies not only across US hospitals but also across regions, conditions, and performance measures.4
Because it is assumed that measuring quality of care is a key component in improving care, quality measurement is playing an increasingly prominent role in quality improvement. For example, quality is measured and in many cases reported for hospitals,5-10 health plans,11 nursing homes,12 home health agencies,13 and physicians.14-16 These efforts are designed to provide an incentive to improve the quality of the care delivered and to influence consumer choice of providers.17,18 These measures also are increasingly being used to determine clinicians' reimbursement.19-21
Recently, the US Centers for Medicare & Medicaid Services (CMS), along with other health care organizations, began participating in the Hospital Quality Alliance, a large-scale public-private collaboration that seeks to make performance information on all acute care nonfederal hospitals accessible to the public, payers, and providers of care.22,23 These performance measures evaluate hospital quality on certain processes of care for patients with acute myocardial infarction (AMI), heart failure, pneumonia, and for surgical infection prevention. The Joint Commission on Accreditation of Healthcare Organizations (JCAHO), another Hospital Quality Alliance participant, also uses these measures, requiring hospitals to report their performance as a part of accreditation for most US hospitals.24 This comparative quality information is now available to the public, in slightly different forms, through the CMS Web site Hospital Compare (http://www.hospitalcompare.hhs.gov/) and the JCAHO Web site Quality Check (http://www.qualitycheck.org).
One way that hospital process performance measures might lead to improvements in health care quality is by providing consumers with the information needed to choose a hospital in which they will receive higher-quality care.17,18 However, like most performance measures, Hospital Compare captures information in a select group of patients with a given condition. Additionally, process performance measures capture information about only a small portion of the overall care delivered during a hospital stay. While some research has documented an association between higher adherence to care guidelines and better outcomes of patients who receive that care,25,26 to date there has been limited evidence demonstrating that hospitals that perform better on process measures also have better overall quality for the average patient. Our objective in this study was to determine whether quality measured with the process measures used in the CMS's Hospital Compare are correlated with and predictive of hospitals' risk-adjusted mortality rates. In other words, from a consumer's perspective, can a hospital's performance on process measures be used to choose a hospital in which patients can expect to have better outcomes?
Participation in Hospital Compare is voluntary and initially only one quarter of hospitals reported their performance data.27 However, the Medicare Modernization Act in 2003 introduced financial incentives for hospitals to report data on 10 performance measures to the CMS. Starting in 2004, hospitals that did not submit performance data for these measures experienced a reduction in the annual Medicare fee schedule update of 0.4 percentage points. After implementation of the Medicare Modernization Act, the proportion of hospitals reporting their performance on these measures grew to 98%.23,28 Seven additional measures are also included in Hospital Compare but are not tied to financial incentives and have significantly lower levels of reporting.28 The data from these 17 measures have been posted on the CMS's Hospital Compare Web site since April 2005 and are updated quarterly.
Process Performance Measures. We evaluated hospital performance based on publicly available data from the CMS on the original 10 process measures included in Hospital Compare. These measures evaluate quality of care for AMI, heart failure, and pneumonia between January 1 and December 31, 2004. Five of the measures assess quality of care for AMI: aspirin use within 24 hours of arrival, β-blocker use within 24 hours of arrival, angiotensin-converting enzyme inhibitor use for left ventricular dysfunction, aspirin prescribed at discharge, and β-blocker prescribed at discharge. Two of the measures assess quality of care for heart failure: assessment of left ventricular function and the use of an angiotensin-converting enzyme inhibitor for left ventricular dysfunction. Three of the measures assess quality of care for pneumonia: the timing of initial antibiotics, pneumococcal vaccination, and assessment of oxygenation within 24 hours of admission.
This evaluation is limited to the original 10 measures because hospital reporting rates are nearly universal. Reporting rates are much lower for the remaining 7 measures, ranging between 18% and 83%.28 For these 7 measures, a hospital's decision to report its data may be nonrandom23,29 (eg, hospitals that perform poorly may be less likely to report their performance) and hence the significant missing data would cause additional and challenging analyses that would require the inclusion of unverifiable assumptions about the selection process. While the JCAHO requires reporting of these measures for accreditation, diminishing concerns about missing data, the data from Hospital Compare were chosen instead because these data are from all US hospitals rather than only accredited hospitals.
For each of the 10 measures, a hospital's performance is calculated as the proportion of patients who received the indicated care out of all the patients who were eligible for the indicated care. All US acute care hospitals that participated in Hospital Compare during 2004 were included. To ensure the stability of the measures, hospitals with fewer than 25 patients in the denominator of a measure were excluded. This is the same convention that the CMS uses to report performance.
Hospital performance also was measured using 2 condition-specific performance measures. First, a composite measure was calculated by aggregating individual measures within conditions using a weighted average of performance across all measures.30 For example, the number of times measured AMI care was received at a hospital was divided by the number of times patients were eligible for all AMI measures. This composite measure was included because it is a metric of hospital performance that is currently used by the JCAHO30 and may provide more meaningful information about hospital performance than individual measures do.
Second, an “all-or-none” measure31 was developed, which identified both hospitals that performed well on all measures within a condition set and hospitals that performed poorly on all measures within a condition set. To do this, we first identified high-performing and low-performing hospitals, or hospitals that performed above the 75th percentile on every measure they reported and hospitals that performed below the 25th percentile on every measure they reported. This all-or-none measure was included because hospitals that perform well on all measures within a condition set may be of higher quality than hospitals that perform well on average, and therefore the relationship between performance and mortality rates may be easier to detect, if it exists.
Risk-Adjusted Mortality Rates. Using the Medicare Provider Analysis and Review (MEDPAR) file, which contains all Medicare Part A claims for 2004, we calculated condition-specific hospital risk-adjusted mortality rates using the standard convention of the ratio of expected mortality rate to the observed mortality rate. For this calculation, each patient's predicted probability of death was calculated using logistic regression and adjusted for 30 comorbidities defined by Elixhauser et al,32 and by age, race, ZIP-code level median income and education, sex, insurance status, and whether the admission was emergent or elective. Each hospital's expected mortality rate was then calculated by summing the predicted probabilities of death for all of the patients divided by the total number of patients in that hospital. The risk-adjusted mortality rate was calculated by taking the ratio of the observed to expected mortality rates standardized by the average nationwide mortality rate.
Condition-specific inpatient, 30-day, and 1-year risk-adjusted mortality rates were calculated for each hospital. For these calculations, 3 cohorts of patients were defined: all patients who had been admitted to the hospital with the principal diagnosis of AMI (International Classification of Diseases, Ninth Revision, Clinical Modification [ICD-9-CM] codes 410.0-410.9), heart failure (ICD-9-CM codes 402.01, 402.11, 402.91, 404.01, 404.03, 404.11. 404.13, 404.91, 404.93, or 428.0-428.9), or pneumonia (ICD-9-CM codes 480.8, 480.9, 481, or 482.0-487.0). In the pneumonia cohort, patients who had a principal diagnosis of septicemia (ICD-9-CM codes 038.0-038.9) or respiratory failure (ICD-9-CM codes 518.81 or 518.84) and a diagnosis of pneumonia also were included. Finally, patients who did not have an admitting diagnosis of pneumonia were excluded. This was done to exclude patients who developed nosocomial pneumonia during their hospital stay.
Characteristics of Hospitals
The characteristics of each hospital were obtained from Medicare's 2004 Provider of Service file. Characteristics included in the analyses included profit status, number of beds, teaching status, and whether a facility performs open heart surgery (as a measure of hospital technology).33 These measures were chosen because they are often used as implicit measures of hospital quality and are known to affect patient outcomes.33-35
For descriptive purposes, hospitals were grouped into thirds based on their average 1-year risk-adjusted mortality rate for AMI, heart failure, and pneumonia. Hospital characteristics were then aggregated within tertiles. The association between hospital characteristics and tertiles of risk-adjusted mortality rate were assessed.
A Bayesian approach was used to test the relationship between hospital performance on the 10 performance measures and composite measures with condition-specific inpatient, 30-day, and 1-year risk-adjusted mortality rates unadjusted for hospital characteristics. To do this, Bayesian “shrinkage” was applied to each hospital's observed and expected condition-specific mortality rates.36 This process weights the hospital's mortality rates based on the degree of uncertainty used to calculate those rates. Rates from hospitals with small caseloads have greater uncertainty and hence these rates are shrunken more toward the population mean. Similarly, hospitals with large caseloads have a relatively smaller amount of shrinkage and the shrunken estimate is closer to the hospital's observed rate. Bayesian shrinkage helps account for expected regression to the mean in mortality rates.37 Mortality rate estimates that are unadjusted for regression to the mean are biased because the observed mortality rates will usually be farther from the population mean than the true mortality rate. We estimated the relationship between each hospital's risk-adjusted mortality rate and performance, while controlling for other hospital characteristics, using a linear model. The relationship between each performance measure and condition-specific risk-adjusted mortality rates were modeled separately.
After estimating the relationship between hospital performance and mortality rates, we estimated the distance between Medicare beneficiaries and the closest high-performing hospital. Hospitals were classified as high-performing if they performed in the 75th percentile based on their condition-specific composite performance. Distances were calculated based on miles between a hospital's ZIP-code centroid and a Medicare beneficiary's ZIP-code centroid. The median distance between Medicare beneficiaries and the closest high-performing hospital and the proportion of beneficiaries living within 30 miles of a high-performing hospital were calculated. Thirty miles was chosen as a cut point because we were examining acute conditions that may benefit from minimizing the time between symptom onset and initiating treatment.
Finally, a sensitivity analysis was conducted due to concerns about the accuracy and completeness of the Hospital Compare data.23 If poor-performing hospitals either submitted inaccurate data or excluded more patients from the denominator of the performance measures, this could attenuate the relationship between performance and mortality. Therefore, new data were simulated to try to reflect this possibility. To do this, it was first assumed that the 10% of hospitals with the lowest condition-specific composite performance had inflated their performance by 5%. Second, it was assumed that a more extreme case in which the 20% of hospitals with the lowest condition-specific composite performance inflated their performance by 10%. These assumed perturbations in the data were corrected for by systematically deflating the performance of affected hospitals and by rerunning the analyses.
The study protocol and waiver of informed consent was reviewed and approved by the institutional review board of the University of Pennsylvania. All Bayesian analyses were performed using WinBUGS version 1.4 (MRC, Cambridge, England). All other analyses were performed using Stata version 9.0 (StataCorp, College Station, Tex). A P value of less than .05 was considered statistically significant for all analyses.
Of the 4048 hospitals in the Hospital Compare database, 21 did not report performance on any of the 10 performance measures. A total of 284 hospitals were excluded because all of the measures they reported were based on fewer than 25 patients. An additional 86 hospitals included in Hospital Compare were not identified in the 2004 MEDPAR file and were dropped from the analyses. A total of 3657 hospitals were included in the final analyses.
The characteristics of the hospitals are summarized in Table 1. After grouping hospitals into tertiles based on their average 1-year risk-adjusted mortality rates by AMI, heart failure, and pneumonia, the average risk-adjusted mortality rate ranged from 0.27 to 0.40. Among the high-mortality hospitals, a smaller proportion of hospitals were large, were for profit, were teaching hospitals, or had open heart surgery capabilities (P≤.001 for all comparisons). Hospital performance for each of the 10 individual performance measures and the condition-specific composite measures appear in Table 2.
Relationship Between Hospital Performance and Risk-Adjusted Mortality Rates
After adjusting for hospital characteristics, the adjusted difference in risk-adjusted mortality rates for hospitals whose performance measures were in the 25th percentile compared with those in the 75th percentile were estimated, which we found to be a plausible range of hospital performance within individual health care markets (Table 3). For example, among hospitals performing in the 25th percentile for the measure of aspirin at admission for AMI, the risk-adjusted inpatient mortality rate at these hospitals is 0.074. Among hospitals performing in the 75th percentile on this measure, the predicted risk-adjusted inpatient mortality rate at these hospitals is 0.068. This resulted in an absolute reduction in mortality rates of 0.006.
Across all AMI performance measures, the absolute reduction in risk-adjusted mortality rates between hospitals performing in the 25th percentile vs those performing in the 75th percentile was 0.005 for inpatient mortality, 0.006 for 30-day mortality, and 0.012 for 1-year mortality. For the heart failure performance measures, the absolute mortality reduction was smaller, ranging from 0.001 for inpatient mortality to 0.002 for 1-year mortality. For the pneumonia performance measures, the absolute reduction in mortality ranged from 0.001 for 30-day mortality to 0.005 for inpatient mortality.
The adjusted difference in risk-adjusted mortality rates was also estimated using the all-or-none quality measure (Table 4). Between 8% and 14% of hospitals qualified as high-performing using this measure. Differences in mortality rates for hospitals performing above the 75th percentile for all measures within a condition vs those performing lower than the 25th percentile on all reported measures ranged between 0.008 and 0.018 for AMI and between 0.003 and 0.014 for pneumonia. Differences in heart failure–related mortality between high-performing and low-performing hospitals were not statistically significant.
Distance to High-Performing Hospitals
The median distance between Medicare beneficiaries and the closest high-performing hospitals ranged from 26.1 miles (41.8 km) to 29.7 miles (47.5 km) by AMI, heart failure, and pneumonia (Table 5). Distances were longer in the Midwest, South, and West compared with the Northeast, where the median distance to a high-performing hospital was less than 15 miles (24 km). Approximately half of Medicare beneficiaries live within 30 miles (48 km) of a high-performing hospital.
To test whether our results were affected by the possibility that low-performing hospitals systematically alter their measured performance by submitting inaccurate data or excluding more patients from the denominator of the performance measures, we assumed that the lowest-performing hospitals reported performance that was better than it actually was. After simulating data to adjust for this possible perturbation and rerunning the analyses, these changes only minimally changed the magnitude of the effect we found and did not change the overall interpretation of our results (available from author on request).
Hospital performance on the Hospital Compare measures were modestly correlated with condition-specific risk-adjusted mortality rates. Although hospital performance predicted differences in risk-adjusted mortality rates that were statistically significant in some cases, these differences were small. Based on these results, the ability of performance measures to detect clinically meaningful differences in quality across hospitals is questionable.
One way to interpret the magnitude of these results is in terms of the number of lives at risk from patients receiving care at low-performing hospitals instead of high-performing hospitals. For example, approximately 750 000 patients are hospitalized annually for AMI.38 If one third of those patients who received care at the lowest-performing hospitals received care at high-performing hospitals instead, in which mortality rates are 1.2 percentage points lower, approximately 3000 more lives could have been saved. When put in the context of the number of individuals who have access to high-performing hospitals, the number of lives at risk becomes smaller.
While a large majority of Medicare beneficiaries live within 30 miles of a high-performing hospital in the Northeast, only one third of Medicare beneficiaries do in the South. For those with commercial insurance, the number of patients with access to these hospitals may be even further diminished due to selective contracting between hospitals and insurers.39 Hospital crowding40 may further diminish access to high-performing hospitals. Despite our findings of a statistically significant relationship between hospital performance and mortality, the clinical significance of differences in hospital performance may be small due to the small effect size and the constraints limiting patient selection of high-performing hospitals.
There are plausible reasons for the modest effect of hospital performance on risk-adjusted mortality rates. First, Hospital Compare measures discrete aspects of care delivery rather than assessing global quality. While these measures have been tightly linked to patient outcomes in clinical trials41-49 and are included in clinical practice guidelines,50-53 risk-adjusted mortality rates are likely influenced by many factors that are independent of these measures, such as whether a hospital uses electronic health records, staffing levels, activities of quality oversight committees, or other clinical strategies that may impact the clinical outcomes of a hospital. Additionally, the measures used in this study include only a subset of measurable activities at hospitals. It is possible that a larger set of performance measures would represent a more complete picture of performance and would be more tightly linked to mortality. Second, there is relatively little variation in some performance measures across hospitals. For example, in the case of giving aspirin at admission for AMI, hospitals performing in the 25th percentile give aspirin 91% of the time, only 7 percentage points below hospitals performing in the 75th percentile. There is strong evidence to suggest that all eligible patients should receive aspirin for AMI.42,50,51 However, the small amount of variation across hospitals makes it difficult to detect the effect of aspirin. Third, in the case of discharge performance measures, such as prescribing β-blockers at hospital discharge after AMI, the care of patients after hospitalization may be beyond the control of the hospital. Past research has suggested that the persistence of long-term pharmacological therapies is low, with a substantial number of patients stopping treatment within the first 6 months.54
It is also possible that our finding of a modest relationship between hospital performance and mortality rates is an artifact due to possible low reliability of the data used to calculate performance. The CMS audits the data that is reported to it for performance measurement, and others have found that the accuracy of these data is high.23 However, there is no mechanism to monitor whether the data are complete, or whether hospitals include all eligible patients in performance measurement.23
Differences in how hospitals identify patients to include in performance measurement could attenuate the relationship between performance and mortality. For example, poor-performing hospitals may try to game the system by selecting patients who receive the processes that are being measured and excluding those who do not, thereby artificially inflating their measured performance. This strategy would make it more difficult to detect a relationship between performance and mortality if it exists. However, our simulations to control for this type of gaming revealed that even if the worst 20% of hospitals were inflating their performance by 10%, the effect of hospital performance on mortality would remain modest. This suggests that our findings do not result from systematic differences in how hospitals report data for performance measurement.
Our results have important implications. Some may judge the value of performance measures as low because they predict small differences in mortality rates. While the information in Hospital Compare may be minimally informative for consumers wishing to choose high-quality hospitals, consumers may nonetheless judge the revealed differences in risk as important. They may also value the data for other reasons, such as the reassurance it provides regarding the quality of medical care. In addition, public reporting may motivate hospitals to improve the quality of care they deliver. Numerous studies have found that measuring performance and providing feedback to clinicians leads to improvement in performance.55-57 Hospitals that receive feedback on their quality also improve their performance.24 Whether improvements in mortality will follow improvements in performance remains to be seen. The small relationship between performance and mortality documented herein suggests that such improvements in mortality also may be small.
Our results should be interpreted in light of how they inform efforts to tie performance to financial incentives. The CMS is currently experimenting with using measured performance to determine hospital reimbursement.19 Such incentive programs will have to be structured to maximize their impact. This might be done by focusing incentive programs on the lowest performing hospitals rather than trying to improve quality of care among all hospitals because the differences in performance and outcomes are small between hospitals in the 25th percentile and the 75th percentile.
There are potential limitations to our study. First, there are limitations inherent to a cross-sectional study. It is impossible to assess causation of improved performance leading to improved mortality. It is possible that the demonstrated relationship between process and outcomes is confounded by patient or hospital factors. Patient and hospital characteristics were controlled for in our analyses to reduce the likelihood of this sort of bias. Additionally, sensitivity analyses were performed to limit the analyses to academic hospitals and to different regions of the country (results not reported herein). These sensitivity analyses have not changed the documented relationship between process and outcomes. Nonetheless, it is possible that some of the differences in mortality rates between hospitals are attributable to unobserved patient and hospital characteristics. In particular, the persistence of mortality differences 1 year after hospitalization may suggest that the modest effects found in this analysis are due in part to factors other than hospital performance that are unobserved in our data.
Second, our risk-adjustment models were based on administrative data in which comorbidities may be underreported. However, other research has suggested that underreporting in administrative data does not impact the accuracy of hospital quality rankings based on risk-adjusted mortality rates.58 Furthermore, a well-validated risk-adjustment model designed specifically for use with administrative data was used,32,59,60 and this risk model had high predictive power in our testing.
Third, our estimates of hospital risk-adjusted mortality rates were based only on Medicare beneficiaries rather than all payers. However, Medicare beneficiaries make up more than 50% of hospital admissions for the conditions studied herein and health care dollars spent on Medicare beneficiaries significantly exceeds that of other payers. In addition, hospital quality research based on Medicare data may be generalizable to the broader US population.61
Fourth, mortality as an outcome was assessed. It is possible that hospital performance is an important predictor of other outcomes, such as length of stay, failure to rescue, or readmission rates.
The CMS has initiated performance measurement not only in hospitals but also in nursing homes,12 and home health agencies,13 and efforts are under way to extend measurement to physicians.62 The use of performance measurement will be expanded in 2007 with the Deficit Reduction Act and efforts are under way to link performance to payment.19,62 These efforts have been made under the assumption that improvements in processes of care will lead to improvements in patient outcomes.63
Our study suggests that in the case of hospital performance, the CMS's current set of performance measures are not tightly linked to patient outcomes. These findings should not undermine current efforts to improve health care quality through performance measurement and reporting. However, attention should be focused on finding measures of health care quality that are more tightly linked to patient outcomes. Only then will performance measurement live up to expectations for improving health care quality.
Corresponding Author: Rachel M. Werner, MD, PhD, 1230 Blockley Hall, 423 Guardian Dr, Philadelphia, PA 19104 (rwerner@mail.med.upenn.edu).
Author Contributions: Dr Werner had full access to all of the data in the study and takes responsibility for the integrity of the data and the accuracy of the data analysis.
Study concept and design: Werner, Bradlow.
Acquisition of data: Werner.
Analysis and interpretation of data: Werner, Bradlow.
Drafting of the manuscript: Werner.
Critical revision of the manuscript for important intellectual content: Werner, Bradlow.
Statistical analysis: Werner, Bradlow.
Obtained funding: Werner.
Administrative, technical, or material support: Werner.
Study supervision: Werner, Bradlow.
Financial Disclosures: None reported.
Funding/Support: Dr Werner was supported by a career development award from the Department of Veterans Affairs.
Role of the Sponsor: The Department of Veterans Affairs had no role in the design and conduct of the study; collection, management, analysis, and interpretation of the data; and preparation, review, or approval of the manuscript.
Acknowledgment: We thank David Asch, MD, MBA (Center for Health Equity Research and Promotion, Philadelphia Veterans Affairs Medical Center Division of General Internal Medicine, University of Pennsylvania School of Medicine and Leonard Davis Institute of Health Economics), for comments on drafts of this article. Dr Asch was not compensated for his review.
1.McGlynn EA, Asch SM, Adams J.
et al. The quality of health care delivered to adults in the United States.
N Engl J Med. 2003;348:2635-264512826639
Google ScholarCrossref 2.Institute of Medicine. Crossing the Quality Chasm: A New Health System for the 21st Century. Washington, DC: National Academy Press; 2001
3.Asch SM, Kerr EA, Keesey J.
et al. Who is at greatest risk for receiving poor-quality health care?
N Engl J Med. 2006;354:1147-115616540615
Google ScholarCrossref 4.Jha AK, Li Z, Orav EJ, Epstein AM. Care in US hospitals—the hospital quality alliance program.
N Engl J Med. 2005;353:265-27416034012
Google ScholarCrossref 5.Leapfrog Group. The Leapfrog Group Hospital Patient Safety Survey, April 2003-March 2004. Washington, DC: Leapfrog Group; 2004
6.Pennsylvania Health Care Cost Containment Council. Hospital Performance Report. Harrisburg: Pennsylvania Health Care Cost Containment Council; 2005
17.Werner RM, Asch DA. The unintended consequences of publicly reporting quality information.
JAMA. 2005;293:1239-124415755946
Google ScholarCrossref 18.Marshall MN, Shekelle PG, Leatherman S, Brook RH. The public release of performance data: what do we expect to gain? a review of the evidence.
JAMA. 2000;283:1866-187410770149
Google ScholarCrossref 20.Epstein AM, Lee TH, Hamel MB. Paying physicians for high-quality care.
N Engl J Med. 2004;350:406-41014736934
Google ScholarCrossref 21.Doran T, Fullwood C, Gravelle H.
et al. Pay-for-performance programs in family practices in the United Kingdom.
N Engl J Med. 2006;355:375-38416870916
Google ScholarCrossref 23.US Governement Accountability Office. Hospital Quality Data: CMS Needs More Rigorous Methods to Ensure Reliability of Publicly Released Data. Washington, DC: US Government Accountability Office; 2006. Publication GAO-06-54
24.Williams SC, Schmaltz SP, Morton DJ, Koss RG, Loeb JM. Quality of care in US hospitals as reflected by standardized measures, 2002-2004.
N Engl J Med. 2005;353:255-26416034011
Google ScholarCrossref 25.Peterson ED, Roe MT, Mulgund J.
et al. Association between hospital process performance and outcomes among patients with acute coronary syndromes.
JAMA. 2006;295:1912-192016639050
Google ScholarCrossref 26.Higashi T, Shekelle PG, Adams JL.
et al. Quality of care is associated with survival in vulnerable older patients.
Ann Intern Med. 2005;143:274-28116103471
Google ScholarCrossref 29.McCormick D, Himmelstein DU, Woolhandler S, Wolfe SM, Bor DH. Relationship between low quality-of-care scores and HMOs' subsequent public disclosure of quality-of-care scores.
JAMA. 2002;288:1484-149012243635
Google ScholarCrossref 30.Joint Commision on Accreditation of Healthcare Organizations Web site. Quality report user guide.
http://www.jointcommission.org/. Accessibility verified November 1, 2006 31.Nolan T, Berwick DM. All-or-none measurement raises the bar on performance.
JAMA. 2006;295:1168-117016522838
Google ScholarCrossref 32.Elixhauser A, Steiner C, Harris DR, Coffey RM. Comorbidity measures for use with administrative data.
Med Care. 1998;36:8-279431328
Google ScholarCrossref 33.Aiken LH, Clarke SP, Sloane DM, Sochalski J, Silber JH. Hospital nurse staffing and patient mortality, nurse burnout, and job dissatisfaction.
JAMA. 2002;288:1987-199312387650
Google ScholarCrossref 34.Ayanian JZ, Weissman JS, Chasan-Taber S, Epstein AM. Quality of care for two common illnesses in teaching and nonteaching hospitals.
Health Aff (Millwood). 1998;17:194-2059916369
Google ScholarCrossref 35.Hartz AJ, Krakauer H, Kuhn EM.
et al. Hospital characteristics and mortality rates.
N Engl J Med. 1989;321:1720-17252594031
Google ScholarCrossref 36.Christiansen CL, Morris CN. Improving the statistical approach to health care provider profiling.
Ann Intern Med. 1997;127:764-7689382395
Google ScholarCrossref 38.Agency for Health Care Research and Quality. HCUP-3 Nationwide Inpatient Sample 2003. Rockville, Md: Agency for Health Care Research and Quality; 2003
39.Glied S. Managed care. In: Culyer AJ, Newhouse JP, eds. Handbook of Health Economics, Vol 1A. Amsterdam, the Netherlands: Elsevier Science; 2000:707-753
40.Kolata G, Abelson R. A bonus for health, payable to the doctor. New York Times. April 15, 2005:C1
41. Timolol-induced reduction in mortality and reinfarction in patients surviving acute myocardial infarction.
N Engl J Med. 1981;304:801-8077010157
Google ScholarCrossref 42.ISIS-2 (Second International Study of Infarct Survival) Collaborative Group. Randomised trial of intravenous streptokinase, oral aspirin, both, or neither among 17,187 cases of suspected acute myocardial infarction: ISIS-2.
Lancet. 1988;2:349-3602899772
Google Scholar 43.Antiplatelet Trialists' Collaboration. Collaborative overview of randomised trials of antiplatelet therapy: prevention of death, myocardial infarction, and stroke by prolonged antiplatelet therapy in various categories of patients.
BMJ. 1994;308:81-1068298418
Google ScholarCrossref 44.Pfeffer MA, Braunwald E, Moye LA.
et al. SAVE Investigators. Effect of captopril on mortality and morbidity in patients with left ventricular dysfunction after myocardial infarction: results of the survival and ventricular enlargement trial.
N Engl J Med. 1992;327:669-6771386652
Google ScholarCrossref 45.First International Study of Infarct Survival Collaborative Group. Randomised trial of intravenous atenolol among 16,027 cases of suspected acute myocardial infarction: ISIS-1.
Lancet. 1986;2:57-662873379
Google Scholar 46.SOLVD Investigators. Effect of enalapril on survival in patients with reduced left ventricular ejection fractions and congestive heart failure.
N Engl J Med. 1991;325:293-3022057034
Google ScholarCrossref 47.CONSENSUS Trial Study Group. Effects of enalapril on mortality in severe congestive heart failure: results of the Cooperative North Scandinavian Enalapril Survival Study (CONSENSUS).
N Engl J Med. 1987;316:1429-14352883575
Google ScholarCrossref 48.Shapiro ED, Berg AT, Austrian R.
et al. The protective efficacy of polyvalent pneumococcal polysaccharide vaccine.
N Engl J Med. 1991;325:1453-14601944423
Google ScholarCrossref 49.Houck PM, Bratzler DW, Nsa W, Ma A, Bartlett JG. Timing of antibiotic administration and outcomes for Medicare patients hospitalized with community-acquired pneumonia.
Arch Intern Med. 2004;164:637-64415037492
Google ScholarCrossref 50.Antman EM, Anbe DT, Armstrong PW.
et al. ACC/AHA Guidelines for the Management of Patients With ST-Elevation Myocardial Infarction. Washington, DC and Dallas, Tex: American College of Cardiology Foundation and American Heart Association; 2004
51.Braunwald E, Antman EM, Beasley JW.
et al. ACC/AHA guidelines for the management of patients with unstable angina and non-ST-segment elevation myocardial infarction: a report of the American College of Cardiology/American Heart Association task force on practice guidelines (committee on the management of patients with unstable angina).
J Am Coll Cardiol. 2000;36:970-106210987629
Google ScholarCrossref 52.Hunt SA, Abraham WT, Chin MH.
et al. Writing Committee to Update the 2001 Guidelines for the Evaluation and Management of Heart Failure. ACC/AHA 2005 Guideline Update for the Diagnosis and Management of Chronic Heart Failure in the Adult: A Report of the American College of Cardiology/American Heart Association Task Force on Practice Guidelines. Washington, DC and Dallas, Tex: American College of Cardiology Foundation and American Heart Association; 2005
53.Niederman MS, Mandell LA, Anzueto A.
et al. Guidelines for the management of adults with community-acquired pneumonia: diagnosis, assessment of severity, antimicrobial therapy, and prevention.
Am J Respir Crit Care Med. 2001;163:1730-175411401897
Google ScholarCrossref 54.Benner JS, Glynn RJ, Mogun H, Neumann PJ, Weinstein MC, Avorn J. Long-term persistence in use of statin therapy in elderly patients.
JAMA. 2002;288:455-46112132975
Google ScholarCrossref 55.Jha AK, Perlin JB, Kizer KW, Dudley RA. Effect of the transformation of the Veterans Affairs health care system on the quality of care.
N Engl J Med. 2003;348:2218-222712773650
Google ScholarCrossref 56.Jencks SF, Huff ED, Cuerdon T. Change in the quality of care delivered to Medicare beneficiaries, 1998-1999 to 2000-2001.
JAMA. 2003;289:305-31212525231
Google ScholarCrossref 57.Ferguson TB, Peterson ED, Coombs LP.
et al. Use of continuous quality improvement to increase use of process measures in patients undergoing coronary artery bypass graft surgery: a randomized controlled trial.
JAMA. 2003;290:49-5612837711
Google ScholarCrossref 58.Austin PC. The impact of under coding of cardiac severity an comorbid diseases on the accuracy of hospital report cards.
Med Care. 2005;43:801-80916034294
Google ScholarCrossref 59.Southern DA, Quan H, Ghali WA. Comparison of the Elixhauser and Charlson/Deyo methods of comorbidity measurement in administrative data.
Med Care. 2004;42:355-36015076812
Google ScholarCrossref 60.Stukenborg GJ, Wagner DP, Connors AF Jr. Comparison of the performance of two comorbidity measures, with and without information from prior hospitalizations.
Med Care. 2001;39:727-73911458137
Google ScholarCrossref 61.Needleman J, Buerhaus PI, Mattke S, Stewart M, Zelevinsky K. Measuring hospital quality: can Medicare data substitute for all-payer data?
Health Serv Res. 2003;38:1487-150814727784
Google ScholarCrossref 63.Palmer RH. Process-based measures of quality: the need for detailed clinical data in large health care databases.
Ann Intern Med. 1997;127:733-7389382389
Google ScholarCrossref