The blog of Ashish Jha — physician, health policy researcher, and advocate for the notion that an ounce of data is worth a thousand pounds of opinion.

Tag Archives: Patient experience

The recent articles in the New York Times about the Hospital Corporation of America (HCA) have once again raised important questions about the role of for-profit hospitals in the U.S. healthcare system. For-profits make up about 20% of all hospitals and many of them are part of large chains (such as HCA). Critics of for-profit hospitals have argued that these institutions sacrifice good patient care in their search for better financial returns. Supporters argue that there is little evidence that their behavior differs substantially from non-profit institutions or that their care is meaningfully worse.

To me, this is essentially an empirical question. Yet, as I read through the articles, I was struck by the dearth of data provided on the quality of care at these hospitals. Based on the comments that followed the stories, it was clear that many readers came away thinking that these hospitals had sacrificed quality in order to maximize profits. Here, I thought an ounce of evidence might be helpful.

Measuring quality:

There is no perfect way to measure the quality of a hospital. However, the science of quality measurement has made huge progress over the past decade. There is increasing consensus around a set of metrics, many of which are now publicly reported by the government and even are part of pay-for-performance schemes. While one can criticize every one of these metrics as imperfect, taken together, they paint a relatively good, broad picture of the quality of care in an institution. We focused on five metrics with widespread acceptance:

Mortality rates (Proportion of people who die within 30 days of hospitalization, taking into account the “sickness” of the patient)

Readmission rates (Proportion of people who are readmitted within 30 days of discharge, taking into account the “sickness” of the patient)

Hospital Safety Score (a measure of how effective a hospital likely is at preventing medically-induced harm to patients).

An important caveat: The NY Times article highlighted terrible, unethical practices by some physicians at HCA hospitals who appear to put in cardiac stents when there was no clinical indication. We don’t have the data to examine whether this practice occurs more often at HCA hospitals than at other institutions. Therefore, I’ve decided to focus more broadly at hospital quality. Most of the metrics above have been approved by the National Quality Forum, are widely regarded by “experts” as being good, and are used by Medicare to judge and pay for quality.

How we analyzed the data:

We examined all U.S. hospitals in four groups: Privately-owned non-profit hospitals, government-owned public hospitals, for-profit hospitals that were not part of the HCA chain, and HCA hospitals. In our analysis we “adjusted” for characteristics that are beyond the hospital’s control such as size, teaching status, urban versus rural location, and region of the country (adjusting is important: imagine that all the for-profit hospitals were large and large hospitals generally had better quality. Without adjustment, we’d say for-profit hospitals were better and therefore, we should encourage more for-profit hospitals. With adjustment, we’d be able to hold size differences constant and examine the actual relationship between quality and the profit status of the hospital).

What we found:

In the table below, we use “non-profit” hospitals as the reference group because it’s the largest group of hospitals. All the scores that are statistically different (at p-value <0.05) are highlighted either in red if they are significantly worse or in green if they are significantly better.

Interpretation

The best part of looking at data is that you get to draw your own conclusions. Here are mine. Public hospitals are struggling on nearly every metric. For-profit hospitals outside of the HCA are a mixed bag – they do worse on patient experience (as we’ve found before), better on processes measures, and somewhat worse on mortality and readmission rates. They are about average on the Leapfrog safety score.

However, HCA hospitals look pretty good. They tend to have good patient experience scores, really excellent process quality (adherence to guidelines) and are average or above average on mortality and readmissions (pneumonia mortality does appear to be high, though not statistically significant). They do very well on the Leapfrog safety score* (nearly half got an “A”).

My takeaway is that although which hospital you go to has a profound impact on whether you live or die, whether the hospital is “for-profit” or “not for profit” has very little to do with it. What really matters is leadership, focus on quality, and a dedication to improvement. That appears to happen equally well (or badly, depending on your perspective) in both for-profit and non-profit hospitals.

So, when it comes to quality, it’s time to stop thinking about it as an issue of “for-profit versus non-profit” hospitals. Instead, it’s time to start talking about the large number of relatively poor performing hospitals where patients are being hurt or killed un-necessarily. Those hospitals come in all sizes, shapes, and yes, ownership structures, and we have to figure out how to make them better.

Finally, these analyses were run by Sidney T. Le, a terrific young analyst in our group. You should follow him on twitter (@sidtle) although his love of Stanford sports can be a challenge. Consider yourself warned.

*The Leapfrog safety score was developed by a group of experts (full disclosure: I was on that panel – but don’t worry, there were many people much smarter than me on the panel).

After years of breaking down, my sedan recently died. Finding myself in the market for a new car, I did what most Americans would do – went to the web. Reading reviews and checking rankings, it quickly became clear that each website emphasized something different: Some valued fuel-efficiency and reliability, while others made safety the primary concern. Others clearly put a premium on style and performance. It was enough to make my head spin, until I stopped to consider: What really mattered to me? I decided that safety and reliability were my primary concerns and how fun a car was to drive was an important, if somewhat distant, third consideration.

For years, many of us have complained about the lack of similarly accessible, reliable information about healthcare. These issues are particularly salient when we consider hospital care. Despite a long-standing belief that all hospitals are the same, the truth is startlingly different: where you go has a profound impact on whether you live or die, whether you are harmed or not. There is an urgent need for better information, especially as consumers spend more money out of pocket on healthcare. Until recently, this type of transparent, consumer-focused information simply didn’t exist.

Over the past couple of months, things have begun to change. Three major organizations recently released groundbreaking hospital rankings. The Leapfrog Group, a well-respected organization focused on promoting safer hospital care, assigned hospitals grades (“A” through “F”) based on how well it cared for patients without harming them*. Consumer Reports, famous for testing cars and appliances, came out with its first “safety score” for hospitals. U.S. News & World Report, probably the most widely known source for rating colleges, released its annual hospital rankings, but this time with an important change in methodology. While others have also gotten into this game, these three organizations bring the highest amount of care, transparency, and credibility.

What is particularly striking, though, is how differently the hospitals come out in the recommendations. Just like the different car rating websites that emphasize different things, these three have also taken approaches very different from each other. If we want to understand why some hospitals get an “A” on Leapfrog or are rated one of America’s Best Hospitals but get an abysmal score on Consumer Reports, we have to look “under the hood.”

It is worth noting that there are important similarities across the rating systems. They each include hospital infection rates, though to varying degrees. All three incorporate hospital mortality rates, though they weight them differently. In accompanying blogs, colleagues from the Leapfrog Group and Consumer Reports each explain their methods and approach in far greater detail than I can, so I highlight what each emphasizes, and focus on their differences.

Let’s start with Leapfrog, whose primary focus is ensuring that when you walk into the hospital, your chances of being hurt or killed by the medical care you receive (as opposed to dying from the underlying condition that brought you in) are minimized. Each year, as many as 100,000 people die in U.S. hospitals due to medical errors, a stunning number that has changed very little over the past decade. The most common causes of death and injury in hospitals are medication errors and healthcare-associated infections, though there is no shortage of ways that you can be injured. Leapfrog creates a composite safety score, emphasizing whether a hospital has computerized prescribing, how well it staffs its intensive care units, and what its rates of infections are. Your chances of being injured in the hospital are generally lower at an “A” hospital than a “C” hospital.

Consumer Reports (CR) also calculates a “safety score,” but its focus is very different. First, CR also incorporates infection rates. It also emphasizes three areas not incorporated by Leapfrog. The first is “double” CT scans (which, thankfully, are rare) and represent un-necesary radiation. The second is patients’ reports on how well their providers communicated about medications and how well providers explained what would happen after discharge. I’m a huge fan of patient-reported measures (it’s the one chance we consumers get to rate hospitals). The challenge is that it captures both patient expectationsas well as hospital performance. The latest evidence suggests that patients’ expectations, including their tendency to give high scores, vary a lot based on who they are and where they live. A final area of emphasis is readmissions, a very complicated metric. The best evidence to date suggests that only a small proportion of readmissions are preventable by hospitals. In fact, good social support at home and the ability to get in and see your primary care physician probably matters much more than what the hospital does. The other factors in the CR score that are emphasized less (including mortality) can be found here.

U.S. News has been rating hospitals for a while, though they made a very important methodological change recently. Three things make up their score: mortality (1/3), reputation (1/3) and other stuff (1/3). The other stuff is a collection of safety and technology indicators. The reputation score comes from a survey of specialists in each field and is meant to capture the intangible knowledge about hospital care that the rest of us don’t have. Finally, and most importantly, they now emphasize mortality rates, which in many ways are “the bottom line” for a hospital’s performance. Their risk-adjusted mortality rate captures both how well the hospital provides the right treatments (a.k.a. “effectiveness”) and how well it avoids harming patients (a.k.a. “safety”).

Implications: Who wins and who loses?

Different approaches lead to different winners and losers (see table below). In the Leapfrog scores, there aren’t large variations by hospital type (i.e. size, teaching status). There aren’t large regional variations either – in every part of the country, there are lots of winners and losers. It is worth noting that safety-net hospitals (as we have previously defined) generally do worse on Leapfrog (19% of those who got an “A” were safety-net hospitals, compared to 31% who got a “C”).

Because CR emphasizes infections but also patient experiences and readmissions, big “winners” on their score are small, community-based hospitals with very few minority or poor patients. Hospitals in the Midwest do particularly well. Major teaching hospitals do extremely poorly (three times more likely to end up in the worst quartile). These worst quartile hospitals also have a lot more minority and poor patients – most likely because hospitals that care for poor patients have higher readmission rates (remember, poor social supports at home likely drive this) and worse patient experience scores.

U.S. News top-ranked hospitals are usually big, academic teaching hospitals. On this metric, at least, it’s nearly impossible for a small community hospital to be rated as one of the best in the country.

What’s a consumer to do?

If you’re lucky enough to find a hospital that gets rated highly by all three organizations, I’d take that in a heartbeat. It’s like finding a car that drives well, looks stylish, is reliable, and safe! No brainer. Unfortunately, those hospitals are rare. In Boston, Massachusetts General Hospital was ranked the #1 hospital in the country by U.S. News. It got an “A” from Leapfrog. It was near the bottom of Massachusetts hospitals in the CR rating, receiving a score of just 45 out of 100.

It can be easy to decide if the safety, or the style, or the performance of a car is most important to you. Unfortunately, choosing what’s most important in health care can make us ask difficult and seemingly unreasonable questions. Is my primary goal to survive my hospitalization, avoid infections and medication errors, or have a reasonably good experience? Every individual has to decide what matters most. If a low mortality rate is most important, U.S. News is your best bet. If you care most about patient safety, then Leapfrog is the way to go. Consumer Reports emphasizes infections, unnecessary radiation and patient experience. If those matter most, CR is your best bet.

My personal list ranks mortality as most important (by far), followed by safety, with patient experience an important but distant third. Others will make different choices. This is why we need different lists and different types of hospitals. It’s time for consumers to use this new information to make better choices. I know that there is little precedent for consumers choosing healthcare providers based on quality. I’ve always believed it is because they lacked good data. In an era of greater transparency, if consumers vote with their feet, it’ll make hospital care better for everyone.

*Please note that I served on a blue-ribbon panel that advised the Leapfrog Group on their methodology. I received no financial compensation for this effort. The final choice of metrics and the scoring methodology for the Leapfrog Safety Score was made by the Leapfrog Group and not by the blue-ribbon panel.