The evening before she was to be discharged, I received a call at home from a colleague asking me if I knew that my patient was scheduled to go to the operating room the following morning. Stunned, I immediately phoned the surgeon, who explained that he was being pushed to operate by Dr. Adelman, who was “pissed we got heart failure involved … It’s a very political situation,” he said apologetically. “Adelman is a big referrer.”
After getting off the phone, I called Rajiv to ask him what to do. “Tell him it’s unacceptable,” my brother cried. “She’s your patient!”
So I phoned the surgeon back. “I am the only cardiologist leaving notes,” I told him politely. “I’ve been recommending no surgery for days.”
The surgeon said he was sorry and explained that Dr. Adelman had called him directly. “My whole career I have always acted with the feeling that it’s not worth it to go to town over one case,” he said. “Adelman sends me a lot of business. I don’t want to lose it—or your business either.” He wondered if we should get a third opinion. He mentioned asking a private cardiologist he knew to rubber-stamp the decision to send Ms. Harris to surgery.
While the surgeon and I were conferring, Rajiv called me back. Inexplicably, he had changed his mind. “Just let it go,” he now said. “Tell him you’re a junior guy and you’ll defer to whatever he says.”
“Are you joking?” I said, perplexed by his change of heart.
“Look, you can’t always insist on having your way,” Rajiv snapped. “Sometimes you have to let people make their own mistakes.”
A group of us had an urgent meeting in Rajiv’s office the following morning before the operation. “Just let them do it,” a colleague told me. “Medically, it may or may not be the right thing, but politically—well, if she goes back to Adelman with shortness of breath, he’s going to say, ‘What the fuck, we sent her over there for an operation, and those guys didn’t do anything.’”
Rajiv agreed. “You did what you thought was right,” he said. He reminded me that there were differences of opinion over whether mitral valve surgery was warranted in elderly patients like Mildred Harris. How could I be so sure that my judgment was correct? I told him that at the very least there should be another opinion in the chart. How could I agree to send my patient for surgery now when I had dismissed the idea in all my previous notes?
“No one is going to blame you,” Rajiv said coolly. “If they take the patient to surgery without your permission, the burden is on them.” He smiled slyly. “If anything, you may be subpoenaed to testify, but that’s all.”
When I went upstairs to talk to my patient, transporters were already there with a stretcher to take her to the OR. I asked Ms. Harris how she was feeling. “I couldn’t sleep all night,” she said, as the orderlies transferred her to the narrow transport gurney. “I thought you told me I could go home.” I told her the decision had been changed.
Fortunately, the surgery went well, and she was discharged from the hospital after a few days. I phoned Santo Russo at Columbia to tell him about what had transpired. “This sort of thing happens all the time,” he told me. “Just last week I said no to mitral valve surgery on a ninety-three-year-old. When the surgeon asked me about it, I said it was just my opinion. If they don’t listen, I sign off the case. I do it in a nice way, a politically correct way. I try not to ruffle too many feathers.”
“But she was high risk,” I said, hoping for some reassurance. “Her symptoms were controlled with medications. Plus, she didn’t want the surgery.” I remembered James Irey, the patient Santo and I had forced into repeating a catheterization—with a fatal result.
“I agree with you,” Santo replied. “I would have done exactly the same thing.” He acknowledged the frustrations of working in a hospital where referring physicians were pushing your hand. “If you mess up relations with a referrer, you can get fired,” he warned.
I told him that I had good relations with most physicians, except for maybe one or two.
“Well, if everyone likes you, then you’re probably not doing a good job,” he said.
* * *
In my first year as an attending I also sat in on meetings on how to improve the hospital’s compliance with certain quality indicators. These “core measures” sprung from a general quality-improvement program called pay for performance (P4P). Employers and insurers, including Medicare, had started about a hundred such initiatives across the country. The general intent was to reward doctors for providing better care. For example, doctors received bonuses if they prescribed ACE inhibitor drugs, which reduce blood pressure, to patients with congestive heart failure. (Only about two-thirds of heart failure patients nationwide were receiving life-prolonging ACE inhibitors, a deadly oversight by busy physicians.) Hospitals got bonuses if they administered antibiotics to pneumonia patients in a timely manner. On the surface, this seemed like a good idea: reward doctors and hospitals for quality, not just quantity. It seemed like the perfect solution to the fee-for-service problem. But pay for performance, I quickly learned, could have untoward consequences.
A colleague once asked me for help in treating a patient with congestive heart failure who had just been transferred from another hospital. The patient was a charming black man in his early sixties who applied mild exasperation to virtually every remark. (“I didn’t like the food, Doc. But I ate it anyways!”) When I looked over his medical chart, I noticed that he was receiving an intravenous antibiotic every day. No one seemed to know why. Apparently it had been started in the emergency room at the other hospital because doctors there thought he might have pneumonia.
But he did not appear to have pneumonia or any other infection. He had no fever. His white blood cell count was normal, and he wasn’t coughing up sputum. His chest X-ray did show a vague marking that could be interpreted as a pneumonic infiltrate, but given the clinical picture, that was probably just fluid in the lungs from heart failure.
I ordered the antibiotic stopped—but not in time to prevent the patient from developing a severe diarrheal infection called C. difficile colitis, often caused by antibiotics. He became dehydrated. His temperature spiked to alarming levels. His white blood cell count almost tripled. In the end, with different antibiotics, the infection was brought under control, but not before the patient had spent almost two weeks in the hospital.
The complication stemmed from the requirement from Medicare that antibiotics be administered to a pneumonia patient within six hours of arriving at the hospital. The trouble is that doctors often cannot diagnose pneumonia that quickly. You have to talk to and examine a patient, wait for blood tests and chest X-rays, and then often observe the patient over time to determine the true mechanism of disease. Under P4P, there is pressure to treat even when the diagnosis isn’t firm, as was the case with this gentleman. So more and more antibiotics are being used in emergency rooms today, despite all-too-evident dangers like antibiotic-resistant bacteria and antibiotic-associated infections.
I spoke with a senior health care quality consultant about this problem at a hospital quality meeting. “We’re in a difficult situation,” he said. “We’re introducing these things without thinking, without looking at the consequences. Doctors who wrote care guidelines never expected them to become performance measures.” In other words, he explained, recommended care in certain situations had become mandated care in all situations.
The guidelines could have a chilling effect, he said. “What about hospitals that stray from the guidelines in an effort to do even better? Should they be punished for trying to innovate? Will they have to take a hit financially until performance measures catch up with current research?” Moreover, how do you correct for patient mix? Hospitals with larger catchment areas will have longer delays in treating heart attacks because of increased patient transit times. How do you adjust for these demographic factors?
To better understand the potential problems, one simply has to look at another quality-improvement program: surgical report cards. In t
he early 1990s, New York and Pennsylvania started publishing mortality statistics on hospitals and surgeons performing coronary bypass surgery. The purpose of these report cards was to improve the quality of cardiac surgery by pointing out deficiencies in hospitals and surgeons. The idea was that surgeons who did not measure up would be forced to improve.
Of course, surgical deaths are affected by myriad factors, so models were created to predict surgical risks and avoid penalizing surgeons who took on the most difficult cases. For example, a fifty-year-old man, otherwise healthy, who underwent coronary bypass surgery was judged to have a risk of death of about 1 percent. For a seventy-year-old on intravenous nitroglycerin with a history of heart surgery, congestive heart failure, emphysema, and other medical problems, the risk was about 20 percent or higher. Many surgeons, however, thought that the models underestimated surgical risk, particularly for the sickest patients. They criticized the models for oversimplifying heart surgery. Surgery, they argued, is a team sport, involving referring physicians, technicians, nurses, anesthesiologists, and surgeons. Many variables can affect patient outcomes that are beyond a surgeon’s control. Among other things, they said, the models did not account for simple bad luck.
Surgeons began to fastidiously report—some would say overreport—medical conditions that could affect the outcome of surgery. In some New York hospitals the prevalence in surgical patients of emphysema, a condition known to increase surgical risk, increased from a few percent to more than 50 percent after report cards came into use. Some surgeons even made a habit of routinely putting patients on intravenous nitroglycerin because being on the drug conferred added risk, thus covering for a possibly poor outcome. Others tried to “hide” surgical deaths by transferring patients to hospice programs right before they died.
“It’s all about the numbers,” a surgeon at NYU, where I first learned about report cards during my fellowship, told me. “We have to start coding for everything, and you guys have to help us out.” There would be no high-risk surgeries, he added, unless the risk was documented in detail. “If I don’t operate again until next year, that’s okay with me.”
Despite these excesses, in the beginning there were high hopes for this quality-improvement program. In the first few years there were major gains in surgical outcomes. The most striking results were in New York State, where mortality rates for coronary bypass surgery declined a whopping 41 percent, and outcomes improved for all hospitals at all levels. In a 1994 article in Annals of Internal Medicine, the cardiologists Eric Topol and Robert Califf wrote that “appropriate implementation of score cards could ultimately lead to a substantial improvement in the quality of U.S. cardiovascular medicine.”
But not everyone believed that report cards were causing real improvements in care. Some entertained a more disturbing possibility. Were surgeons’ numbers improving because of better performance or because sicker patients were not getting the operations they needed?
In 2003, researchers at Northwestern and Stanford tried to answer this question. Using Medicare data, they studied all elderly patients in the United States who had had heart attacks or coronary bypass surgery in 1987 (before report cards were used) and 1994 (after they had taken effect). They compared New York and Pennsylvania, states with mandatory surgical report cards, with the rest of the country. They discovered a significant amount of “cherry-picking” in the states with mandatory report cards, and learned that patients generally were worse-off for it. They wrote: “Mandatory reporting mechanisms inevitably give providers the incentive to decline to treat more difficult and complicated patients,” adding that “observed mortality declined as a result of a shift in incidence of surgeries toward healthier patients, not because report cards improved the outcomes of care for individuals with heart disease.” Doctors agreed with these conclusions. In a survey in New York State, 63 percent of cardiac surgeons acknowledged that because of report cards, they were accepting only relatively healthy patients for heart bypass surgery. And 59 percent of cardiologists said it had become harder to find a surgeon to operate on their most severely ill patients.
(Of course, it isn’t only heart surgeons who are feeling this pressure. Similar pressures are being brought to bear in other areas of medicine. For example, there is evidence that report cards on interventional cardiologists have resulted in a drop in the number of angioplasty procedures performed on very sick patients. “I said, ‘Just treat her with medicine,’” an interventional cardiologist at NYU told me about a critically ill patient in shock on whom he had refused to do angioplasty. “I didn’t tell them it was because I didn’t want a death on the table.”)
Whenever you try to dictate professional behavior, there are bound to be unintended consequences. With surgical report cards, surgeons’ numbers improved not only because of better performance but also because dying patients were not getting the operations they needed. Pay for performance is likely to have similar repercussions.
For example, doctors today are being encouraged to voluntarily report to Medicare on sixteen quality indicators, including prescribing aspirin and beta-blocker drugs to patients who have suffered heart attacks and strict cholesterol and blood pressure control for diabetics. Those who perform well receive cash bonuses.
But what to do about complex patients with multiple medical problems? Half of Medicare beneficiaries over sixty-five have at least three chronic conditions. Twenty-one percent have five or more. P4P quality measures are focused on acute illness. It isn’t at all clear they should be applied to elderly patients with multiple disorders who may have trouble keeping track of their medications. With P4P doling out bonuses, many doctors worry that they will feel pressured to prescribe “mandated” drugs, even to elderly patients who may not benefit, and to cherry-pick patients who can comply with the measures.
Moreover, which doctor should be held responsible for meeting the quality guidelines? Medicare patients see on average two primary care physicians in any given year and five specialists working in four different practices. Care is widely dispersed, so it is difficult to assign responsibility to one doctor. If a doctor assumes responsibility for only a minority of her patients, then there is little financial incentive to participate in P4P. If she assumes too much responsibility, she may be unfairly blamed for any lapses in quality.
Nor is it even clear that pay for performance actually results in better care, because it may end up benefiting mainly those physicians who already meet the guidelines. A few years ago, researchers at Harvard conducted a study on the impact of P4P at one of the nation’s largest health plans, PacifiCare Health Systems. In 2003, PacifiCare began paying bonuses to medical groups in the Pacific Northwest if they met or exceeded ten quality targets. The researchers compared the performance of these groups with a control group on three measures of clinical quality: cervical cancer screening, mammography, and diabetes testing. For all three measures, physician groups with better performance at baseline improved the least but got the most payments (per enrollee, the maximum annual bonus was about $27). If they could collect bonuses by maintaining the status quo, what was the incentive for these doctors to improve? Another study also showed no difference in thirty-day mortality for patients hospitalized with one of four conditions—heart failure, myocardial infarction, coronary bypass surgery, or pneumonia—at 252 hospitals that participated in P4P as compared with more than 3,000 control hospitals that did not.
Several simple reforms could improve P4P. Insurers could use less stringent requirements for antibiotic delivery. For example, to minimize antibiotic misuse they could set the clock running after a diagnosis of pneumonia is made, instead of when a patient is first brought into the ER. They could also use percentile-based rankings or rolling averages over extended time periods so hospitals don’t feel pressured to be at 100 percent compliance all the time. The irony is that lowering the benchmark—allowing for a bit of wiggle room—is more likely to result in proper care.
P4P not surprisingly is deeply unpopular
among most American physicians. It forces them to follow certain clinical priorities—“cookbook medicine,” “a rule book”—leaving them deeply dissatisfied with the loss of autonomy. Many doctors say they feel like pawns in a game being played by regulators. Instead of being allowed to exercise their professional judgments and deliver “patient-centered” care, physicians believe they are being guided on what to do with a burgeoning menu of incentive payments or strict regulations. One recently wrote online:
We as a profession are partly to blame. We allowed the insurance companies to become the intermediary between us and our patients. There was more money at first but now we suffer the “controls” they have put in place. With regulations reaching the point of insanity, the Department of Health says jump and we don’t have the backbone to question or fight back. If we don’t begin to take back the controls of our profession, we will become mere technicians working on the government dole.
Another wrote:
What makes this particularly difficult is that [P4P] was not imposed on us against our will. We, through our professional societies, have adopted it voluntarily. If we could simply band together and fight the external enemy who did this to us, I would have high hopes. Since we did it to ourselves, the solution will be orders of magnitude more difficult.
Doctors have seldom been rewarded for excellence, at least not in any tangible way. In medical school there are tests, board exams, and lab practicals, but once you go into clinical practice, these traditional measures fall away. Whether pay for performance can remedy this problem is still unclear. But from what I learned in my first year as an attending, it has the potential to compromise patient care in unexpected ways.
* * *
What then is the solution to health care overuse? One option is to hire doctors as employees and put them on a salary, as they do at the Mayo and Cleveland clinics, taking away the financial incentive to overtest. Chronically ill patients in the last two years of life cost Medicare tens of thousands of dollars less when treated at the Cleveland Clinic, where teams of doctors follow established best-practice models, than at many other medical centers. Nevertheless, many self-employed doctors recoil at the idea of institutional employment and intrusion on their decision-making authority. Another option is to use bundled payments. A major driver of overutilization is that doctors are paid piecework. There is less of an incentive to increase volume if payments are packaged (e.g., for an entire hospitalization) rather than discrete for every service. Yet another possibility is “accountable care organizations” advanced by Obamacare, in which teams of doctors would be responsible (and paid accordingly) for their patients’ clinical outcomes. Of course, such a scheme would force doctors to work together and to coordinate care. Unfortunately, most doctors, notoriously independent and already smothered in paperwork, have generally performed poorly in this regard.
Doctored Page 12