Me vais a perdonar por el "off-topic", pero acabo de leer este artículo sobre errores humanos en medicina que me ha parecido magnífico:
The Wrong Stuff : Risky Business: James Bagian—NASA astronaut turned patient safety expert—on Being Wrong
Siempre me ha parecido que los médicos tendían bastante a taparse los unos a los otros, pero la razón es que, como dice el artículo, los fallos no se analizan para evitar que no ocurran más (o que ocurran lo menos posible), sino que directamente se dice que ha sido culpa de alguien y se le dice "Ten más cuidado y fíjate más", como si eso fuera a arreglar algo... os traduzco (libremente) un párrafo que creo que resume muy bien el sentido del artículo:
Tomemos un ejemplo muy sencillo: Una enfermera le da al paciente de la cama A la medicina del paciente de la cama B. ¿Qué dirías? ¿"La enfermera se ha equivocado"? Eso es cierto, pero ¿cual es la solución? ¿"Enfermera, por favor, sea más cuidadosa"? Decir a la gente que tenga cuidado no es efectivo. Los humanos no funcionan así; algunos son mejores que otros, pero nadie es perfecto. Necesitamos una solución que no consista en hacer a la gente perfecta.
Por lo tanto, preguntémonos: "¿Por qué la enfermera se equivocó?" Quizá las dos medicinas sean muy parecidas... ese es un problema de empaquetado; podemos solucionarlo. O quizá se espera que la enfermera reparta las píldoras de diez pacientes en cinco minutos. Ese es un problema de organización; podemos solucionarlo. Y estas soluciones pueden tener un enorme impacto. Entre el 7% y el 10% de las medicinas que se administran a pacientes tienen algún problema: medicina incorrecta, dosis incorrecta, paciente incorrecto... pero si introduces un código de barras para la administración de medicinas, la tasa de errores cae a un 0.1%; eso es enorme!!
Además, me encanta este artículo porque no sólo es aplicable a la medicina; para casi cualquier aspecto de la vida, si cuando algo falla dejamos de pensar en quién la ha cagado, y pensamos en por qué ha fallado y qué podemos hacer para que el error se repita lo menos posible, el mundo sería un lugar muchísimo mejor...
Para quien no tenga problemas con el inglés, y a riesgo de que la Sinde decida chaparme el blog por fusilar contenidos ajenos, os copio el artículo íntegro... estoy seguro de que, con todos los enlaces y atribuciones correctamente puestos, a Kathryn Schulz no le importará que dé a conocer su trabajo entrevistando a James Bagian:
In the weeks after I launched this series, several readers e-mailed me to suggest that I interview a man named James Bagian. When I began looking into his background, it became clear to me why: Name a high-stakes industry, and odds are Bagian has been involved in trying to make it safer. He is, among other things, an engineer, an anesthesiologist, a NASA astronaut (he was originally scheduled to be on the fatal Challenger mission), a private pilot, an Air Force-qualified freefall parachutist, and a mountain rescue instructor. And then there's his current job: director of the Veteran Administration's National Center for Patient Safety. In that capacity, Bagian is responsible for overseeing the reduction and prevention of harmful medical mistakes at the VA's 153 hospitals.
Given that most of us are far more likely to find ourselves in a health clinic than a space shuttle, it's sobering to hear Bagian compare the overall attitude toward error in his various fields. "If you look at the percent of budget we spend on safety activity in healthcare versus, say, nuclear power or aviation or the chemical industry, it's not even close," he told me. "Granted, that's just one metric, and I'm not saying money is the be-all end-all. But if people in industry look at what happens in healthcare, they say, ‘Man, this doesn't look like anything we recognize.'" In the below interview, Bagian and I talk about how to make medicine safer, why he doesn't like the word "error," and what it was like to dodge the Challenger bullet.
***
How does the healthcare industry compare to engineering and aeronautics when it comes to dealing with human error?
Not favorably. Much of my background is in what's called high-reliability industries—the ones that operate under conditions of high hazard yet seldom have a bad event—and people in those fields tend to have a systems perspective. We're not terribly interested in what some individual did. We want to know what led up to a bad event and what changes we need to make to reduce the likelihood of that event ever happening again.
When I got into healthcare, I felt like I'd stepped into an entirely different world. It was all about, "Let's figure out who screwed up and blame them and punish them and explain to them why they're stupid." To me, it's almost like whistling past the grave. When we demonize the person associated with a bad event, it makes us feel better. It's like saying, "We're not stupid so it won't happen to us." Whereas in fact it could happen to us tomorrow.
Why do you think healthcare lags so far behind in this respect?
For one thing, in healthcare there's tons of variation, in both biology and behavior, so physicians are rightly skeptical of the cookie-cutter approach. They think you have to tailor everything to the individual. There's some truth to that, but the tailoring should be based on what helps the patient, not on your own personal preference.
And then, too, medicine is much older than these other fields, eons old, and for most of that time there wasn't PubMed or the AMA or what have you. It was all about the expertise of the individual practitioner. It's a short step from there to assuming that problems in medicine stem from problematic individuals. That's why we have this whole "train and blame" mentality in medical culture; someone makes a mistake, you train them not to do it anymore, and then you punish them if it happens again. I think we've ridden that horse about as far as we can.
That suggests that the biggest obstacle to reducing medical error is medical culture, rather than our understanding of the human body or the quality of the available technologies and treatments.
It's all those things, but first and foremost, yes, it's cultural. But I should say before we go any further that I don't usually use the term "error." For starters, it distracts people from the real goal, which isn't reducing error but reducing harm. And it also feeds into precisely the cultural problem we're discussing. It has a punitive feel, and it suggests that the right answer was available at the time, which isn't always the case.
I appreciate that attitude, but some things really are medical errors, right? Bad outcomes don't only happen because a certain piece of information was unknowable or a certain event was unforeseeable. Sometimes doctors just write the wrong prescriptions or operate on the wrong body parts.
That's true, but if at the end of the day all you can say is, "So-and-so made a mistake," you haven't solved anything. Take a very simple example: A nurse gives the patient in Bed A the medicine for the patient in Bed B. What do you say? "The nurse made a mistake"? That's true, but then what's the solution? "Nurse, please be more careful"? Telling people to be careful is not effective. Humans are not reliable that way. Some are better than others, but nobody's perfect. You need a solution that's not about making people perfect.
So we ask, "Why did the nurse make this mistake?" Maybe there were two drugs that looked almost the same. That's a packaging problem; we can solve that. Maybe the nurse was expected to administer drugs to ten patients in five minutes. That's a scheduling problem; we can solve that. And these solutions can have an enormous impact. Seven to 10 percent of all medicine administrations involve either the wrong drug, the wrong dose, the wrong patient, or the wrong route. Seven to 10 percent. But if you introduce bar coding for medication administration, the error rate drops to one tenth of one percent. That's huge.
If the biggest obstacles to improving medical safety are cultural, how do you go about changing the culture?
Some of it is about philosophy. We're very, very clear about the fact that patient safety is not someone else's issue. We say, "Everyone here is your patient." Even if they're not directly yours, they're being cared for by this organization, and if you see something that puts someone at greater risk, it is your moral responsibility to intervene with their caregiver to make sure the right thing happens. You don't just say "That's not my business." Baloney. If it was your kid, would you get involved? Then why don't you do it for this patient?
And some of it is about tools. You can't change the culture by saying, ‘Let's change the culture.' It's not like we're telling people, "Oh, think in a systems way." That doesn't mean anything to them. You change the culture by giving people new tools that actually work. The old culture has tools, too, but they're foolish: "Be more careful," "Be more diligent," "Do a double-check," "Read all the medical literature." Those kinds of tools don't really work.
What kinds of tools have you introduced that do work?
One thing we do that's unusual is we look at close calls. In the beginning, nobody did that in healthcare. Even today probably less than 10 percent of hospital facilities require that close calls be reported, and an even smaller percentage do root cause analyses on them. At the VA, 50 percent of all the root cause analyses we do are on close calls. We think that's hugely important. So does aviation. So does engineering. So does nuclear power. But you talk to most people in healthcare, they'll say, "Why bother? Nothing really happened. What's the big deal?"
How do you get people to tell you about their close calls, or for that matter about their actual errors? Getting people to report problems has always been tricky in medicine.
Yeah, reporting is a huge issue, because obviously you can't fix a problem if you don't know about it. Back in 1998, we conducted a huge cultural survey on patient safety, and one of the questions we asked was, "Why don't you report?" And the major reason—most people think it's going to be fear of malpractice or punishment, but it wasn't those. It was embarrassment, humiliation. So the question became, How do you get people to not be afraid of that? We talked about it a lot, and we devised what we called a blameworthy act, which we defined as possessing one of the following three characteristics: it involves assault, rape, or larceny; the caregiver was drunk or on illicit drugs; or he or she did something that was purposefully unsafe. If you commit a blameworthy act, that's not a safety issue, although it might manifest as one. That's going to get handled administratively, and probably you should be embarrassed. But we made it clear that blameworthiness was a very narrow case.
At the time that we conducted this survey, we were already considered to be a good reporting healthcare organization; our people reported more than in most places. But in the ten months after we implemented this definition, our reporting went up 30 fold. That's 3,000 percent. And it has continued to go up ever since—not as dramatically, but a couple of percentage points every year.
It's pretty sobering that the reporting rate can go up so much. I realize that that's good news, but it also suggests that there was (and to a lesser extent presumably still is) a lot of bad stuff going on out there that we never hear about.
That's true. But the only reason to have reporting is to identify vulnerabilities, not to count the number of incidents. Reports are never good for determining incidence or prevalence, because they're essentially voluntary. Even if you say "You must report," people will only report when they feel like it's in their interest to do so.
Do you punish people for failing to report serious medical issues?
No. In theory, punishment sounds like a good idea, but in practice, it's a terrible one. All it does is create a system where it's not in people's interest to report a problem.
What about public reporting? If the primary purpose of reporting is to identify vulnerabilities, is there any value to making such reports public? There certainly seems to be some movement in that direction within healthcare.
It depends what you're reporting. If you look at our Web site and publications, you'll see that we post risks and advisories, we're open about the problems we have and the steps we need to take. And we don't mince words; it's not like we're afraid to talk about these things.
But the reports that come in raw from the field—I don't think it makes sense to make those public. They're too misleading. People don't understand what they mean, they don't have the knowledge and sophistication and opportunity to get the full facts, and the way something looks at first blush is often not how it looks after an investigation.
What about these scorecards and similar public metrics that some states and institutions are now using? Are you in favor of those?
I don't think they're always bad, but I do think they often kid people about what's going on. When people think they're going to be graded, they're very likely to take action to make themselves look good, to give themselves a business advantage. And that can be dangerous.
Let me give you an analogy. In the United States, airlines are legally prohibited from advertising based on their safety record. The feeling was, if you let airlines compete for customers based on safety, there will be an incentive not to report problems. Suppose I work for an airlines where the ad campaign is "We're the safest company, we've never had a flight canceled for maintenance problems." And I'm a mechanic and I see a maintenance issue and I think: "We should hold this flight." But someone above me is saying, "This is going to destroy our advertising campaign, this is an investor-owned airline" and so forth. So I say, "Well, maybe it's not that big of a deal, we'll catch it at the next scheduled maintenance instead of dealing with it now." Do we really want to create that kind of perverse incentive?
On the subject of public awareness of medical safety, I want to ask you about some recent incidents at VA facilities, such as the non-sterile equipment that might have exposed patients to HIV and hepatitis.
Other places have had the same thing happen or worse and done nothing about it. At the Senate hearings, people from other medical systems showed up and said, "Oh, this stuff happens all the time, it's just that the VA told you about it." The joint commissioner said, "The VA's done more than anybody in terms of looking at this and making it better." You have to have a thick enough hide to tolerate some of the unsophisticated responses to the fact that you're publicizing a problem. All those responses do is make some managers who don't have as much courage say "Let's not talk about this in the future." And that means the problems don't get fixed.
What about the VA facility in Philadelphia with the so-called "rogue cancer unit," where almost a hundred patients received inappropriate radiation treatment?
The fact is, there was not a robust quality-control system in place for that kind of treatment—not in Philly and not anywhere. The profession didn't even have standards about how to decide what amount of radioactive seed to give and how to follow up. Penn, which was providing the service, said, "There's no standard, so we didn't violate it." And we said, "Well, there should be a standard, and we should've been enforcing it. We should have stepped into the gap in healthcare, as we've done in many other situations." We hadn't. Now we have. We went and changed it all, and along the way we took our lumps in the press.
But here again, the important fact is that we didn't say, "Let's not talk about it." The easiest thing would have been to fix this one problem and not make a big deal of it and let everybody else fend for themselves. We didn't take that approach. I think that's a good thing. Does it hurt the organization in some ways when people read about it and say, "Oh, look at what's happening in the VA?" A little. But what they don't know is that the same thing or worse is happening in their own hospital.
As a government institution, is the VA legally required to behave differently in terms of reporting and investigating problems than private hospitals and healthcare systems?
No. Nobody told us we had to look at close calls or tell people they can sue us or any of that. It's just what we decided to do. In the VA system, we ask, "What's the right thing for the patient?" That's what guides us. Whereas people in the private sector sometimes say, "We could lose market share if we talk about this publically, let's not do it."
It's got to be traumatic to be a healthcare provider who is involved in a major medical error. How do you support practitioners in that situation?
We don't deal with that on the safety side. In my opinion, it's a nice thing to do, but it's not the major issue. Quite honestly, I think: "Get over it and grow up." I come from aviation, and we don't have pilot support groups. Would it be helpful to have them? Maybe a little. But I think the far more important thing is: Don't blame people where they shouldn't be blamed. Don't humiliate them publically. Don't disclose who they are if it wasn't an intentional act. And show them that the problem they reported was fixed—that it was worth taking that risk, because it made things better for other patients.
Let's talk about the patients. What do you do for victims of medical error?
If a patient is harmed by something we've done, we tell them. We explain what happened, we tell them that they're eligible for monetary compensation, and we tell them they can sue us. I don't know any other place that says, "Here's how to bring a tort claim against us." We do. We figure that if we hurt you, whether through malfeasance or not, we should make restitution.
Do you get sued a lot?
There's a ton of information in the malpractice literature about what are called closed claims—the ones that are resolved in court—and what you see is that when the patient feels like they've been dealt with less than candidly, that's when they really go in for the kill. It's like, "Okay, if that's the way you treated me, let's see who pays in the end." It becomes about getting even, which I can understand.
So we make it easy. We just tell them. And we end up getting more torts filed, but the aggregate payment is less, because people aren't trying to get revenge. Most people just want us to pay for something specific, to take care of the problem we created. And a lot of people say, "Thanks for telling me, I'm not glad it happened, but I understand that it wasn't intentional." And that's that.
Let me shift gears for a bit and ask you a couple of questions about your career as an astronaut. Given that, as you've said, aviation is historically far better about safety issues than medicine, what kinds of things still go wrong up there?
You can never make the probability of failure zero. You can make it really low, but you can never make it vanish. In a high-stakes, high-value system like a space shuttle, we go to great lengths to understand, say, the failure probability for a valve or a fitting or bolt. And then we do what's called a probabilistic risk assessment: we put all of those probabilities together and say, "What level of confidence do we want to have that we won't have a catastrophe?" And management has to decide what that level is. It really comes down to risk—to how much of it we're willing to accept.
It seems that NASA's been willing to accept a fair amount of it, given the tragedies that have bedeviled our space program.
NASA has been terrible on this. Not because of how much risk they're willing to accept, and not because they don't do the work to understand it. They always know what the probability of failure is. But, historically at least, they haven't been honest and forthright in taking to the public about it. In the early '80s, before the Challenger accident, they would say—and this is where I think they actually lied, I don't think that's too strong a word—they would say, "Flying the shuttle is like flying a 727 to Disney World."
That's absurd. Not only are you more likely to get killed in the shuttle than in an airplane; you're more likely to get killed going up once in the shuttle than if you had flown combat missions for two years in Vietnam. That's a statistical fact, but NASA doesn't make it clear. They might tell the House [of Representatives] that there's a 1.5 percent failure rate, but most Americans don't understand what that means. I mean, 1.5, what is that? Is that a lot? You have to relate it to something that means something to somebody. Otherwise, people have the perception that space flight is safe, and when there's an accident, they're shocked. It's like, "We gotta stop flying." If we want to add additional safeguards because now we're feeling emotional about it, okay, we can do that. But if we're still meeting our design specs for loss, why would we stop flying?
Pretty much everyone who lived in the U.S. at the time can tell you exactly where they were and what they were doing when they found out about the Challenger disaster. You were supposed to be on the Challenger—and then a few months before the fatal mission, your crew was switched out. Where were you instead?
I was there. I helped get everything ready. I babysat the vehicle during tanking. I was at the pad in case there was a pad emergency. When it happened, I was looking at my watch, because every time the shuttle launched, I would think, "I wonder if we're going to lose it during launch this time." That wasn't a fleeting thought, like a low-probability thing that just crossed my mind. It was something I thought seriously about every single time. The riskiest time is between throttle down and throttle up, approximately between 30 seconds and 75 seconds into the mission. So I'm looking at my watch and I'm like, "Okay, we're back to full throttle and it didn't blow. We're over the hump." And then a second later, it went off.
How did you feel?
Was I sad that it happened? Of course. Was I surprised? Not really. I knew it was going to happen sooner or later—and not that much later. At the time, the loss rate was about 4 percent, or one in 25 missions. Challenger was the 25th mission. That's not how statistics works, of course—it's not like you're guaranteed to have 24 good flights and then one bad one, it just happened that way in this case—but still, you think we're going to fly a bunch of missions with a 4 percent failure rate and not have any failures? You gotta be kidding. Anybody who's a realist knows you're going to have losses. Even at 1.8 percent, which is the estimated failure rate these days—how many missions did we go? We went another 70-some missions and had another loss. Well, we were looking at a one in 80 loss rate. That's right on schedule.
That's a really high risk level. Handling it institutionally and politically is one thing, but how do you handle it emotionally?
Everybody's different. People who hadn't been around the high risk stuff themselves, it changed their whole appetite for it. Others looked at it much as I did: It's a shame but it happens, let's go on. I had worked at a test pilot school and some of my best friends were killed while I was there, so it was not an abstract concept to me that people I worked with would be killed doing the job I do.
You were part of the team that investigated the Challenger accident. Were you satisfied with how that investigation was handled?
Overall I didn't have big problems with it. But one thing that was deliberately buried was what happened to the crew. I did that part of the investigation, and there was tremendous political pressure not to tell anyone what happened—not even the other people in the crew office. They didn't learn for months, which was totally inappropriate. They wouldn't even let us put in checklists about what to do in the case of a breakup similar to Challenger. There's ways you could probably survive it, but politically we weren't allowed to discuss that for years, which to me is total hogwash. There are still many people that don't understand that the crew of the Challenger didn't die until they hit the water. They were all strapped into their seats in a basically intact crew module; their hearts were still beating when they hit the water. People think they were blown to smithereens, but that's not what happened. And the problem with that is the same one we were talking about with regard to medicine: if you don't learn what you can from a tragedy, you can't mitigate that risk in the future.
If you could hear someone else interviewed about wrongness, who would it be?
I've been thinking about these issues one way or another for my entire adult life and I've talked to most of the major hitters, so I'm guessing I'm not going to hear much that strikes me as new. But I would be interested to hear what the president [Obama] thinks when outcomes are less than what would be desired.
Kathryn Schulz is the author of Being Wrong: Adventures in the Margin of Error. She can be reached at [email protected]. You can follow her on Facebook here, and on Twitter here.
This interview is part of a series of Q and As in which notable people discuss their relationship to being wrong. You can read past interviews with hedge fund manager Victor Niederhoffer, mountaineer Ed Viesturs, This American Life host Ira Glass, celebrity chef Anthony Bourdain, Sports Illustrated senior writer Joe Posnanski, education scholar and activist Diane Ravitch, and criminal defense lawyer and pundit Alan Dershowitz.