Become a fan of Slashdot on Facebook

 



Forgot your password?
typodupeerror
×
AI IBM Medicine Software Hardware Technology

IBM Watson Reportedly Recommended Cancer Treatments That Were 'Unsafe and Incorrect' 103

An anonymous reader quotes a report from Gizmodo: Internal company documents from IBM show that medical experts working with the company's Watson supercomputer found "multiple examples of unsafe and incorrect treatment recommendations" when using the software, according to a report from Stat News. According to Stat, those documents provided strong criticism of the Watson for Oncology system, and stated that the "often inaccurate" suggestions made by the product bring up "serious questions about the process for building content and the underlying technology." One example in the documents is the case of a 65-year-old man diagnosed with lung cancer, who also seemed to have severe bleeding. Watson reportedly suggested the man be administered both chemotherapy and the drug "Bevacizumab." But the drug can lead to "severe or fatal hemorrhage," according to a warning on the medication, and therefore shouldn't be given to people with severe bleeding, as Stat points out. A Memorial Sloan Kettering (MSK) Cancer Center spokesperson told Stat that they believed this recommendation was not given to a real patient, and was just a part of system testing.

According to the report, the documents blame the training provided by IBM engineers and on doctors at MSK, which partnered with IBM in 2012 to train Watson to "think" more like a doctor. The documents state that -- instead of feeding real patient data into the software -- the doctors were reportedly feeding Watson hypothetical patients data, or "synthetic" case data. This would mean it's possible that when other hospitals used the MSK-trained Watson for Oncology, doctors were receiving treatment recommendations guided by MSK doctors' treatment preferences, instead of an AI interpretation of actual patient data. And the results seem to be less than desirable for some doctors.
This discussion has been archived. No new comments can be posted.

IBM Watson Reportedly Recommended Cancer Treatments That Were 'Unsafe and Incorrect'

Comments Filter:
  • by Crashmarik ( 635988 ) on Wednesday July 25, 2018 @05:45PM (#57009658)

    Really where is the there, here ? You'll have doctors frequently dispute what the correct treatment is and with diseases like cancer it doesn't help that the best you can often do is offer a statistical improvement of someone's chances.

    Far better that more people can afford treatment faster than this remain the province of the priesthood.

    • Cancer is a huge money industry for medicine. This is why the huge focus is on screening / early detection, because those allow tons of unnecessary treatment for perfectly healthy people. People get done with treatment and get told they're in the clear. Everybody is happy and celebrates. Nobody sues for fraud when nothing was wrong in the first place.

      https://qz.com/1335348/google-is-building-virtual-agents-to-handle-call-centers-grunt-work/
    • Comment removed based on user account deletion
    • They found "multiple examples of unsafe and incorrect treatment recommendations". How many exactly, what's the %? What's the relevancy of the "incorrectness" (totally, or mildly?). Doctors have to protect their interests and, probably, discredit AI, thus any mild error is to be publicized. Similar to complaints against Airbnb. Airbnb does close to a million rentals a day, and when an infinitesimal part of that (twice a year) makes trouble, it's largely publicized.
  • by rnturn ( 11092 ) on Wednesday July 25, 2018 @05:51PM (#57009690)

    ... but it will the patient. Is that a problem?"

    Doctor (shaking his head): Yes, Watson... that is a problem.

    (Who trained Watson for this job anyway?)

  • by Tablizer ( 95088 ) on Wednesday July 25, 2018 @05:57PM (#57009714) Journal

    The purpose of such a tool should be to make suggestions that a doctor may not consider themselves. It should be up to the doctor(s) to vet the suggestions or leads before any treatment is actually rendered. A doctor would have to be born in Stupidville to accept bot suggestions as-is.

    • This is why you want Dr Who, not Dr Watson.

      Dr Who knows how to use a screwdriver, and she does it much better than Dr Watson does.

      • by Tablizer ( 95088 )

        Dr. Who only knows how to use a sonic screwdriver. A muggle's screwdriver baffles the daylights out of her/him/it.

        • Dr. Who only knows how to use a sonic screwdriver. A muggle's screwdriver baffles the daylights out of her/him/it.

          She's The Doctor, not an Engineer.

      • I was going to say they need to make an IBM House.

        You'd end up with suggestions like punching the patients in the face, or abusing the staff. Good times.

    • That is in fact what it does

      • by Tablizer ( 95088 )

        It's ultimately what the doctor does with the info that really matters. I would hope they are properly trained to use the system and know its limitations. Disclaimer notices wouldn't hurt as reminder.

  • Really no surprise (Score:5, Interesting)

    by gweihir ( 88907 ) on Wednesday July 25, 2018 @05:59PM (#57009720)

    This is a statistics-driven automaton that has zero insight or understanding. Calling it "AI" is a marketing lie, even if the AI field has given in and calls things like this "weak AI", which is the AI without "I". As such, this machine can find statistical correlations, but it cannot do plausibility checks, because that requires insight. It cannot do predictions either, because that also requires insight. The real strength of Watson (and it is quite an accomplishment) is that unlike older comparable systems, you can feed the training data and the queries into it in natural language. This means you can train a lot cheaper, but at the cost of accuracy, as the effect described in the story nicely shows.

    It is time for this "AI" hype to die down. All it shows is that many people do not chose to use what they have in general intelligence and rather mindlessly follow a crows of cheer-leaders.

    • by Tablizer ( 95088 ) on Wednesday July 25, 2018 @06:07PM (#57009756) Journal

      Calling it "AI" is a marketing lie

      In practice the term "AI" is vague and continuous rather than a Boolean designation ("is" versus "is-not"). The term is not worth sweating over. The exception may be if you are making a big purchase and/or investment based on something being "AI". In that case, inspect it carefully rather than assume something with "AI" is smart and/or useful. But that's good advice for any significant purchase: test drive it & ask detailed questions rather than rely on the brochure.

      • by gweihir ( 88907 ) on Wednesday July 25, 2018 @10:26PM (#57010628)

        It actually is pretty Boolean: Use it for anything real and you are a liar. Because exactly nothing that deserves the description "AI" does exist. Qualify it with "weak" and you use an obviously inappropriate term.

        • by Tablizer ( 95088 )

          Terms are ultimately defined by common usage, not necessarily by what's logical, clear, useful, or fair.

          Defining "natural intelligence" is sticky also. I remember debating for weeks over what "intent" means. Great nerdy fun. (This was before Emailgate, by the way.)

          • by gweihir ( 88907 )

            We are in science and engineering here. Terms have real meaning and are not defined by common use outside of that field.

            • by Tablizer ( 95088 )

              The issue was "AI". If you can supply a precise and unambiguous definition, please do.

              Further, what it means colloquially (regular press) and what it means in technical journals could vary. The audience scope or target thus may also matter.

        • It actually is pretty Boolean: Use it for anything real and you are a liar.

          Who the fuck appointed you the arbitrator of what's "intelligent" and what isn't?

          Besides, anybody who has read your previous posts knows that you consider
          intelligence to be some kind of supernatural hocus-pocus,
          so of course a machine can't have it.

    • by ShanghaiBill ( 739463 ) on Wednesday July 25, 2018 @06:13PM (#57009782)

      As such, this machine can find statistical correlations, but it cannot do plausibility checks, because that requires insight. It cannot do predictions either, because that also requires insight.

      Neither of these require "insight". They just require more data. With enough examples, statistical correlation is all you need.

      • by Anonymous Coward

        You'll never capture everything in the training set.

        In this case what was required is being able to read the medicine's instructions and do some common sense reasoning to see how it's relevant to the patient. Between reading and common sense we're well beyond what Watson is capable of.

      • by gweihir ( 88907 )

        You will never have enough data for that.

      • With enough examples, statistical correlation is all you need.

        A: We have to withhold this treatment because 100% of people with this condition last year died within a month.

        B: Were they treated for it?

        A. No, because we have to withhold treatment.

    • The AI hype is sound and on solid footing compared to the blockchain hype: I've never seen so much effort poured into such a useless technology, cthulu be praised.
      • by gweihir ( 88907 )

        So you think something that does not exist is "solid" in comparison to something that does exist but it pretty useless? Strange priorities you have there...

  • by erp_consultant ( 2614861 ) on Wednesday July 25, 2018 @06:01PM (#57009726)

    I'll take Incorrect Diagnosis for $200, Alex.

  • by greenwow ( 3635575 ) on Wednesday July 25, 2018 @06:01PM (#57009728)

    But is Watson cheaper than a doctor?

  • The survival rate for lung cancer can sometimes be as low as 4% over five years. Even if the drug combination had a 90% chance to outright kill the patient it might raise their overall chances of survival enough to actually be worth the risk. Based on what I know about lung cancer dying from severe hemorrhaging could be preferable to the relatively slow agonizing death some experience otherwise, especially if your overall chances of survival are higher.

  • So? (Score:5, Insightful)

    by 50000BTU_barbecue ( 588132 ) on Wednesday July 25, 2018 @06:06PM (#57009752) Journal

    How many human doctors did the same or worse?

    • Asking society to put its trust in a machine with the justification that at its best it fucks up no more often than some humans at their worst is a non-starter.

      • by yusing ( 216625 )

        Yeah but ... no health benefits! no retirement! no vacations!

        Great deal for the vendors, not so much for their victims.

    • by AHuxley ( 892839 )
      Human doctors face peer review of all work in good advanced teaching hospitals.
      The best teaching hospitals can ensure only a nations very best medical professionals are working every decade.
    • by Anonymous Coward

      "Part of the system testing"

      I think what we are reading is leaked info from someone working on the trail who releases that it's going frighteningly well.

      Basically:
      Watson will get things wrong. Especially in testing. It should not be used on it's own for the foreseeable future. It needs a trained doctor to review the decisions... it is a remarkable assistant. It will only get better and bring a standard of healthcare to a vast number of people who could never afford to access/reach a doctor.

  • by GerryGilmore ( 663905 ) on Wednesday July 25, 2018 @06:10PM (#57009770)
    ...is this: "A Memorial Sloan Kettering (MSK) Cancer Center spokesperson told Stat that they believed this recommendation was not given to a real patient, and was just a part of system testing."
    Isn't this the kind of thing that testing is designed to uncover? It sounds to me like at least this part of the process is working, unlike the asshole who fed the model "fake data".
  • It just wanted to help impose pro-Darwinian responses to malformed genetic abnormalities.

    Next up: self-driving cars that crash on purpose because their passengers sing songs the AI hates.

  • test data provide test results.
  • by Dread_ed ( 260158 ) on Wednesday July 25, 2018 @06:49PM (#57009892) Homepage

    So the data fed to train Watson wasn't from actual cases? Why does it matter what the computer prescribed, then? The system that is Watson is only as good as the data you feed it. Feed it fake information, get not even wrong results. Sounds more like a smear campaign,
      intentionally designed to fail, and certainly not an experiment designed to measure Watson's recommendations against actual doctor recommendations.

    Here's a better idea...

    Feed the damn thing actual patient records with everything included from first immunization to the patient's ultimate death. If you are looking to see if there are any correlations that humans haven't already made you need to feed that sucker as much data as is inhumanly possible and then let it do the work.

    What we have now is a pseudepigrapha of Watson's capabilities. Sure the results are from Watson, but they are not what Watson would do if given accurate, real life data to work with. They made a forgery of the system and put Watson's name on it.

    Shady, bro. Shady...

  • by blindseer ( 891256 ) <blindseer@noSPAm.earthlink.net> on Wednesday July 25, 2018 @06:51PM (#57009902)

    An AI can only be as good as the data used to train it. The article pointed out that Watson was trained using what was possibly based as much on objective data as much as it was on subjective preferences of the physicians that fed it data.

    I recall reading an article about someone doing a study on medical procedures done throughout the USA and they noticed "hot spots" of procedures being done in certain areas. What they found was that in these places they'd see physicians that would recommend procedures out of personal preference. One example was a an area with a lot of tonsillectomies, because a physician felt that any throat infection meant the tonsils had to come out. Another area had an elevated number of hysterectomies, because a physician felt that post-menopause women had an elevated risk of developing cysts and cancers on the uterus. The article went on to say that while such treatments may be unusual no one was willing to consider this malpractice.

    So, Watson recommended a treatment for someone that might aggravate an existing problem of severe bleeding. Is this bad coding for not taking this into account? Or, is there a physician that entered such a prescription for their patient with similar symptoms? It's real difficult to second guess a physician. It's real easy to second guess the computer. Even if both the computer and the human came to the same recommendation for treatment.

    • So, you're holding it wrong?
  • by yusing ( 216625 ) on Wednesday July 25, 2018 @10:40PM (#57010670) Journal

    What would happen if we started calling Ai 'Fake Intelligence' ... Fee Fi Foes?

    As I understand the current fashions, AI has a fatal flaw: it's result is non-deterministic ... noone can be sure how it arrives at an answer. That might be okay for face recognition, or 'computer art' ... but for locating potential automobile collision victims, or deterministically arriving at a sound treatment for a patient? Wrong model.

    I'd guess that the 'expert systems' of 20 years back outperform neural nets. Their logic trees were scrutable.

  • Right now it's still in a early learning proces, and it's a tool to help doctors. So what if it, at this point in development, makes the unsafe/incorrect treatment? It's not like doctors are right all the time, and doctors also have been well know to prescribe wrong treatments. Or, maybe the system did know about it, but calculated the risc factor of the patient dying anyway if he didn't get treatment.
    But we're still at the beginning of having AI determine stuff like this, and yet Watson is already very wel

  • Watson just needs a big data cache of real-life human deaths to learn how to cure cancer.
  • Unless you combine it with dilaftin.
    Which any first-year should know is
    the standard prep medication your patient
    was taking before surgery. Your patient
    should be dead.

  • For the second time today we see evidence that the poor excuse for AI they keep trotting out, in this case probably the most advanced version of it, even, is crap. I maintain that without understanding how a biological brain actually is able to think, there's no way these throw-it-at-the-wall-and-see-if-it-sticks guesses at an approach are going to ever be real AI -- and since we don't have the instrumentality to really truly see how a biological brain works, and map it's connections, in a living subject, w

"The vast majority of successful major crimes against property are perpetrated by individuals abusing positions of trust." -- Lawrence Dalzell

Working...