As AI capabilities advance in complex medical scenarios that doctors face on a daily basis, the technology remains controversial in medical communities.

  • theluddite
    link
    fedilink
    English
    arrow-up
    28
    arrow-down
    4
    ·
    1 year ago

    Ya that’s a fundamental misunderstanding of percentages. For an analogous situation with which we’re all more intuitively familiar, a self driving car that is 99.9% accurate in detecting obstacles crashes into one in one thousand people and/or things. That sucks.

    Also, most importantly, LLMs are incapable of collaboration, something very important in any complex human endeavor but difficult to measures, and therefore undervalued by our inane, metrics-driven business culture. Chatgpt won’t develop meaningful, mutually beneficial relationships with its colleagues, who can ask each other for their thoughts when they don’t understand something. It’ll just spout bullshit when it’s wrong, not because it doesn’t know, but because it has no concept of knowing at all.

    • deweydecibel@lemmy.world
      link
      fedilink
      English
      arrow-up
      10
      arrow-down
      1
      ·
      edit-2
      1 year ago

      It really needs to be pinned to the top of every single discussion around chatgbt:

      It does not give answers because it knows. It gives answers because it thinks it looks right.

      Remember back in school when you didn’t study for a test and went through picking answers that “looked right” because you vaguely remember hearing the words in Answer B during class at some point?

      It will never have wisdom and intuition from experience, and that’s critically important for doctors.

        • ourob@discuss.tchncs.de
          link
          fedilink
          English
          arrow-up
          5
          ·
          1 year ago

          “Looks right” in a human context means the one that matches a person’s actual experience and intuition. “Looks right” in an LLM context means the series of words have been seen together often in the training data (as I understand it, anyway - I am not an expert).

          Doctors are most certainly not choosing treatment based on what words they’ve seen together.