• UlyssesT [he/him]@hexbear.net
    link
    fedilink
    English
    arrow-up
    1
    ·
    edit-2
    2 months ago

    Firstly, how do you filter for medical information in a way that works 100% of the time

    That is a good question, and it goes double for what you’re apparently running interference for. How exactly does a sheer volume of bullshit justify bullshit being generated by LLMs in medical fields?

    You are going to miss a lot of medical questions

    Yes, and the magic of LLMs means those medical questions are going to be missed a lot more often and faster than ever before.

    Having fact checkers will lull people into a false sense of security, which will be very bad when they inevitably get things wrong.

    That’s an amazing take: the errors aren’t as bad as attempts to mitigate the flood of errors from the planet-burning bullshit machines. galaxy-brain

    • gay_king_prince_charles [she/her, he/him]@hexbear.net
      link
      fedilink
      English
      arrow-up
      3
      ·
      edit-2
      2 months ago

      If you see a note saying “This was confirmed to be correct by our well-trained human fact checkers” and one saying “[Gemini] can make mistakes. Check important info.”, you are more likely to believe the first than the second. The solution here is to look at actual articles with credited authors, not to have an army of people reviewing every single medical query.

      • UlyssesT [he/him]@hexbear.net
        link
        fedilink
        English
        arrow-up
        2
        ·
        2 months ago

        I’m still not seeing a safe or even meaningful use case for the treat printers here, especially not for the additional electricity and waste carbon costs. Were medical data queries impossible before LLMs? No, they were not.

        • LLM usage here doesn’t help, that’s true. But medical queries weren’t good before LLM’s either, just because it’s an incredibly complex field with many edge cases. There is a reason self diagnosis is dangerous and it isn’t because of technology.technology.

          • UlyssesT [he/him]@hexbear.net
            link
            fedilink
            English
            arrow-up
            1
            ·
            2 months ago

            LLM usage here doesn’t help

            Yes, I’m glad we can agree on that.

            The rest of what you were saying seems kind of like a pointless derail because you were defending something that’s already indefensibly bad for its supposed use case here.