• Barbarian@lemmy.reckless.dev
    link
    fedilink
    English
    arrow-up
    5
    ·
    1 year ago

    So, there’s a fundamental issue here. A lot of the systems that Amanda is talking about aren’t actually AI.

    Chat-GPT, contrary to the blogosphere, is not actually AI. It does not have the capability for thought. It doesn’t have the capacity to understand truth or fiction as concepts, let alone tell them apart.

    Chat-GPT and similar systems are probabilistic language models. Essentially, I start it off with sentences (a list of tokens, if you want to get technical). Then it responds by essentially looking at the training data it’s been supplied with and picking out the sequence of tokens that most likely is the answer the user is expecting, given the input. Notice that bolded text? The user is expecting. Not anything else. These language models are trained to spit out what users expect, nothing more, nothing less. If a user doesn’t like the response, they give a thumbs down and the model recalibrates, introducing more noise and randomness into the result.

    These language models are actually really great at reducing manual labor at certain tasks (writing cover letters, delivering predictable essays, I’ve personally used Chat-GPT for Shadowrun world-building) but they need to have a knowledgeable person using them because they absolutely will not reliably say true things. They will say whatever their training data says is the most likely thing the user is asking for.

    • cavemeat@beehaw.org
      link
      fedilink
      English
      arrow-up
      3
      ·
      1 year ago

      A lot of the talk around AI has made this mistake. I think its in part because Chat-GPT sounds very smart and human, especially to average people.