It’s pretty easy to see the problem here: The Internet is brimming with misinformation, and most large language models are trained on a massive body of text obtained from the Internet.

Ideally, having substantially higher volumes of accurate information might overwhelm the lies. But is that really the case? A new study by researchers at New York University examines how much medical information can be included in a large language model (LLM) training set before it spits out inaccurate answers. While the study doesn’t identify a lower bound, it does show that by the time misinformation accounts for 0.001 percent of the training data, the resulting LLM is compromised.

  • Pennomi@lemmy.world
    link
    fedilink
    English
    arrow-up
    5
    ·
    20 hours ago

    Even curation seems unlikely to fix the problem. I bet a new algorithm is required that allows LLMs to validate their response before it’s returned. Basically an “inner monologue” to avoid saying stupid things.

    • Ech@lemm.ee
      link
      fedilink
      English
      arrow-up
      4
      arrow-down
      1
      ·
      18 hours ago

      These models are so shit they need a translator. Hilarious.