• drspod@lemmy.ml
    link
    fedilink
    arrow-up
    56
    ·
    3 days ago

    The problem is that an LLM is a language model, not an objective reality model, so the best it can do is estimate the probability of a particular sentence appearing in the language, but not the probability that the sentence represents a true statement according to our objective reality.

    They seem to think that they can use these confidence measures to filter the output when it is not confident of being correct, but there are an infinite number of highly probable sentences in a language which are false in reality. An LLM has no way of distinguishing between unlikely and false, or between likely and true.

    • nymnympseudonym@lemmy.world
      link
      fedilink
      arrow-up
      3
      arrow-down
      13
      ·
      3 days ago

      language model, not an objective reality model

      Sort of. It’s a generic prediction model. Multi-modal models work the same way as text-only models in this sense.

      So do organic brains.

      Right now, you are “hallucinating” most of your visual field outside the fovea centralis.

      This aspect of your conscious perceptual system is exactly the same kind of high-dimensional interpolation that ML neural networks do.