• Xerxos@lemmy.ml
    link
    fedilink
    arrow-up
    45
    arrow-down
    1
    ·
    14 hours ago

    There was a paper about this not long ago. The problem is, how LLMs get trained: a right answer gets a point, everything else gets no points. This rewards guessing (produces a point sometimes) over answering “I don’t know/I can’t do this” (produces never a point)

    • ipkpjersi@lemmy.ml
      link
      fedilink
      arrow-up
      9
      ·
      9 hours ago

      It’s like when developers give a wrong answer during technical interviews, rather than say “I’d have to look it up” or “I’d have to check the documentation” etc.