• invalidusernamelol [he/him]@hexbear.net
    link
    fedilink
    English
    arrow-up
    17
    ·
    edit-2
    12 hours ago

    A nice portion of the technical work that goes into these models is maintenance of the facade.

    If there’s an article written about a specific question, they will really quickly go in and just hardcode an answer to it.

    Situations like this shouldn’t be taken as specific, but as general criticism of the reasoning methodology behind these models that still hasn’t been solved, because the system itself is built in a way that monkey patching statistical anomalies is the only way.

    • ChatGPT-5 though the web has a temperature greater than 0. The correct behavior is more likely a result of being a non-deterministic system than a concentrated effort to crawl the Internet for articles about bugs and hardcore solutions. The first token in this question will be “yes” or “no” and all further output is likely to support that. Because gpt-5 isn’t a CoT model, it can’t mimic knowledge of future tokens and almost has to maintain previous output, so there’s a good chance of going either way.