Apple Machine Learning Research has a new preprint: “The Illusion of Thinking: Understanding the Strengths and Limitations of Reasoning Models via the Lens of Problem Complexity.” [Apple, PDF] “Lar…
The term “reasoning model” is as gaslighting a marketing term as “hallucination”. When an LLM is “Reasoning” it is just running the model multiple times. As this report implies, using more tokens appears to increase the probability of producing a factually accurate response, but the AI is not “reasoning”, and the “steps” of it “thinking” are just bullshit approximations.
The term “reasoning model” is as gaslighting a marketing term as “hallucination”. When an LLM is “Reasoning” it is just running the model multiple times. As this report implies, using more tokens appears to increase the probability of producing a factually accurate response, but the AI is not “reasoning”, and the “steps” of it “thinking” are just bullshit approximations.
AI agent that can do anything you want!
looks inside
state machines and if statements