The Illusion of Thinking

← Back to logbook

Gonna toot my own horn here: Apple just dropped this paper today: "The Illusion of Thinking: Understanding the Strengths and Limitations of Reasoning Models via the Lens of Problem Complexity" which hits on a LOT of points I just made on my post yesterday. Good timing i guess

A tl/dr of the paper is that Apple did some research on large reasoning models (LRMs) in controllable puzzle environments like Tower of Hanoi, river crossing, checkers, etc. They found that they fail to have generalizable reasoning capabilities in higher complexity problem spaces and struggle with overthinking in low complexity problems. They did some unique work looking at reasoning traces (not just output) to see how the model can follow instructions and move step-by-step in these problems, allowing them to see surprisingly poor instruction-following capabilities.

Similar to what I said yesterday, they confirmed:

Caveats:

Future considerations:

Some relevant papers that also are skeptical of reasoning models n stuff: