A new study titled The Illusion of Thinking, questions the very foundation of AI advanced reasoning models, or LRMs (Large Reasoning Models), and their capacity to solve complex problems reliably.
In the groundbreaking report, Apple researchers have revealed troubling limitations in today’s most powerful AI systems, raising fresh concerns about the tech industry’s rapid pursuit of artificial general intelligence (AGI).
While AI models like OpenAI’s o3, Google’s Gemini, and Anthropic’s Claude 3.7 continue to impress users with fluent responses and problem-solving capabilities, Apple’s paper suggests these models may be fundamentally flawed when faced with tasks requiring deep logical reasoning.
Read the Apple research paper: The Illusion of Thinking
Details of “Illusion of Thinking” Study
Apple’s researchers set out to explore whether current large reasoning models could demonstrate scalable, reliable thinking. The results were startling. While LRMs showed strong performance in solving simple problems, they experienced a “complete accuracy collapse” on more complex tasks, such as classic puzzle-based reasoning tests.