Flutterby™! : Simulated reason models simulate reasoning

Next unread comment / Catchup all unread comments User Account Info | Logout | XML/Pilot/etc versions | Long version (with comments) | Weblog archives | Site Map | | Browse Topics

Simulated reason models simulate reasoning

2025-07-22 19:31:45.47121+02 by Dan Lyke 0 comments

New Apple study challenges whether AI models truly “reason” through problems

In early June, Apple researchers released a study suggesting that simulated reasoning (SR) models, such as OpenAI's o1 and o3, DeepSeek-R1, and Claude 3.7 Sonnet Thinking, produce outputs consistent with pattern-matching from training data when faced with novel problems requiring systematic thinking. The researchers found similar results to a recent study by the United States of <span data-huuid="4951512150826999828">America Mathematical Olympiad</span> (USAMO) in April, showing that these same models achieved low scores on novel mathematical proofs.

The article is pretty good, rounding up a bunch of pro and con arguments.

[ related topics: Apple Computer Invention and Design Theater & Plays Television Mathematics Macintosh Artificial Intelligence ]

comments in ascending chronological order (reverse):

Comment policy

We will not edit your comments. However, we may delete your comments, or cause them to be hidden behind another link, if we feel they detract from the conversation. Commercial plugs are fine, if they are relevant to the conversation, and if you don't try to pretend to be a consumer. Annoying endorsements will be deleted if you're lucky, if you're not a whole bunch of people smarter and more articulate than you will ridicule you, and we will leave such ridicule in place.


Flutterby™ is a trademark claimed by

Dan Lyke
for the web publications at www.flutterby.com and www.flutterby.net.