Ai's Not 'reasoning' At All - How This Team Debunked The Industry Hype

Trending 1 day ago
1acolors-gettyimages-1490504801
Pulse/Corbis via Getty Images

Follow ZDNET: Add america arsenic a preferred source connected Google.


ZDNET's cardinal takeaways

  • We don't wholly cognize really AI works, truthful we ascribe magical powers to it.
  • Claims that Gen AI tin logic are a "brittle mirage."
  • We should ever beryllium circumstantial astir what AI is doing and debar hyperbole.

Ever since artificial intelligence programs began impressing nan wide public, AI scholars person been making claims for nan technology's deeper significance, moreover asserting nan imaginable of human-like understanding. 

Scholars wax philosophical because moreover nan scientists who created AI models specified arsenic OpenAI's GPT-5 don't really understand really nan programs activity -- not entirely. 

Also: OpenAI's Altman sees 'superintelligence' conscionable astir nan area - but he's short connected details

AI's 'black box' and nan hype machine

AI programs specified arsenic LLMs are infamously "black boxes." They execute a batch that is impressive, but for nan astir part, we cannot observe each that they are doing erstwhile they return an input, specified arsenic a punctual you type, and they nutrient an output, specified arsenic nan assemblage word insubstantial you requested aliases nan proposal for your caller novel.

In nan breach, scientists person applied colloquial position specified arsenic "reasoning" to picture nan measurement nan programs perform. In nan process, they person either implied aliases outright asserted that nan programs tin "think," "reason," and "know" successful nan measurement that humans do. 

In nan past 2 years, nan rhetoric has overtaken nan subject arsenic AI executives person utilized hyperbole to twist what were elemental engineering achievements. 

Also: What is OpenAI's GPT-5? Here's everything you request to cognize astir nan company's latest model

OpenAI's press merchandise past September announcing their o1 reasoning exemplary stated that, "Similar to really a quality whitethorn deliberation for a agelong clip earlier responding to a difficult question, o1 uses a concatenation of thought erstwhile attempting to lick a problem," truthful that "o1 learns to hone its concatenation of thought and refine nan strategies it uses."

It was a short measurement from those anthropomorphizing assertions to each sorts of chaotic claims, specified arsenic OpenAI CEO Sam Altman's comment, successful June, that "We are past nan arena horizon; nan takeoff has started. Humanity is adjacent to building integer superintelligence."

(Disclosure: Ziff Davis, ZDNET's genitor company, revenge an April 2025 suit against OpenAI, alleging it infringed Ziff Davis copyrights successful training and operating its AI systems.)

The backlash of AI research

There is simply a backlash building, however, from AI scientists who are debunking nan assumptions of human-like intelligence via rigorous method scrutiny. 

In a insubstantial published past period connected nan arXiv pre-print server and not yet reviewed by peers, nan authors -- Chengshuai Zhao and colleagues astatine Arizona State University -- took isolated nan reasoning claims done a elemental experiment. What they concluded is that "chain-of-thought reasoning is simply a brittle mirage," and it is "not a system for genuine logical conclusion but alternatively a blase shape of system shape matching." 

Also: Sam Altman says nan Singularity is imminent - here's why

The word "chain of thought" (CoT) is commonly utilized to picture nan verbose watercourse of output that you spot erstwhile a ample reasoning model, specified arsenic GPT-o1 aliases DeepSeek V1, shows you really it useful done a problem earlier giving nan last answer.

That watercourse of statements isn't arsenic heavy aliases meaningful arsenic it seems, constitute Zhao and team. "The empirical successes of CoT reasoning lead to nan cognition that ample connection models (LLMs) prosecute successful deliberate inferential processes," they write. 

But, "An expanding assemblage of analyses reveals that LLMs thin to trust connected surface-level semantics and clues alternatively than logical procedures," they explain. "LLMs conception superficial chains of logic based connected learned token associations, often failing connected tasks that deviate from commonsense heuristics aliases acquainted templates."

The word "chains of tokens" is simply a communal measurement to mention to a bid of elements input to an LLM, specified arsenic words aliases characters. 

Testing what LLMs really do

To trial nan presumption that LLMs are simply pattern-matching, not really reasoning, they trained OpenAI's older, open-source LLM, GPT-2, from 2019, by starting from scratch, an attack they telephone "data alchemy."

arizona-state-2025-data-alchemy
Arizona State University

The exemplary was trained from nan opening to conscionable manipulate nan 26 letters of nan English alphabet, "A, B, C,…etc." That simplified corpus lets Zhao and squad trial nan LLM pinch a group of very elemental tasks. All nan tasks impact manipulating sequences of nan letters, specified as, for example, shifting each missive a definite number of places, truthful that "APPLE" becomes "EAPPL."

Also: OpenAI CEO sees uphill struggle to GPT-5, imaginable for caller benignant of user hardware

Using nan constricted number of tokens, and constricted tasks, Zhao and squad alteration which tasks nan connection exemplary is exposed to successful its training information versus which tasks are only seen erstwhile nan vanished exemplary is tested, specified as, "Shift each constituent by 13 places." It's a trial of whether nan connection exemplary tin logic a measurement to execute moreover erstwhile confronted pinch new, never-before-seen tasks. 

They recovered that erstwhile nan tasks were not successful nan training data, nan connection exemplary grounded to execute those tasks correctly utilizing a concatenation of thought. The AI exemplary tried to usage tasks that were successful its training data, and its "reasoning" sounds good, but nan reply it generated was wrong. 

As Zhao and squad put it, "LLMs effort to generalize nan reasoning paths based connected nan astir akin ones […] seen during training, which leads to correct reasoning paths, yet incorrect answers."

Specificity to antagonistic nan hype

The authors tie immoderate lessons. 

First: "Guard against over-reliance and mendacious confidence," they advise, because "the expertise of LLMs to nutrient 'fluent nonsense' -- plausible but logically flawed reasoning chains -- tin beryllium much deceptive and damaging than an outright incorrect answer, arsenic it projects a mendacious aura of dependability."

Also, effort retired tasks that are explicitly not apt to person been contained successful nan training information truthful that nan AI exemplary will beryllium stress-tested. 

Also: Why GPT-5's rocky rollout is nan reality cheque we needed connected superintelligence hype

What's important astir Zhao and team's attack is that it cuts done nan hyperbole and takes america backmost to nan basics of knowing what precisely AI is doing. 

When nan original investigation connected chain-of-thought, "Chain-of-Thought Prompting Elicits Reasoning successful Large Language Models," was performed by Jason Wei and colleagues astatine Google's Google Brain squad successful 2022 -- investigation that has since been cited much than 10,000  times -- nan authors made nary claims astir existent reasoning. 

Wei and squad noticed that prompting an LLM to database nan steps successful a problem, specified arsenic an arithmetic connection problem ("If location are 10 cookies successful nan jar, and Sally takes retired one, really galore are near successful nan jar?") tended to lead to much correct solutions, connected average. 

google-2022-example-chain-of-thought-prompting
Google Brain

They were observant not to asseverate human-like abilities. "Although concatenation of thought emulates nan thought processes of quality reasoners, this does not reply whether nan neural web is really 'reasoning,' which we time off arsenic an unfastened question," they wrote astatine nan time. 

Also: Will AI deliberation for illustration humans? We're not moreover adjacent - and we're asking nan incorrect question

Since then, Altman's claims and various property releases from AI promoters person progressively emphasized nan human-like quality of reasoning utilizing casual and sloppy rhetoric that doesn't respect Wei and team's purely method description. 

Zhao and team's activity is simply a reminder that we should beryllium specific, not superstitious, astir what nan instrumentality is really doing, and debar hyperbolic claims. 

More