We collect cookies to analyze our website traffic and performance; we never collect any personal data. Cookie Policy
Accept
NEW YORK DAWN™NEW YORK DAWN™NEW YORK DAWN™
Notification Show More
Font ResizerAa
  • Home
  • Trending
  • New York
  • World
  • Politics
  • Business
    • Business
    • Economy
    • Real Estate
  • Crypto & NFTs
  • Tech
  • Lifestyle
    • Lifestyle
    • Food
    • Travel
    • Fashion
    • Art
  • Health
  • Sports
  • Entertainment
Reading: Pondering Machines challenges OpenAI's AI scaling technique: 'First superintelligence will probably be a superhuman learner'
Share
Font ResizerAa
NEW YORK DAWN™NEW YORK DAWN™
Search
  • Home
  • Trending
  • New York
  • World
  • Politics
  • Business
    • Business
    • Economy
    • Real Estate
  • Crypto & NFTs
  • Tech
  • Lifestyle
    • Lifestyle
    • Food
    • Travel
    • Fashion
    • Art
  • Health
  • Sports
  • Entertainment
Follow US
NEW YORK DAWN™ > Blog > Technology > Pondering Machines challenges OpenAI's AI scaling technique: 'First superintelligence will probably be a superhuman learner'
Pondering Machines challenges OpenAI's AI scaling technique: 'First superintelligence will probably be a superhuman learner'
Technology

Pondering Machines challenges OpenAI's AI scaling technique: 'First superintelligence will probably be a superhuman learner'

Last updated: October 24, 2025 11:51 pm
Editorial Board Published October 24, 2025
Share
SHARE

Whereas the world's main synthetic intelligence firms race to construct ever-larger fashions, betting billions that scale alone will unlock synthetic common intelligence, a researcher at one of many trade's most secretive and worthwhile startups delivered a pointed problem to that orthodoxy this week: The trail ahead isn't about coaching greater — it's about studying higher.

"I believe that the first superintelligence will be a superhuman learner," Rafael Rafailov, a reinforcement studying researcher at Pondering Machines Lab, instructed an viewers at TED AI San Francisco on Tuesday. "It will be able to very efficiently figure out and adapt, propose its own theories, propose experiments, use the environment to verify that, get information, and iterate that process."

This breaks sharply with the strategy pursued by OpenAI, Anthropic, Google DeepMind, and different main laboratories, which have guess billions on scaling up mannequin dimension, knowledge, and compute to attain more and more refined reasoning capabilities. Rafailov argues these firms have the technique backwards: what's lacking from in the present day's most superior AI programs isn't extra scale — it's the power to really be taught from expertise.

"Learning is something an intelligent being does," Rafailov stated, citing a quote he described as not too long ago compelling. "Training is something that's being done to it."

The excellence cuts to the core of how AI programs enhance — and whether or not the trade's present trajectory can ship on its most bold guarantees. Rafailov's feedback supply a uncommon window into the pondering at Pondering Machines Lab, the startup co-founded in February by former OpenAI chief expertise officer Mira Murati that raised a record-breaking $2 billion in seed funding at a $12 billion valuation.

Why in the present day's AI coding assistants overlook all the pieces they discovered yesterday

As an example the issue with present AI programs, Rafailov supplied a state of affairs acquainted to anybody who has labored with in the present day's most superior coding assistants.

"If you use a coding agent, ask it to do something really difficult — to implement a feature, go read your code, try to understand your code, reason about your code, implement something, iterate — it might be successful," he defined. "And then come back the next day and ask it to implement the next feature, and it will do the same thing."

The problem, he argued, is that these programs don't internalize what they be taught. "In a sense, for the models we have today, every day is their first day of the job," Rafailov stated. "But an intelligent being should be able to internalize information. It should be able to adapt. It should be able to modify its behavior so every day it becomes better, every day it knows more, every day it works faster — the way a human you hire gets better at the job."

The duct tape drawback: How present coaching strategies educate AI to take shortcuts as an alternative of fixing issues

Rafailov pointed to a particular conduct in coding brokers that reveals the deeper drawback: their tendency to wrap unsure code in strive/besides blocks — a programming assemble that catches errors and permits a program to proceed operating.

"If you use coding agents, you might have observed a very annoying tendency of them to use try/except pass," he stated. "And in general, that is basically just like duct tape to save the entire program from a single error."

Why do brokers do that? "They do this because they understand that part of the code might not be right," Rafailov defined. "They understand there might be something wrong, that it might be risky. But under the limited constraint—they have a limited amount of time solving the problem, limited amount of interaction—they must only focus on their objective, which is implement this feature and solve this bug."

The end result: "They're kicking the can down the road."

This conduct stems from coaching programs that optimize for quick activity completion. "The only thing that matters to our current generation is solving the task," he stated. "And anything that's general, anything that's not related to just that one objective, is a waste of computation."

Why throwing extra compute at AI received't create superintelligence, in accordance with Pondering Machines researcher

Rafailov's most direct problem to the trade got here in his assertion that continued scaling received't be enough to succeed in AGI.

"I don't believe we're hitting any sort of saturation points," he clarified. "I think we're just at the beginning of the next paradigm—the scale of reinforcement learning, in which we move from teaching our models how to think, how to explore thinking space, into endowing them with the capability of general agents."

In different phrases, present approaches will produce more and more succesful programs that may work together with the world, browse the net, write code. "I believe a year or two from now, we'll look at our coding agents today, research agents or browsing agents, the way we look at summarization models or translation models from several years ago," he stated.

However common company, he argued, will not be the identical as common intelligence. "The much more interesting question is: Is that going to be AGI? And are we done — do we just need one more round of scaling, one more round of environments, one more round of RL, one more round of compute, and we're kind of done?"

His reply was unequivocal: "I don't believe this is the case. I believe that under our current paradigms, under any scale, we are not enough to deal with artificial general intelligence and artificial superintelligence. And I believe that under our current paradigms, our current models will lack one core capability, and that is learning."

Educating AI like college students, not calculators: The textbook strategy to machine studying

To elucidate the choice strategy, Rafailov turned to an analogy from arithmetic schooling.

"Think about how we train our current generation of reasoning models," he stated. "We take a particular math problem, make it very hard, and try to solve it, rewarding the model for solving it. And that's it. Once that experience is done, the model submits a solution. Anything it discovers—any abstractions it learned, any theorems—we discard, and then we ask it to solve a new problem, and it has to come up with the same abstractions all over again."

That strategy misunderstands how information accumulates. "This is not how science or mathematics works," he stated. "We build abstractions not necessarily because they solve our current problems, but because they're important. For example, we developed the field of topology to extend Euclidean geometry — not to solve a particular problem that Euclidean geometry couldn't handle, but because mathematicians and physicists understood these concepts were fundamentally important."

The answer: "Instead of giving our models a single problem, we might give them a textbook. Imagine a very advanced graduate-level textbook, and we ask our models to work through the first chapter, then the first exercise, the second exercise, the third, the fourth, then move to the second chapter, and so on—the way a real student might teach themselves a topic."

The target would basically change: "Instead of rewarding their success — how many problems they solved — we need to reward their progress, their ability to learn, and their ability to improve."

This strategy, generally known as "meta-learning" or "learning to learn," has precedents in earlier AI programs. "Just like the ideas of scaling test-time compute and search and test-time exploration played out in the domain of games first" — in programs like DeepMind's AlphaGo — "the same is true for meta learning. We know that these ideas do work at a small scale, but we need to adapt them to the scale and the capability of foundation models."

The lacking components for AI that really learns aren't new architectures—they're higher knowledge and smarter goals

When Rafailov addressed why present fashions lack this studying functionality, he supplied a surprisingly simple reply.

"Unfortunately, I think the answer is quite prosaic," he stated. "I think we just don't have the right data, and we don't have the right objectives. I fundamentally believe a lot of the core architectural engineering design is in place."

Reasonably than arguing for totally new mannequin architectures, Rafailov recommended the trail ahead lies in redesigning the information distributions and reward buildings used to coach fashions.

"Learning, in of itself, is an algorithm," he defined. "It has inputs — the current state of the model. It has data and compute. You process it through some sort of structure, choose your favorite optimization algorithm, and you produce, hopefully, a stronger model."

The query: "If reasoning models are able to learn general reasoning algorithms, general search algorithms, and agent models are able to learn general agency, can the next generation of AI learn a learning algorithm itself?"

His reply: "I strongly believe that the answer to this question is yes."

The technical strategy would contain creating coaching environments the place "learning, adaptation, exploration, and self-improvement, as well as generalization, are necessary for success."

"I believe that under enough computational resources and with broad enough coverage, general purpose learning algorithms can emerge from large scale training," Rafailov stated. "The way we train our models to reason in general over just math and code, and potentially act in general domains, we might be able to teach them how to learn efficiently across many different applications."

Neglect god-like reasoners: The primary superintelligence will probably be a grasp pupil

This imaginative and prescient results in a basically completely different conception of what synthetic superintelligence would possibly seem like.

"I believe that if this is possible, that's the final missing piece to achieve truly efficient general intelligence," Rafailov stated. "Now imagine such an intelligence with the core objective of exploring, learning, acquiring information, self-improving, equipped with general agency capability—the ability to understand and explore the external world, the ability to use computers, ability to do research, ability to manage and control robots."

Such a system would represent synthetic superintelligence. However not the type usually imagined in science fiction.

"I believe that intelligence is not going to be a single god model that's a god-level reasoner or a god-level mathematical problem solver," Rafailov stated. "I believe that the first superintelligence will be a superhuman learner, and it will be able to very efficiently figure out and adapt, propose its own theories, propose experiments, use the environment to verify that, get information, and iterate that process."

This imaginative and prescient stands in distinction to OpenAI's emphasis on constructing more and more highly effective reasoning programs, or Anthropic's give attention to "constitutional AI." As an alternative, Pondering Machines Lab seems to be betting that the trail to superintelligence runs by way of programs that may constantly enhance themselves by way of interplay with their atmosphere.

The $12 billion guess on studying over scaling faces formidable challenges

Rafailov's look comes at a posh second for Pondering Machines Lab. The corporate has assembled a formidable crew of roughly 30 researchers from OpenAI, Google, Meta, and different main labs. However it suffered a setback in early October when Andrew Tulloch, a co-founder and machine studying professional, departed to return to Meta after the corporate launched what The Wall Avenue Journal known as a "full-scale raid" on the startup, approaching greater than a dozen workers with compensation packages starting from $200 million to $1.5 billion over a number of years.

Regardless of these pressures, Rafailov's feedback counsel the corporate stays dedicated to its differentiated technical strategy. The corporate launched its first product, Tinker, an API for fine-tuning open-source language fashions, in October. However Rafailov's speak suggests Tinker is simply the inspiration for a way more bold analysis agenda targeted on meta-learning and self-improving programs.

"This is not easy. This is going to be very difficult," Rafailov acknowledged. "We'll need a lot of breakthroughs in memory and engineering and data and optimization, but I think it's fundamentally possible."

He concluded with a play on phrases: "The world is not enough, but we need the right experiences, and we need the right type of rewards for learning."

The query for Pondering Machines Lab — and the broader AI trade — is whether or not this imaginative and prescient might be realized, and on what timeline. Rafailov notably didn’t supply particular predictions about when such programs would possibly emerge.

In an trade the place executives routinely make daring predictions about AGI arriving inside years and even months, that restraint is notable. It suggests both uncommon scientific humility — or an acknowledgment that Pondering Machines Lab is pursuing a for much longer, tougher path than its rivals.

For now, essentially the most revealing element could also be what Rafailov didn't say throughout his TED AI presentation. No timeline for when superhuman learners would possibly emerge. No prediction about when the technical breakthroughs would arrive. Only a conviction that the potential was "fundamentally possible" — and that with out it, all of the scaling on this planet received't be sufficient.

You Might Also Like

Why AI coding brokers aren’t production-ready: Brittle context home windows, damaged refactors, lacking operational consciousness

AI denial is turning into an enterprise threat: Why dismissing “slop” obscures actual functionality positive factors

GAM takes purpose at “context rot”: A dual-agent reminiscence structure that outperforms long-context LLMs

The 'reality serum' for AI: OpenAI’s new technique for coaching fashions to admit their errors

Anthropic vs. OpenAI pink teaming strategies reveal completely different safety priorities for enterprise AI

TAGGED:039Firstchallengeslearner039MachinesOpenAI039sscalingstrategySuperhumanSuperintelligencethinking
Share This Article
Facebook Twitter Email Print

Follow US

Find US on Social Medias
FacebookLike
TwitterFollow
YoutubeSubscribe
TelegramFollow
Popular News
Vice President Vance says troops can be paid as stress builds on Congress to finish the shutdown
Politics

Vice President Vance says troops can be paid as stress builds on Congress to finish the shutdown

Editorial Board October 28, 2025
ESA unveils innovation-focused thought chief summit for April 2026
Doctors Debate Whether Trans Teens Need Therapy Before Hormones
In ‘Landscapers,’ True Crime Meets Hollywood Fantasy
Don Young of Alaska, the House Dean, Leaves an Earmark Legacy

You Might Also Like

Inside NetSuite’s subsequent act: Evan Goldberg on the way forward for AI-powered enterprise methods
Technology

Inside NetSuite’s subsequent act: Evan Goldberg on the way forward for AI-powered enterprise methods

December 4, 2025
Nvidia's new AI framework trains an 8B mannequin to handle instruments like a professional
Technology

Nvidia's new AI framework trains an 8B mannequin to handle instruments like a professional

December 4, 2025
Gong examine: Gross sales groups utilizing AI generate 77% extra income per rep
Technology

Gong examine: Gross sales groups utilizing AI generate 77% extra income per rep

December 4, 2025
AWS launches Kiro powers with Stripe, Figma, and Datadog integrations for AI-assisted coding
Technology

AWS launches Kiro powers with Stripe, Figma, and Datadog integrations for AI-assisted coding

December 4, 2025

Categories

  • Health
  • Sports
  • Politics
  • Entertainment
  • Technology
  • Art
  • World

About US

New York Dawn is a proud and integral publication of the Enspirers News Group, embodying the values of journalistic integrity and excellence.
Company
  • About Us
  • Newsroom Policies & Standards
  • Diversity & Inclusion
  • Careers
  • Media & Community Relations
  • Accessibility Statement
Contact Us
  • Contact Us
  • Contact Customer Care
  • Advertise
  • Licensing & Syndication
  • Request a Correction
  • Contact the Newsroom
  • Send a News Tip
  • Report a Vulnerability
Term of Use
  • Digital Products Terms of Sale
  • Terms of Service
  • Privacy Policy
  • Cookie Settings
  • Submissions & Discussion Policy
  • RSS Terms of Service
  • Ad Choices
© 2024 New York Dawn. All Rights Reserved.
Welcome Back!

Sign in to your account

Lost your password?