We collect cookies to analyze our website traffic and performance; we never collect any personal data. Cookie Policy
Accept
NEW YORK DAWN™NEW YORK DAWN™NEW YORK DAWN™
Notification Show More
Font ResizerAa
  • Home
  • Trending
  • New York
  • World
  • Politics
  • Business
    • Business
    • Economy
    • Real Estate
  • Crypto & NFTs
  • Tech
  • Lifestyle
    • Lifestyle
    • Food
    • Travel
    • Fashion
    • Art
  • Health
  • Sports
  • Entertainment
Reading: OpenAI overrode considerations of knowledgeable testers to launch sycophantic GPT-4o
Share
Font ResizerAa
NEW YORK DAWN™NEW YORK DAWN™
Search
  • Home
  • Trending
  • New York
  • World
  • Politics
  • Business
    • Business
    • Economy
    • Real Estate
  • Crypto & NFTs
  • Tech
  • Lifestyle
    • Lifestyle
    • Food
    • Travel
    • Fashion
    • Art
  • Health
  • Sports
  • Entertainment
Follow US
NEW YORK DAWN™ > Blog > Technology > OpenAI overrode considerations of knowledgeable testers to launch sycophantic GPT-4o
OpenAI overrode considerations of knowledgeable testers to launch sycophantic GPT-4o
Technology

OpenAI overrode considerations of knowledgeable testers to launch sycophantic GPT-4o

Last updated: May 2, 2025 8:16 pm
Editorial Board Published May 2, 2025
Share
SHARE

It’s been a little bit of a topsy-turvy week for the primary generative AI firm by way of customers.

OpenAI, creator of ChatGPT, launched after which withdrew an up to date model of the underlying multimodal (textual content, picture, audio) giant language mannequin (LLM) that ChatGPT is attached to by default, GPT-4o, as a consequence of it being too sycophantic to customers. The corporate not too long ago reported no less than 500 million lively weekly customers of the hit internet service.

A fast primer on the horrible, no good, sycophantic GPT-4o replace

OpenAI started updating GPT-4o to a more moderen mannequin it hoped can be extra well-received by customers on April twenty fourth, accomplished the up to date by April twenty fifth, then, 5 days later, rolled it again on April 29, after days of mounting complaints of customers throughout social media — primarily on X and Reddit.

The complaints various in depth and in specifics, however all usually coalesced round the truth that GPT-4o seemed to be responding to person queries with undue flattery, help for misguided, incorrect and downright dangerous concepts, and “glazing” or praising the person to an extreme diploma when it wasn’t really particularly requested, a lot much less warranted.

In examples screenshotted and posted by customers, ChatGPT powered by that sycophantic, up to date GPT-4o mannequin had praised and endorsed a enterprise thought for literal “shit on a stick,” applauded a person’s pattern textual content of schizophrenic delusional isolation, and even allegedly supported plans to commit terrorism.

Customers together with prime AI researchers and even a former OpenAI interim CEO mentioned they had been involved that an AI mannequin’s unabashed cheerleading for all these horrible person prompts was greater than merely annoying or inappropriate — that it may trigger precise hurt to customers who mistakenly believed the AI and felt emboldened by its help for his or her worst concepts and impulses. It rose to the extent of an AI security difficulty.

OpenAI then launched a weblog publish describing what went improper — “we focused too much on short-term feedback, and did not fully account for how users’ interactions with ChatGPT evolve over time. As a result, GPT‑4o skewed towards responses that were overly supportive but disingenuous” — and the steps the corporate was taking to deal with the problems. OpenAI’s Head of Mannequin Habits Joanne Jang additionally participated in a Reddit “Ask me anything” or AMA discussion board answering textual content posts from customers and revealed additional details about the corporate’s method to GPT-4o and the way it ended up with an excessively sycophantic mannequin, together with not “bak[ing] in enough nuance,” as to the way it was incorporating person suggestions similar to “thumbs up” actions made by customers in response to mannequin outputs they favored.

Now at this time, OpenAI has launched a weblog publish with much more details about how the sycophantic GPT-4o replace occurred — credited to not any explicit writer, however to “OpenAI.”

CEO and co-founder Sam Altman additionally posted a hyperlink to the weblog publish on X, saying: “we missed the mark with last week’s GPT-4o update. what happened, what we learned, and some things we will do differently in the future.”

What the brand new OpenAI weblog publish reveals about how and why GPT-4o turned so sycophantic

To me, a every day person of ChatGPT together with the 4o mannequin, probably the most placing admission from OpenAI’s new weblog publish concerning the sycophancy replace is how the corporate seems to disclose that it did obtain considerations concerning the mannequin previous to launch from a small group of “expert testers,” however that it seemingly overrode these in favor of a broader enthusiastic response from a wider group of extra normal customers.

As the corporate writes (emphasis mine):

“While we’ve had discussions about risks related to sycophancy in GPT‑4o for a while, sycophancy wasn’t explicitly flagged as part of our internal hands-on testing, as some of our expert testers were more concerned about the change in the model’s tone and style. Nevertheless, some expert testers had indicated that the model behavior “felt” barely off…

“We then had a choice to make: ought to we withhold deploying this replace regardless of constructive evaluations and A/B check outcomes, primarily based solely on the subjective flags of the knowledgeable testers? In the long run, we determined to launch the mannequin because of the constructive indicators from the customers who tried out the mannequin.

“Unfortunately, this was the wrong call. We build these models for our users and while user feedback is critical to our decisions, it’s ultimately our responsibility to interpret that feedback correctly.”

This appears to me like a giant mistake. Why even have knowledgeable testers in the event you’re not going to weight their experience greater than the lots of the gang? I requested Altman about this selection on X however he has but to reply.

Not all ‘reward signals’ are equal

OpenAI’s new autopsy weblog publish additionally reveals extra specifics about how the corporate trains and updates new variations of present fashions, and the way human suggestions alters the mannequin qualities, character, and “personality.” As the corporate writes:

“Since launching GPT‑4o in ChatGPT final Might, we’ve launched 5 main updates centered on modifications to persona and helpfulness. Every replace includes new post-training, and sometimes many minor changes to the mannequin coaching course of are independently examined after which mixed right into a single up to date mannequin which is then evaluated for launch.

“To post-train fashions, we take a pre-trained base mannequin, do supervised fine-tuning on a broad set of excellent responses written by people or present fashions, after which run reinforcement studying with reward indicators from a wide range of sources.

“Throughout reinforcement studying, we current the language mannequin with a immediate and ask it to put in writing responses. We then fee its response based on the reward indicators, and replace the language mannequin to make it extra prone to produce higher-rated responses and fewer prone to produce lower-rated responses.“

Clearly, the “reward signals” utilized by OpenAI throughout post-training have an infinite influence on the ensuing mannequin habits, and because the firm admitted earlier when it overweighted “thumbs up” responses from ChatGPT customers to its outputs, this sign will not be one of the best one to make use of equally with others when figuring out how the mannequin learns to speak and what sorts of responses it must be serving up. OpenAI admits this outright within the subsequent paragraph of its publish, writing:

“Defining the correct set of reward signals is a difficult question, and we take many things into account: are the answers correct, are they helpful, are they in line with our Model Spec⁠, are they safe, do users like them, and so on. Having better and more comprehensive reward signals produces better models for ChatGPT, so we’re always experimenting with new signals, but each one has its quirks.”

Certainly, OpenAI additionally reveals the “thumbs up” reward sign was a brand new one used alongside different reward indicators on this explicit replace.

“the update introduced an additional reward signal based on user feedback—thumbs-up and thumbs-down data from ChatGPT. This signal is often useful; a thumbs-down usually means something went wrong.”

But critically, the corporate doesn’t blame the brand new “thumbs up” information outright for the mannequin’s failure and ostentatious cheerleading behaviors. As an alternative, OpenAI’s weblog publish says it was this mixed with a wide range of different new and older reward indicators, led to the issues: “…we had candidate improvements to better incorporate user feedback, memory, and fresher data, among others. Our early assessment is that each of these changes, which had looked beneficial individually, may have played a part in tipping the scales on sycophancy when combined.”

Reacting to this weblog publish, Andrew Mayne, a former member of the OpenAI technical employees now working at AI consulting agency Interdimensional, wrote on X of one other instance of how delicate modifications in reward incentives and mannequin pointers can influence mannequin efficiency fairly dramatically:

“Early on at OpenAI, I had a disagreement with a colleague (who is now a founder of another lab) over using the word “polite” in a immediate instance I wrote.

They argued “polite” was politically incorrect and wished to swap it for “helpful.”

I identified that focusing solely on helpfulness could make a mannequin overly compliant—so compliant, in reality, that it may be steered into sexual content material inside just a few turns.

After I demonstrated that danger with a easy change, the immediate stored “polite.”

These fashions are bizarre.“

How OpenAI plans to enhance its mannequin testing processes going ahead

The corporate lists six course of enhancements for the best way to keep away from related undesirable and less-than-ideal mannequin habits sooner or later, however to me an important is that this:

“We’ll adjust our safety review process to formally consider behavior issues—such as hallucination, deception, reliability, and personality—as blocking concerns. Even if these issues aren’t perfectly quantifiable today, we commit to blocking launches based on proxy measurements or qualitative signals, even when metrics like A/B testing look good.”

In different phrases — regardless of how vital information, particularly quantitative information, is to the fields of machine studying and synthetic intelligence — OpenAI acknowledges that this alone can’t and shouldn’t be the one means by which a mannequin’s efficiency is judged.

Whereas many customers offering a “thumbs up” may sign a sort of fascinating habits within the brief time period, the long run implications for a way the AI mannequin responds and the place these behaviors take it and its customers, may finally result in a really darkish, distressing, damaging, and undesirable place. Extra is just not all the time higher — particularly if you end up constraining the “more” to a couple domains of indicators.

It’s not sufficient to say that the mannequin handed the entire assessments or acquired various constructive responses from customers — the experience of skilled energy customers and their qualitative suggestions that one thing “seemed off” concerning the mannequin, even when they couldn’t absolutely categorical why, ought to carry way more weight than OpenAI was allocating beforehand.

Let’s hope the corporate — and all the subject — learns from this incident and integrates the teachings going ahead.

Broader takeaways and issues for enterprise decision-makers

Talking maybe extra theoretically, for myself, it additionally signifies why experience is so vital — and particularly, experience in fields past and outdoors of the one you’re optimizing for (on this case, machine studying and AI). It’s the range of experience that permits us as a species to attain new advances that profit our variety. One, say STEM, shouldn’t essentially be held above the others within the humanities or arts.

AI mannequin designers and technical decision-makers at enterprises would do effectively to maintain this broader thought in thoughts when designing metrics round any measurable objective — as a result of even whenever you assume you’re utilizing information to your benefit, it may backfire in methods you didn’t absolutely count on or anticipate, leaving your scrambling to restore the injury and mop up the mess you made, nonetheless inadvertently.

Each day insights on enterprise use instances with VB Each day

If you wish to impress your boss, VB Each day has you lined. We provide the inside scoop on what firms are doing with generative AI, from regulatory shifts to sensible deployments, so you’ll be able to share insights for max ROI.

An error occured.

You Might Also Like

Google claims Gemini 2.5 Professional preview beats DeepSeek R1 and Grok 3 Beta in coding efficiency

Solidroad simply raised $6.5M to reinvent customer support with AI that coaches, not replaces

Google Play launches Diamond District expertise in Roblox

Databricks and Noma sort out CISOs’ AI nightmares round inference vulnerabilities

How a lot data do LLMs actually memorize? Now we all know, because of Meta, Google, Nvidia and Cornell

TAGGED:concernsexpertGPT4oOpenAIoverrodereleasesycophantictesters
Share This Article
Facebook Twitter Email Print

Follow US

Find US on Social Medias
FacebookLike
TwitterFollow
YoutubeSubscribe
TelegramFollow
Popular News
For Financial Help and Counsel, Hunter Biden Turns to Hollywood Lawyer
Politics

For Financial Help and Counsel, Hunter Biden Turns to Hollywood Lawyer

Editorial Board May 10, 2022
More Than a Million Afghans Flee as Economy Collapses
Aaron Rodgers says regardless of Jets’ disappointing season, he desires to return in 2025
Joan Didion Chronicled American Disorder With Her Own Unmistakable Style
The Knicks misplaced Sport 2 to the Pacers as a result of they couldn’t cease one play

You Might Also Like

Play Community wins a number of authorized circumstances in token dispute with Prepared Makers Inc.
Technology

Play Community wins a number of authorized circumstances in token dispute with Prepared Makers Inc.

June 5, 2025
Latent Know-how raises M to alter animation with generative physics
Technology

Latent Know-how raises $8M to alter animation with generative physics

June 5, 2025
Nintendo brings again late-night console launches with debut of Swap 2
Technology

Nintendo brings again late-night console launches with debut of Swap 2

June 5, 2025
Nintendo Change 2 will get official gaming equipment from Belkin
Technology

Nintendo Change 2 will get official gaming equipment from Belkin

June 5, 2025

Categories

  • Health
  • Sports
  • Politics
  • Entertainment
  • Technology
  • World
  • Art

About US

New York Dawn is a proud and integral publication of the Enspirers News Group, embodying the values of journalistic integrity and excellence.
Company
  • About Us
  • Newsroom Policies & Standards
  • Diversity & Inclusion
  • Careers
  • Media & Community Relations
  • Accessibility Statement
Contact Us
  • Contact Us
  • Contact Customer Care
  • Advertise
  • Licensing & Syndication
  • Request a Correction
  • Contact the Newsroom
  • Send a News Tip
  • Report a Vulnerability
Term of Use
  • Digital Products Terms of Sale
  • Terms of Service
  • Privacy Policy
  • Cookie Settings
  • Submissions & Discussion Policy
  • RSS Terms of Service
  • Ad Choices
© 2024 New York Dawn. All Rights Reserved.
Welcome Back!

Sign in to your account

Lost your password?