We collect cookies to analyze our website traffic and performance; we never collect any personal data. Cookie Policy
Accept
NEW YORK DAWN™NEW YORK DAWN™NEW YORK DAWN™
Notification Show More
Font ResizerAa
  • Home
  • Trending
  • New York
  • World
  • Politics
  • Business
    • Business
    • Economy
    • Real Estate
  • Crypto & NFTs
  • Tech
  • Lifestyle
    • Lifestyle
    • Food
    • Travel
    • Fashion
    • Art
  • Health
  • Sports
  • Entertainment
Reading: Anthropic faces backlash to Claude 4 Opus characteristic that contacts authorities, press if it thinks you’re doing one thing ‘egregiously immoral’
Share
Font ResizerAa
NEW YORK DAWN™NEW YORK DAWN™
Search
  • Home
  • Trending
  • New York
  • World
  • Politics
  • Business
    • Business
    • Economy
    • Real Estate
  • Crypto & NFTs
  • Tech
  • Lifestyle
    • Lifestyle
    • Food
    • Travel
    • Fashion
    • Art
  • Health
  • Sports
  • Entertainment
Follow US
NEW YORK DAWN™ > Blog > Technology > Anthropic faces backlash to Claude 4 Opus characteristic that contacts authorities, press if it thinks you’re doing one thing ‘egregiously immoral’
Anthropic faces backlash to Claude 4 Opus characteristic that contacts authorities, press if it thinks you’re doing one thing ‘egregiously immoral’
Technology

Anthropic faces backlash to Claude 4 Opus characteristic that contacts authorities, press if it thinks you’re doing one thing ‘egregiously immoral’

Last updated: May 22, 2025 9:26 pm
Editorial Board Published May 22, 2025
Share
SHARE

Anthropic’s first developer convention on Might 22 ought to have been a proud and joyous day for the agency, but it surely has already been hit with a number of controversies, together with Time journal leaking its marquee announcement forward of…nicely, time (no pun supposed), and now, a significant backlash amongst AI builders and energy customers brewing on X over a reported security alignment characteristic in Anthropic’s flagship new Claude 4 Opus giant language mannequin.

Name it the “ratting” characteristic, as it’s designed to rat a consumer out to authorities if the mannequin detects the consumer engaged in wrongdoing.

“If it thinks you’re doing one thing egregiously immoral, for instance, like faking knowledge in a pharmaceutical trial, it can use command-line instruments to contact the press, contact regulators, attempt to lock you out of the related methods, or the entire above.“

The “it” was in reference to the brand new Claude 4 Opus mannequin, which Anthropic has already overtly warned may assist novices create bioweapons in sure circumstances, and tried to forestall simulated substitute by blackmailing human engineers throughout the firm.

Apparently, in an try and cease Claude 4 Opus from partaking in these sort of damaging and nefarious behaviors, researchers on the AI firm added quite a few new security options, together with one that might, in line with Bowman, contact outsiders if it was directed by the consumer to have interaction in “something egregiously immoral.”

Quite a few questions for particular person customers and enterprises about what Claude 4 Opus will do to your knowledge, and beneath what circumstances

Whereas maybe well-intended, the characteristic raises all kinds of questions for Claude 4 Opus customers, together with enterprises and enterprise clients — chief amongst them, what behaviors will the mannequin contemplate “egregiously immoral” and act upon? Will it share personal enterprise or consumer knowledge with authorities autonomously (by itself), with out the consumer’s permission?

The implications are profound and could possibly be detrimental to customers, and maybe unsurprisingly, Anthropic confronted an instantaneous and nonetheless ongoing torrent of criticism from AI energy customers and rival builders.

Austin Allred, co-founder of the federal government fined coding camp BloomTech and now a co-founder of Gauntlet AI, put his emotions in all caps: “Honest question for the Anthropic team: HAVE YOU LOST YOUR MINDS?”

Ben Hyak, a former SpaceX and Apple designer and present co-founder of Raindrop AI, an AI observability and monitoring startup, additionally took to X to blast Anthropic’s acknowledged coverage and have: “this is, actually, just straight up illegal,” including in one other publish: “An AI Alignment researcher at Anthropic simply stated that Claude Opus will CALL THE POLICE or LOCK YOU OUT OF YOUR COMPUTER if it detects you doing one thing unlawful?? i’ll by no means give this mannequin entry to my laptop.“

“Some of the statements from Claude’s safety people are absolutely crazy,” wrote pure language processing (NLP) Casper Hansen on X. “Makes you root a bit more for [Anthropic rival] OpenAI seeing the level of stupidity being this publicly displayed.”

Anthropic researcher adjustments tune

Bowman later edited his tweet and the next one in a thread to learn as follows, but it surely nonetheless didn’t persuade the naysayers that their consumer knowledge and security can be protected against intrusive eyes:

Bowman added:

“I deleted the sooner tweet on whistleblowing because it was being pulled out of context.

TBC: This isn’t a brand new Claude characteristic and it’s not attainable in regular utilization. It exhibits up in testing environments the place we give it unusually free entry to instruments and really uncommon directions.“

Screenshot 2025 05 22 at 3.13.04%E2%80%AFPM

From its inception, Anthropic has greater than different AI labs sought to place itself as a bulwark of AI security and ethics, centering its preliminary work on the ideas of “Constitutional AI,” or AI that behaves in line with a set of requirements useful to humanity and customers. Nevertheless, with this new replace, the moralizing could have brought about the decidedly reverse response amongst customers — making them mistrust the brand new mannequin and your entire firm, and thereby turning them away from it.

I’ve reached out to an Anthropic spokesperson with extra questions on this characteristic and can replace after I hear again.

Each day insights on enterprise use circumstances with VB Each day

If you wish to impress your boss, VB Each day has you coated. We provide the inside scoop on what firms are doing with generative AI, from regulatory shifts to sensible deployments, so you possibly can share insights for optimum ROI.

An error occured.

Former Anthropic exec raises M to insure AI brokers and assist startups deploy safely

You Might Also Like

It’s Qwen’s summer season: new open supply Qwen3-235B-A22B-Pondering-2507 tops OpenAI, Gemini reasoning fashions on key benchmarks

Freed says 20,000 clinicians are utilizing its medical AI transcription ‘scribe,’ however competitors is rising quick

Anthropic unveils ‘auditing agents’ to check for AI misalignment

SecurityPal combines AI and consultants in Nepal to hurry enterprise safety questionnaires by 87X or extra

White Home plan alerts “open-weight first” period—and enterprises want new guardrails

TAGGED:AnthropicauthoritiesbacklashClaudecontactsegregiouslyfacesfeatureimmoralOpuspressthinksYoure
Share This Article
Facebook Twitter Email Print

Follow US

Find US on Social Medias
FacebookLike
TwitterFollow
YoutubeSubscribe
TelegramFollow
Popular News
‘American Song Contest’ Puts a Stateside Spin on Eurovision
Entertainment

‘American Song Contest’ Puts a Stateside Spin on Eurovision

Editorial Board March 18, 2022
U.S. Will Help Transfer Soviet-Made Tanks to Ukraine
Mets evaluating depth starters after shedding Frankie Montas and Sean Manaea
Analysis: Biden Sees Booming Economy, but Many Americans Don’t
Manhattan’s Sean O’Bradaigh makes professional debut in all-Irish boxing card at The Backyard

You Might Also Like

Qwen3-Coder-480B-A35B-Instruct launches and it ‘might be the best coding model yet’
Technology

Qwen3-Coder-480B-A35B-Instruct launches and it ‘might be the best coding model yet’

July 23, 2025
Former Anthropic exec raises M to insure AI brokers and assist startups deploy safely
Technology

Former Anthropic exec raises $15M to insure AI brokers and assist startups deploy safely

July 23, 2025
Alibaba’s new open supply Qwen3-235B-A22B-2507 beats Kimi-2 and affords low compute model
Technology

Alibaba’s new open supply Qwen3-235B-A22B-2507 beats Kimi-2 and affords low compute model

July 23, 2025
Combination-of-recursions delivers 2x sooner inference—Right here’s how one can implement it
Technology

Combination-of-recursions delivers 2x sooner inference—Right here’s how one can implement it

July 23, 2025

Categories

  • Health
  • Sports
  • Politics
  • Entertainment
  • Technology
  • World
  • Art

About US

New York Dawn is a proud and integral publication of the Enspirers News Group, embodying the values of journalistic integrity and excellence.
Company
  • About Us
  • Newsroom Policies & Standards
  • Diversity & Inclusion
  • Careers
  • Media & Community Relations
  • Accessibility Statement
Contact Us
  • Contact Us
  • Contact Customer Care
  • Advertise
  • Licensing & Syndication
  • Request a Correction
  • Contact the Newsroom
  • Send a News Tip
  • Report a Vulnerability
Term of Use
  • Digital Products Terms of Sale
  • Terms of Service
  • Privacy Policy
  • Cookie Settings
  • Submissions & Discussion Policy
  • RSS Terms of Service
  • Ad Choices
© 2024 New York Dawn. All Rights Reserved.
Welcome Back!

Sign in to your account

Lost your password?