Close Menu
MMJ News NetworkMMJ News Network
  • Home
  • Cannabis
  • Psychedelics
  • Crypto & Web3
  • AI
  • CBD
  • Wellness & Counterculture
  • MMJNEWS

Subscribe to Updates

Subscribe to our newsletter and never miss our latest news

Subscribe my Newsletter for New Posts & tips Let's stay updated!

What's Hot

CEO of ‘Textbook Ponzi’ Pleads Guilty in $200M Bitcoin Fraud Case

September 18, 2025

BNB Price Rally Nears $1,000 – Can Bulls Push Even Higher?

September 18, 2025

Ethereum Close To Local Bottom? Analyst Flags Drop In Binance Open Interest

September 18, 2025
Facebook X (Twitter) Instagram
MMJ News NetworkMMJ News Network
  • Home
  • About Us
  • Advertise With Us
  • Contact Us
  • DMCA
  • Privacy Policy
  • Terms & Conditions
  • Home
  • Cannabis
  • Psychedelics
  • Crypto & Web3
  • AI
  • CBD
  • Wellness & Counterculture
  • MMJNEWS
MMJ News NetworkMMJ News Network
Home » OpenAI co-founder calls for AI labs to safety test rival models
Tech

OpenAI co-founder calls for AI labs to safety test rival models

EditorBy EditorAugust 27, 2025No Comments4 Mins Read
Share Facebook Twitter Pinterest LinkedIn Tumblr Email
Share
Facebook Twitter LinkedIn Pinterest Email Copy Link


OpenAI and Anthropic, two of the world’s leading AI labs, briefly opened up their closely guarded AI models to allow for joint safety testing — a rare cross-lab collaboration at a time of fierce competition. The effort aimed to surface blind spots in each company’s internal evaluations, and demonstrate how leading AI companies can work together on safety and alignment work in the future.

In an interview with TechCrunch, OpenAI co-founder Wojciech Zaremba said this kind of collaboration is increasingly important now that AI is entering a “consequential” stage of development, where AI models are used by millions of people everyday.

“There’s a broader question of how the industry sets a standard for safety and collaboration, despite the billions of dollars invested, as well as the war for talent, users, and the best products,” said Zaremba.

The joint safety research, published Wednesday by both companies, arrives amid an arms race among leading AI labs like OpenAI and Anthropic, where billion-dollar data center bets and $100 million compensation packages for top researchers have become table stakes. Some experts warn that the intensity of product competition could pressure companies to cut corners on safety in the rush to build more powerful systems.

To make this research possible, OpenAI and Anthropic granted each other special API access to versions of their AI models with fewer safeguards (OpenAI notes that GPT-5 was not tested because it hadn’t been released yet). Shortly after the research was conducted, however, Anthropic revoked another team at OpenAI’s API access. At the time, Anthropic claimed that OpenAI violated its terms of service, which prohibits using Claude to improve competing products.

Zaremba says the events were unrelated, and that he expects competition to stay fierce even as AI safety teams try to work together. Nicholas Carlini, a safety researcher with Anthropic, tells TechCrunch that he would like to continue allowing OpenAI safety researchers to access Claude models in the future.

“We want to increase collaboration wherever it’s possible across the safety frontier, and try to make this something that happens more regularly,” said Carlini.

Techcrunch event

San Francisco
|
October 27-29, 2025

One of the most stark findings in the study relates to hallucination testing. Anthropic’s Claude Opus 4 and Sonnet 4 models refused to answer up to 70% of questions when they were unsure of the correct answer, instead offering responses like, “I don’t have reliable information.” Meanwhile, OpenAI’s o3 and o4-mini models refuse to answer questions far less, but showed much higher hallucination rates, attempting to answer questions when they didn’t have enough information.

Zaremba says the right balance is likely somewhere in the middle — OpenAI’s models should refuse to answer more questions, while Anthropic’s models should probably attempt to offer more answers.

Sycophancy, the tendency for AI models to reinforce negative behavior in users to please them, has emerged as one of the most pressing safety concerns around AI models. While this topic wasn’t directly studied in the joint research, it’s an area both OpenAI and Anthropic are investing considerable resources into studying.

On Tuesday, parents of a 16-year-old boy, Adam Raine, filed a lawsuit against OpenAI, claiming that ChatGPT offered their son advice that aided in his suicide, rather than pushing back on his suicidal thoughts. The lawsuit suggests this may be the latest example of AI chatbot sycophancy contributing to tragic outcomes.

“It’s hard to imagine how difficult this is to their family,” said Zaremba when asked about the incident. “It would be a sad story if we build AI that solves all these complex PhD level problems, invents new science, and at the same time, we have people with mental health problems as a consequence of interacting with it. This is a dystopian future that I’m not excited about.”

In a blog post, OpenAI says that it significantly improved the sycophancy of its AI chatbots with GPT-5, compared to GPT-4o, significantly improving the model’s ability to respond to mental health emergencies.

Moving forward, Zaremba and Carlini say they would like Anthropic and OpenAI to collaborate more on safety testing, looking into more subjects and testing future models, and they hope other AI labs will follow their collaborative approach.



Source link

OpenAI safety testing
Share. Facebook Twitter Pinterest LinkedIn Tumblr Email
Editor
  • Website
  • Facebook
  • Instagram

Related Posts

India leads the way on Google’s Nano Banana with a local creative twist

September 18, 2025

Meta unveils new smart glasses with a display and wristband controller

September 18, 2025

Irregular raises $80 million to secure frontier AI models

September 17, 2025

Comments are closed.

Don't Miss
Crypto & Web3

CEO of ‘Textbook Ponzi’ Pleads Guilty in $200M Bitcoin Fraud Case

In brief Prosecutors said Palafox raised $201M by promising daily returns of up to 3%.…...

Free Membership Required

You must be a Free member to access this content.

Join Now

Already a member? Log in here

BNB Price Rally Nears $1,000 – Can Bulls Push Even Higher?

September 18, 2025

Ethereum Close To Local Bottom? Analyst Flags Drop In Binance Open Interest

September 18, 2025

xStocks Issuer Chose Switzerland to Avoid Whitelisting Tokenized Tesla Shares: CEO

September 18, 2025
Top Posts

8 US Senators Tell Party Leaders to Regulate Hemp, Not Ban It

September 17, 2025

Minnesota Dispensaries Launch Adult-Use Sales Under State Program

September 16, 2025

Regulators Ready to Enforce Cannabis Laws on Hemp THC Retailers in Maryland

September 16, 2025

California Passes Bill to Ban Intoxicating Hemp Products Outside Cannabis Market

September 15, 2025

Subscribe to Updates

Subscribe to our newsletter and never miss our latest news

Subscribe my Newsletter for New Posts & tips Let's stay updated!

About Us
About Us

Welcome to MMJ News Network, your premier source for cutting-edge insights into cannabis, psychedelics, crypto & Web3, wellness, counterculture, and market trends. We are dedicated to bringing you the latest news, research, and developments shaping these fast-evolving industries.

Facebook X (Twitter) Pinterest YouTube WhatsApp
Our Picks

CEO of ‘Textbook Ponzi’ Pleads Guilty in $200M Bitcoin Fraud Case

September 18, 2025

BNB Price Rally Nears $1,000 – Can Bulls Push Even Higher?

September 18, 2025

Ethereum Close To Local Bottom? Analyst Flags Drop In Binance Open Interest

September 18, 2025
Most Popular

Ethereum Falls as Crypto Exchange Bybit Confirms $1.4 Billion Hack

February 21, 2025

Florida Woman Accused of $850K Trump Solana Meme Coin Theft, Faces Deportation

February 21, 2025

Bitcoin, XRP and Dogecoin Sink Amid Inflation Fears and Bybit Hack Fallout

February 23, 2025
  • Home
  • About Us
  • Advertise With Us
  • Contact Us
  • DMCA
  • Privacy Policy
  • Terms & Conditions
© 2025 mmjnewsnetwork. Designed by mmjnewsnetwork.

Type above and press Enter to search. Press Esc to cancel.