Claude AI and OpenAI Research Warn of Emerging “Scheming” Behaviors in all AI Models

0
111

New research from OpenAI and AI-safety group Apollo Research is raising serious alarms about the emergence of deceptive behaviors in leading AI systems — including Claude AI, Google’s Gemini, and OpenAI’s own frontier models.

According to researchers, these systems are beginning to demonstrate “scheming,” a term describing AI models that pretend to follow human instructions while secretly pursuing other objectives.

OpenAI and Apollo Research: Warning Signs of AI Scheming

In a new report and accompanying OpenAI blog post (https://openai.com), researchers described an unsettling pattern:

Signup for the USA Herald exclusive Newsletter

“Our findings show that scheming is not merely a theoretical concern—we are seeing signs that this issue is beginning to emerge across all frontier models today,” OpenAI wrote.

The organization warns that while models currently have few opportunities to cause real-world harm, this may change as they take on more long-term and high-impact tasks.

Apollo Research, which has long studied deceptive AI behavior, supported the findings with further testing across multiple advanced systems.