OpenAI’s new safety tools are designed to make AI models harder to jailbreak. Instead, they may give users a false sense of security

November 5, 2025

A small amount of bad data can ‘poison’ even the largest AI models, researchers warn

October 14, 2025

Researchers from top AI labs including Google, OpenAI, and Anthropic warn they may be losing the ability to understand advanced AI models