OpenAI pledges to publish AI safety test results more often
Published on: 2025-07-10 08:38:40
OpenAI is moving to publish the results of its internal AI model safety evaluations more regularly in what the outfit is pitching as an effort to increase transparency.
On Wednesday, OpenAI launched the Safety evaluations hub, a web page showing how the company’s models score on various tests for harmful content generation, jailbreaks, and hallucinations. OpenAI says that it’ll use the hub to share metrics on an “ongoing basis,” and that it intends to update the hub with “major model updates” going forward.
Introducing the Safety Evaluations Hub—a resource to explore safety results for our models.
While system cards share safety metrics at launch, the Hub will be updated periodically as part of our efforts to communicate proactively about safety.https://t.co/c8NgmXlC2Y — OpenAI (@OpenAI) May 14, 2025
“As the science of AI evaluation evolves, we aim to share our progress on developing more scalable ways to measure model capability and safety,” wrote OpenAI in a blog post. “By shar
... Read full article.