Find Related products on Amazon

Shop on Amazon

Researchers claim breakthrough in fight against AI’s frustrating security hole

Published on: 2025-04-26 11:15:44

In the AI world, a vulnerability called "prompt injection" has haunted developers since chatbots went mainstream in 2022. Despite numerous attempts to solve this fundamental vulnerability—the digital equivalent of whispering secret instructions to override a system's intended behavior—no one has found a reliable solution. Until now, perhaps. Google DeepMind has unveiled CaMeL (CApabilities for MachinE Learning), a new approach to stopping prompt-injection attacks that abandons the failed strategy of having AI models police themselves. Instead, CaMeL treats language models as fundamentally untrusted components within a secure software framework, creating clear boundaries between user commands and potentially malicious content. Prompt injection has created a significant barrier to building trustworthy AI assistants, which may be why general-purpose big tech AI like Apple's Siri doesn't currently work like ChatGPT. As AI agents get integrated into email, calendar, banking, and document- ... Read full article.