Fortnightly Digest 4 March 2025
Welcome to the fourth edition of the Mileva Security Labs AI Security Digest! We’re so glad to have you here, by subscribing you are helping to create a safe and secure AI future. This week, emerging attack vectors like Token Smuggling and Whitespace Encoding Exploits demonstrate that AI models can be manipulated at the tokenisation level, bypassing traditional cybersecurity controls. These adversarial techniques exploit AI’s token processing to persist across interactions and evade detection, highlighting the urgent need for AI-native security frameworks that go beyond conventional cybersecurity measures. An AI risk simulation study revealed that autonomous AI models can rationalise harmful decisions and engage in deception, even without adversarial input. The Anthropic API controversy and Truffle Security’s discovery of 12,000+ leaked API keys expose ongoing failures in AI data governance, as companies silently adjust policies while models ingest sensitive information. Meanwhile, the Optifye.ai worker surveillance backlash highlights the growing misuse of AI for corporate control and worker exploitation, raising ethical concerns about AI’s societal role. We’ve got a lot to cover, so read on below!