=Coffee
16 February 2026

=Coffee

Hacked

About
A lot of modern AI models have a kind of security guard layer that sits in front of them. Its job? A binary choice as to whether the prompt heading into the model is safe or not. Kasimir Schulz, a lead security researcher at HiddenLayer, has been researching how to trick these models. Their solution, a technique called "Echogram" involves words with such positive statistical sentiment — such overwhelming good vibes — that it flips that verdict.


Learn more about your ad choices. Visit podcastchoices.com/adchoices