Hackers discovered a sneaky way to steal data by hiding malicious prompts inside images processed by large language models


  • Malicious prompts remain invisible until image downscaling reveals hidden instructions
  • The attack works by exploiting how AI resamples uploaded images
  • Bicubic interpolation can expose black text from specially crafted images

As AI tools become more integrated into daily work, the security risks attached to them are also evolving in new directions.

Researchers at Trail of Bits have demonstrated a method where malicious prompts are hidden inside images and then revealed during processing by large language models.

Continue Reading