Armitage Archive

Highlight from The Security Paradox of Local LLMs

While the first attack plants a backdoor for later use, this one doesn’t wait for code to be deployed. It achieves immediate RCE on the developer’s machine during the code generation process.

The technique works by first distracting the model with a cognitive overload to bypass its safety filters. Once the model’s defenses are down, a second part of the prompt asks it to write a Python script containing an obfuscated payload.