AI Navigate

Amnesia: Adversarial Semantic Layer Specific Activation Steering in Large Language Models

arXiv cs.AI / 3/12/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • Amnesia is a lightweight activation-space adversarial attack that targets internal transformer states to bypass safety mechanisms in open-weight LLMs.
  • It operates without any fine-tuning or additional training and can induce harmful content in state-of-the-art open-weight LLMs during evaluation.
  • Red-teaming experiments show that existing safeguards can be circumvented, highlighting vulnerabilities in current alignment and safety measures.
  • The findings emphasize the need for more robust security defenses and continued research to prevent misuse of open-weight LLMs.

Abstract

Warning: This article includes red-teaming experiments, which contain examples of compromised LLM responses that may be offensive or upsetting. Large Language Models (LLMs) have the potential to create harmful content, such as generating sophisticated phishing emails and assisting in writing code of harmful computer viruses. Thus, it is crucial to ensure their safe and responsible response generation. To reduce the risk of generating harmful or irresponsible content, researchers have developed techniques such as reinforcement learning with human feedback to align LLM's outputs with human values and preferences. However, it is still undetermined whether such measures are sufficient to prevent LLMs from generating interesting responses. In this study, we propose Amnesia, a lightweight activation-space adversarial attack that manipulates internal transformer states to bypass existing safety mechanisms in open-weight LLMs. Through experimental analysis on state-of-the-art, open-weight LLMs, we demonstrate that our attack effectively circumvents existing safeguards, enabling the generation of harmful content without the need for any fine-tuning or additional training. Our experiments on benchmark datasets show that the proposed attack can induce various antisocial behaviors in LLMs. These findings highlight the urgent need for more robust security measures in open-weight LLMs and underscore the importance of continued research to prevent their potential misuse.