AI Navigate

Beyond Creed: A Non-Identity Safety Condition A Strong Empirical Alternative to Identity Framing in Low-Data LoRA Fine-Tuning

arXiv cs.CL / 3/17/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper investigates four safety supervision formats for low-data LoRA finetuning across three instruction-tuned model families (Llama 3.1 8B, Qwen2.5 7B, Gemma 3 4B) and evaluates them on HarmBench.
  • The non-identity condition D yields 74.4% refusal on Llama, 76.9% on Gemma, and 74.1% on Qwen on the full 320-behavior HarmBench set.
  • Creed-style framing (B) improves over plain constitutional rules (A) for Llama and Gemma, but remains below D, giving an overall ordering: D > B > C ≥ A > baseline.
  • Capability evaluations on MMLU and ARC-Challenge show no meaningful trade-offs across the four conditions.

Abstract

How safety supervision is written may matter more than the explicit identity content it contains. We study low-data LoRA safety fine-tuning with four supervision formats built from the same core safety rules: constitutional rules (A), creed-style identity framing (B), a B-matched creed condition with a worldview/confession identity-maintenance tail (C), and a matched non-identity condition (D). Across three instruction-tuned model families (Llama 3.1 8B, Qwen2.5 7B, and Gemma 3 4B), we evaluate HarmBench using a reconciled dual-judge pipeline combining Bedrock-hosted DeepSeek v3.2 and Sonnet 4.6, with disagreement and boundary cases manually resolved. The non-identity condition D is the strongest group on all three model families on the full 320-behavior HarmBench set, reaching 74.4% refusal on Llama, 76.9% on Gemma, and 74.1% on Qwen. By comparison, creed-style framing (B) improves over plain constitutional rules (A) on Llama and Gemma, but remains substantially below D, yielding an overall descriptive ordering of D > B > C \geq A > baseline. This provides a bounded empirical challenge to a strong version of the identity-framing hypothesis: explicit creed-style identity language is not necessary for the strongest gains observed here. Capability evaluations on MMLU and ARC-Challenge show no meaningful trade-off across conditions.