Noise Steering for Controlled Text Generation: Improving Diversity and Reading-Level Fidelity in Arabic Educational Story Generation

arXiv cs.CL / 4/7/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper studies “noise steering” for Arabic educational story generation, aiming to increase narrative diversity while maintaining strict constraints on vocabulary, reading level, and narrative structure.
  • It proposes injecting calibrated Gaussian perturbations into transformer internal representations at inference time as a training-free method, and evaluates four injection strategies across five small Arabic-centric language models (7–9B).
  • Residual-stream noise improves story diversity while incurring minimal loss in overall quality or constraint adherence and maintains early-grade reading level across all tested models.
  • Attention-entropy noise injection (AENI) is reported to stabilize attention-logit noise injection and recover quality, outperforming less stable approaches.
  • In contrast, high-temperature sampling increases reading grade level and can lead to “catastrophic collapse” in several models, suggesting internal perturbations are better than output-level randomness for constrained educational generation.

Abstract

Generating diverse, pedagogically valid stories for Arabic early-grade reading assessments requires balancing tight constraints on vocabulary, reading level, and narrative structure against the need to avoid repetitive plots that undermine assessment validity. We investigate noise steering, injecting calibrated Gaussian perturbations into the internal representations of transformer models at inference time, as a training-free diversity method evaluated across five small Arabic-centric language models (7-9B parameters). We compare four injection strategies against high-temperature sampling baselines, measuring diversity, quality, constraint adherence, and reading grade level. Residual stream noise consistently improves narrative diversity with minimal quality or constraint cost and preserves early-grade reading level across all models. Attention entropy noise injection (AENI) stabilizes the otherwise unreliable attention-logit noise while recovering quality. High-temperature sampling inflates reading grade level and causes catastrophic collapse on several models. We find internal representation-level perturbation to be a more suitable diversity strategy than output-level stochasticity for constrained educational content generation.