LayerTracer: A Joint Task-Particle and Vulnerable-Layer Analysis framework for Arbitrary Large Language Model Architectures

arXiv cs.CL / 4/23/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper introduces LayerTracer, an architecture-agnostic, end-to-end analysis framework that can be applied to diverse LLM architectures beyond standard Transformers.
  • LayerTracer examines each layer’s hidden states and links them to vocabulary probability distributions to jointly identify where a model begins executing a given task and which layers are most vulnerable.
  • It defines a “task particle” as the key layer where the target token probability first increases significantly, enabling task-effective layer localization.
  • It defines a “vulnerable layer” using the maximum Jensen–Shannon divergence between output distributions before and after mask perturbation to quantify sensitivity to disturbances.
  • Experiments across different model sizes suggest task particles appear primarily in deep layers regardless of parameter count, while larger models show stronger hierarchical robustness, providing guidance for hybrid architecture design and optimization.

Abstract

Currently, Large Language Models (LLMs) feature a diversified architectural landscape, including traditional Transformer, GateDeltaNet, and Mamba. However, the evolutionary laws of hierarchical representations, task knowledge formation positions, and network robustness bottleneck mechanisms in various LLM architectures remain unclear, posing core challenges for hybrid architecture design and model optimization. This paper proposes LayerTracer, an architecture-agnostic end-to-end analysis framework compatible with any LLM architecture. By extracting hidden states layer-by-layer and mapping them to vocabulary probability distributions, it achieves joint analysis of task particle localization and layer vulnerability quantification. We define the task particle as the key layer where the target token probability first rises significantly, representing the model's task execution starting point, and the vulnerable layer is defined as the layer with the maximum Jensen-Shannon (JS) divergence between output distributions before and after mask perturbation, reflecting its sensitivity to disturbances. Experiments on models of different parameter scales show that task particles mainly appear in the deep layers of the model regardless of parameter size, while larger-parameter models exhibit stronger hierarchical robustness. LayerTracer provides a scientific basis for layer division, module ratio, and gating switching of hybrid architectures, effectively optimizing model performance. It accurately locates task-effective layers and stability bottlenecks, offering universal support for LLM structure design and interpretability research.