Beyond N-gram: Data-Aware X-GRAM Extraction for Efficient Embedding Parameter Scaling

arXiv cs.CL / 4/24/2026

📰 NewsDeveloper Stack & InfrastructureModels & Research

Key Points

  • The paper introduces X-GRAM, a frequency-aware dynamic token-injection method to improve the parameter efficiency of large token-indexed lookup tables used for memory-augmented scaling.
  • It argues that prior limitations stem from Zipfian under-training of long-tail tokens, different demand patterns across layers, and “slot collapse” that yields redundant embeddings.
  • X-GRAM uses hybrid hashing and alias mixing to compress the long tail while keeping head capacity, and then refines retrieved vectors with a normalized SwiGLU + ShortConv module to better capture diverse local n-gram features.
  • The extracted signals are injected into attention value streams and inter-layer residuals via depth-aware gating, creating a memory-to-context alignment that decouples model capacity from FLOPs.
  • Experiments at 0.73B and 1.15B model scales report up to +4.4 average accuracy over the vanilla backbone and +3.2 over strong retrieval baselines, while achieving these gains with substantially smaller tables in a 50% setting, with code released on GitHub.

Abstract

Large token-indexed lookup tables provide a compute-decoupled scaling path, but their practical gains are often limited by poor parameter efficiency and rapid memory growth. We attribute these limitations to Zipfian under-training of the long tail, heterogeneous demand across layers, and "slot collapse" that produces redundant embeddings. To address this, we propose X-GRAM, a frequency-aware dynamic token-injection framework. X-GRAM employs hybrid hashing and alias mixing to compress the tail while preserving head capacity, and refines retrieved vectors via normalized SwiGLU ShortConv to extract diverse local n-gram features. These signals are integrated into attention value streams and inter-layer residuals using depth-aware gating, effectively aligning static memory with dynamic context. This design introduces a memory-centric scaling axis that decouples model capacity from FLOPs. Extensive evaluations at the 0.73B and 1.15B scales show that X-GRAM improves average accuracy by as much as 4.4 points over the vanilla backbone and 3.2 points over strong retrieval baselines, while using substantially smaller tables in the 50% configuration. Overall, by decoupling capacity from compute through efficient memory management, X-GRAM offers a scalable and practical paradigm for future memory-augmented architectures. Code aviliable in https://github.com/Longyichen/X-gram.