Arithmetic in the Wild: Llama uses Base-10 Addition to Reason About Cyclic Concepts

arXiv cs.AI / 5/5/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper studies how Llama-3.1-8B reasons about cyclic concepts (like “what month is six months after August”) and finds the model does not directly perform modular arithmetic in the cycle’s period (e.g., 12 months).
  • Instead, Llama-3.1-8B appears to use a generic, task-agnostic addition mechanism: it first computes a base-10 sum of the two inputs, then maps that intermediate result back into the cyclic concept space.
  • The authors argue that the summation is carried out using task-agnostic Fourier features whose periods align with base-10 addition (e.g., 2, 5, 10) rather than the cyclic concept’s own period.
  • Mechanistically, the study identifies a small, reusable set of 28 MLP neurons at layer 18 (about 0.2% of that layer’s MLP) that cluster into groups, each implementing a sum for a Fourier feature with a distinct period.
  • Overall, the work connects causal abstraction and feature-geometry choices to improve mechanistic interpretability of language models.

Abstract

Does structure in representations imply structure in computation? We study how Llama-3.1-8B reasons over cyclic concepts (e.g., "what month is six months after August?"). Even though Llama-3.1-8B's representations for these concepts are circularly structured, we find that instead of directly computing modular addition in the period of the cyclic concept (e.g., 12 for months), the model re-uses a generic addition mechanism across tasks that operates independently of concept-specific geometry. First, it computes the sum of its two inputs using base-10 addition (six + August=14). Then, it maps this sum back to cyclic concept space (14->February). We show that Llama-3.1-8B uses task-agnostic Fourier features to compute these sums--in fact, these features have periods that respect standard base-10 addition, e.g., 2, 5, and 10, rather than the cyclic concept period (e.g., 12 for months). Furthermore, we identify a sparse set of 28 MLP neurons re-used across all tasks (approximately 0.2% of the MLP at layer 18) that can be partitioned into disjoint clusters, each computing the sum for a Fourier feature with a different period. Our work highlights how an interplay between causal abstraction and feature geometry can deepen our mechanistic understanding of LMs.