GeoLAN: Geometric Learning of Latent Explanatory Directions in Large Language Models

arXiv cs.LG / 3/23/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • GeoLAN introduces a training framework that treats token representations as geometric trajectories and uses two differentiable regularizers, KT-CW and KT-Attn, to promote isotropy and diverse attention.
  • The approach aims to enhance mechanistic interpretability and reduce certain fairness biases, with experiments on Gemma-3 and Llama-3-8B showing maintained task accuracy alongside improved geometric metrics, especially in mid-sized models.
  • Results reveal scale-dependent trade-offs between geometric precision and performance, suggesting geometry-aware training as a promising direction for future LLM research.
  • The work highlights a new line of geometry-informed training within models-research and ideas-deep-analysis, emphasizing interpretability over immediate industrial deployment.

Abstract

Large language models (LLMs) demonstrate strong performance, but they often lack transparency. We introduce GeoLAN, a training framework that treats token representations as geometric trajectories and applies stickiness conditions inspired by recent developments related to the Kakeya Conjecture. We have developed two differentiable regularizers, Katz-Tao Convex Wolff (KT-CW) and Katz-Tao Attention (KT-Attn), that promote isotropy and encourage diverse attention. Our experiments with Gemma-3 (1B, 4B, 12B) and Llama-3-8B show that GeoLAN frequently maintains task accuracy while improving geometric metrics and reducing certain fairness biases. These benefits are most significant in mid-sized models. Our findings reveal scale-dependent trade-offs between geometric precision and performance, suggesting that geometry-aware training is a promising approach to enhance mechanistic interpretability.