Weaves, Wires, and Morphisms: Formalizing and Implementing the Algebra of Deep Learning

arXiv cs.LG / 4/9/2026

💬 OpinionDeveloper Stack & InfrastructureIdeas & Deep AnalysisModels & Research

Key Points

  • The paper argues that although deep learning models compute precise functions, there is no widely used formal mathematical framework for describing and composing model architectures in a rigorous way.
  • It proposes a categorical framework that formalizes tensor broadcasting using a new concept called axis-stride and introduces array-broadcasted categories.
  • The framework expresses the mathematical functions of architectures in a compositional, manipulable manner and translates those definitions into both human-readable diagrams and machine-readable data structures.
  • To demonstrate the approach, the authors provide reference implementations in Python (pyncd) and TypeScript (tsncd) with capabilities such as algebraic construction, graph conversion, PyTorch compilation, and diagram rendering.
  • The work aims to enable a systematic, formal workflow for deep learning model design and analysis, potentially reducing reliance on ad-hoc notation and pseudocode.

Abstract

Despite deep learning models running well-defined mathematical functions, we lack a formal mathematical framework for describing model architectures. Ad-hoc notation, diagrams, and pseudocode poorly handle nonlinear broadcasting and the relationship between individual components and composed models. This paper introduces a categorical framework for deep learning models that formalizes broadcasting through the novel axis-stride and array-broadcasted categories. This allows the mathematical function underlying architectures to be precisely expressed and manipulated in a compositional manner. These mathematical definitions are translated into human manageable diagrams and machine manageable data structures. We provide a mirrored implementation in Python (pyncd) and TypeScript (tsncd) to show the universal aspect of our framework, along with features including algebraic construction, graph conversion, PyTorch compilation and diagram rendering. This lays the foundation for a systematic, formal approach to deep learning model design and analysis.