Dive into Claude Code: The Design Space of Today's and Future AI Agent Systems

arXiv cs.LG / 4/17/2026

💬 OpinionDeveloper Stack & InfrastructureIdeas & Deep AnalysisModels & Research

Key Points

  • Claude Code is an agentic coding tool that can execute shell commands, edit files, and call external services, and the study analyzes its full architecture from its public TypeScript source code.
  • The research frames the architecture around five human-centric values—human decision authority, safety/security, reliable execution, capability amplification, and contextual adaptability—and maps them into thirteen concrete design principles.
  • The system’s core control flow is a simple while-loop that alternates between calling the model and running tools, while most sophistication comes from surrounding components like a permission system, context compaction pipeline, extensibility mechanisms, and isolated subagent delegation.
  • A detailed comparison with OpenClaw shows that similar agent design questions lead to different architectural choices depending on deployment context, such as action-level safety classification versus perimeter access control.
  • The paper outlines six open research/design directions for future AI agent systems, based on recent empirical, architectural, and policy literature.

Abstract

Claude Code is an agentic coding tool that can run shell commands, edit files, and call external services on behalf of the user. This study describes its comprehensive architecture by analyzing the publicly available TypeScript source code and further comparing it with OpenClaw, an independent open-source AI agent system that answers many of the same design questions from a different deployment context. Our analysis identifies five human values, philosophies, and needs that motivate the architecture (human decision authority, safety and security, reliable execution, capability amplification, and contextual adaptability) and traces them through thirteen design principles to specific implementation choices. The core of the system is a simple while-loop that calls the model, runs tools, and repeats. Most of the code, however, lives in the systems around this loop: a permission system with seven modes and an ML-based classifier, a five-layer compaction pipeline for context management, four extensibility mechanisms (MCP, plugins, skills, and hooks), a subagent delegation mechanism with worktree isolation, and append-oriented session storage. A comparison with OpenClaw, a multi-channel personal assistant gateway, shows that the same recurring design questions produce different architectural answers when the deployment context changes: from per-action safety classification to perimeter-level access control, from a single CLI loop to an embedded runtime within a gateway control plane, and from context-window extensions to gateway-wide capability registration. We finally identify six open design directions for future agent systems, grounded in recent empirical, architectural, and policy literature.