ClawLess: A Security Model of AI Agents

arXiv cs.AI / 4/10/2026

💬 OpinionDeveloper Stack & InfrastructureIdeas & Deep AnalysisModels & Research

Key Points

  • The paper introduces ClawLess, a security framework for autonomous LLM-based AI agents that assumes a worst-case scenario where the agent may be adversarial.
  • It argues that training or prompting-based controls cannot provide fundamental security guarantees and instead proposes formally verified policies.
  • ClawLess defines a fine-grained security model covering system entities, trust scopes, and permissions, with policies that can adapt to an agent’s runtime behavior.
  • The framework translates the formal policies into enforceable security rules and implements enforcement via a user-space kernel using BPF-based syscall interception.

Abstract

Autonomous AI agents powered by Large Language Models can reason, plan, and execute complex tasks, but their ability to autonomously retrieve information and run code introduces significant security risks. Existing approaches attempt to regulate agent behavior through training or prompting, which does not offer fundamental security guarantees. We present ClawLess, a security framework that enforces formally verified policies on AI agents under a worst-case threat model where the agent itself may be adversarial. ClawLess formalizes a fine-grained security model over system entities, trust scopes, and permissions to express dynamic policies that adapt to agents' runtime behavior. These policies are translated into concrete security rules and enforced through a user-space kernel augmented with BPF-based syscall interception. This approach bridges the formal security model with practical enforcement, ensuring security regardless of the agent's internal design.