From Skill Text to Skill Structure: The Scheduling-Structural-Logical Representation for Agent Skills

arXiv cs.CL / 4/28/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper argues that current LLM agent “skills” are often stored as text-heavy artifacts where machine-usable evidence is embedded in natural-language, making it hard for agents to reason over invocation interfaces and side effects.
  • It proposes the Scheduling-Structural-Logical (SSL) representation, which disentangles scheduling signals, execution structure, and logic-level action/resource-use evidence into an explicit structured form.
  • The authors implement SSL using an LLM-based normalizer and test it on a skill corpus for two tasks: Skill Discovery and Risk Assessment.
  • SSL outperforms text-only baselines, improving MRR for Skill Discovery (0.573 → 0.707) and macro F1 for Risk Assessment (0.744 → 0.787).
  • The results suggest explicit, source-grounded structure makes agent skills easier to search, review, and reuse, positioning SSL as a practical step toward more inspectable and operationally actionable skill representations.

Abstract

LLM agents increasingly rely on reusable skills, capability packages that combine instructions, control flow, constraints, and tool calls. In most current agent systems, however, skills are still represented by text-heavy artifacts, including SKILL{.}md-style documents and structured records whose machine-usable evidence remains embedded largely in natural-language descriptions. This poses a challenge for skill-centered agent systems: managing skill collections and using skills to support agent both require reasoning over invocation interfaces, execution structure, and concrete side effects that are often entangled in a single textual surface. An explicit representation of skill knowledge may therefore help make these artifacts easier for machines to acquire and leverage. Drawing on Memory Organization Packets, Script Theory, and Conceptual Dependency from Schank and Abelson's classical work on linguistic knowledge representation, we introduce what is, to our knowledge, the first structured representation for agent skill artifacts that disentangles skill-level scheduling signals, scene-level execution structure, and logic-level action and resource-use evidence: the Scheduling-Structural-Logical (SSL) representation. We instantiate SSL with an LLM-based normalizer and evaluate it on a corpus of skills in two tasks, Skill Discovery and Risk Assessment, and superiorly outperform the text-only baselines: in Skill Discovery, SSL improves MRR from 0.573 to 0.707; in Risk Assessment, it improves macro F1 from 0.744 to 0.787. These findings reveal that explicit, source-grounded structure makes agent skills easier to search and review. They also suggest that SSL is best understood as a practical step toward more inspectable, reusable, and operationally actionable skill representations for agent systems, rather than as a finished standard or an end-to-end mechanism for managing and using skills.