Specializing Large Models for Oracle Bone Script Interpretation via Component-Grounded Multimodal Knowledge Augmentation

arXiv cs.CV / 4/9/2026

📰 NewsSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper proposes an agent-driven vision-language model framework to decipher Oracle Bone Script by explicitly grounding character components and then reasoning over their semantics to close the “interpretation gap” left by closed-set image recognition methods.
  • It combines a vision-language model for component-level visual grounding with an LLM-based agent that automates a reasoning pipeline including component identification, graph-based knowledge retrieval, and relationship inference.
  • The authors introduce OB-Radix, a new expert-annotated dataset containing 1,022 character images (934 unique) and 1,853 fine-grained component images spanning 478 components with verified explanations and structural/semantic labels.
  • Experiments across three benchmarks indicate the approach produces more detailed and more precise decipherments than baseline methods, emphasizing the benefit of component reuse and transferable pictographic semantics.
  • The work is positioned as a specialized large-model method for a historical-visual decoding task, suggesting a reusable blueprint for other interpretive domains where objects are built from semantically meaningful subcomponents.

Abstract

Deciphering ancient Chinese Oracle Bone Script (OBS) is a challenging task that offers insights into the beliefs, systems, and culture of the ancient era. Existing approaches treat decipherment as a closed-set image recognition problem, which fails to bridge the ``interpretation gap'': while individual characters are often unique and rare, they are composed of a limited set of recurring, pictographic components that carry transferable semantic meanings. To leverage this structural logic, we propose an agent-driven Vision-Language Model (VLM) framework that integrates a VLM for precise visual grounding with an LLM-based agent to automate a reasoning chain of component identification, graph-based knowledge retrieval, and relationship inference for linguistically accurate interpretation. To support this, we also introduce OB-Radix, an expert-annotated dataset providing structural and semantic data absent from prior corpora, comprising 1,022 character images (934 unique characters) and 1,853 fine-grained component images across 478 distinct components with verified explanations. By evaluating our system across three benchmarks of different tasks, we demonstrate that our framework yields more detailed and precise decipherments compared to baseline methods.