SCHK-HTC: Sibling Contrastive Learning with Hierarchical Knowledge-Aware Prompt Tuning for Hierarchical Text Classification

arXiv cs.CL / 4/20/2026

📰 NewsModels & Research

Key Points

  • The paper addresses few-shot hierarchical text classification (HTC), where texts must be assigned to labels arranged in a tree structure under limited data.
  • It argues that existing methods’ reliance on hierarchical consistency constraints is insufficient, especially for separating semantically similar sibling classes when domain knowledge is scarce.
  • The proposed method, SCHK-HTC, combines a hierarchical knowledge extraction module with sibling contrastive learning guided by hierarchical knowledge-aware prompt tuning.
  • By learning discriminative representations at multiple levels of the label hierarchy, the approach improves the separability of confusing classes.
  • Experiments on three benchmark datasets show performance gains over prior state-of-the-art methods in most cases, with accompanying code released on GitHub.

Abstract

Few-shot Hierarchical Text Classification (few-shot HTC) is a challenging task that involves mapping texts to a predefined tree-structured label hierarchy under data-scarce conditions. While current approaches utilize structural constraints from the label hierarchy to maintain parent-child prediction consistency, they face a critical bottleneck, the difficulty in distinguishing semantically similar sibling classes due to insufficient domain knowledge. We introduce an innovative method named Sibling Contrastive Learning with Hierarchical Knowledge-aware Prompt Tuning for few-shot HTC tasks (SCHK-HTC). Our work enhances the model's perception of subtle differences between sibling classes at deeper levels, rather than just enforcing hierarchical rules. Specifically, we propose a novel framework featuring two core components: a hierarchical knowledge extraction module and a sibling contrastive learning mechanism. This design guides model to encode discriminative features at each hierarchy level, thus improving the separability of confusable classes. Our approach achieves superior performance across three benchmark datasets, surpassing existing state-of-the-art methods in most cases. Our code is available at https://github.com/happywinder/SCHK-HTC.