GaitKD: A Universal Decoupled Distillation Framework for Efficient Gait Recognition

arXiv cs.CV / 4/30/2026

📰 NewsDeveloper Stack & InfrastructureTools & Practical UsageModels & Research

Key Points

  • The paper introduces GaitKD, a knowledge distillation framework aimed at improving efficient gait recognition models that are otherwise hard to deploy due to heavy architectures.
  • GaitKD decouples distillation into two parts—decision-level distillation using part-calibrated logit distillation, and boundary-level distillation using an activation-boundary objective to preserve the teacher’s embedding-space partitioning.
  • Instead of direct feature regression, the framework uses boundary preservation to achieve more stable student performance, especially for part-structured supervision signals.
  • Experiments on multiple gait recognition benchmarks with various teacher–student configurations show consistent gains over strong baselines without adding inference-time cost.
  • The authors report that the two transfer components are complementary and release source code on GitHub for replication and further use.

Abstract

Gait recognition is an attractive biometric modality for long-range and contact-free identification, but high-performing gait models often rely on deep and computationally expensive architectures that are difficult to deploy in practice. Knowledge distillation (KD) offers a natural way to transfer knowledge from a powerful teacher to an efficient student; however, standard KD is often less effective for part-structured gait models, where supervision is formed from both part-wise classification logits and part-wise retrieval embeddings. In this paper, we propose GaitKD, a distillation framework that decouples gait knowledge transfer into two complementary components: decision-level distillation and boundary-level distillation. Specifically, GaitKD aligns the teacher and student through part-calibrated logit distillation to transfer inter-class decision relations, while preserving the teacher-induced partitioning of the embedding space through an activation-boundary objective instead of direct feature regression. With a simple aligned part-wise design, GaitKD supports heterogeneous teacher-student gait models without introducing additional inference cost. Experimental results across multiple gait recognition benchmarks and teacher-student configurations show consistent improvements over strong gait baselines. Our study demonstrates that the two transfer components are complementary, and boundary-preserving distillation provides more stable performance than direct feature regression. Source code is available at https://github.com/liyiersan/GaitKD/