Learning Human-Object Interaction for 3D Human Pose Estimation from LiDAR Point Clouds
arXiv cs.CV / 3/18/2026
📰 NewsModels & Research
Key Points
- The paper proposes a Human-Object Interaction Learning (HOIL) framework to improve 3D human pose estimation from LiDAR point clouds by leveraging human-object interactions to mitigate spatial ambiguity and class imbalance in interaction regions.
- It introduces HOICL, a human-object interaction-aware contrastive learning module that enhances feature discrimination between human and object points, especially where interactions occur.
- It introduces CPPool, a contact-aware part-guided pooling mechanism that adaptively reallocates representation capacity by compressing overrepresented non-interacting points while preserving informative points from interacting body parts.
- It also offers an optional contact-based temporal refinement that uses contact cues over time to refine per-frame keypoint estimates; code will be released.
Related Articles

Math needs thinking time, everyday knowledge needs memory, and a new Transformer architecture aims to deliver both
THE DECODER
Kreuzberg v4.5.0: We loved Docling's model so much that we gave it a faster engine
Reddit r/LocalLLaMA
Today, what hardware to get for running large-ish local models like qwen 120b ?
Reddit r/LocalLLaMA
Running mistral locally for meeting notes and it's honestly good enough for my use case
Reddit r/LocalLLaMA
[D] Single-artist longitudinal fine art dataset spanning 5 decades now on Hugging Face — potential applications in style evolution, figure representation, and ethical training data
Reddit r/MachineLearning