HarassGuard: Detecting Harassment Behaviors in Social Virtual Reality with Vision-Language Models

arXiv cs.CV / 4/2/2026

📰 NewsSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper introduces HarassGuard, a vision-language model–based system for detecting physical harassment behaviors in social VR using only visual input to reduce privacy risks from biometric data.
  • It reports the creation of an IRB-approved harassment vision dataset and describes fine-tuning VLMs with prompt engineering and contextual information to improve detection in social VR scenes.
  • Experimental results show competitive performance against traditional baselines (LSTM/CNN and Transformer), with up to 88.09% accuracy for binary classification and 68.85% for multi-class classification.
  • The authors claim HarassGuard can achieve similar baseline performance with substantially fewer fine-tuning samples (200 vs. 1,115), highlighting improved data efficiency and contextual reasoning benefits.

Abstract

Social Virtual Reality (VR) platforms provide immersive social experiences but also expose users to serious risks of online harassment. Existing safety measures are largely reactive, while proactive solutions that detect harassment behavior during an incident often depend on sensitive biometric data, raising privacy concerns. In this paper, we present HarassGuard, a vision-language model (VLM) based system that detects physical harassment in social VR using only visual input. We construct an IRB-approved harassment vision dataset, apply prompt engineering, and fine-tune VLMs to detect harassment behavior by considering contextual information in social VR. Experimental results demonstrate that HarassGuard achieves competitive performance compared to state-of-the-art baselines (i.e., LSTM/CNN, Transformer), reaching an accuracy of up to 88.09% in binary classification and 68.85% in multi-class classification. Notably, HarassGuard matches these baselines while using significantly fewer fine-tuning samples (200 vs. 1,115), offering unique advantages in contextual reasoning and privacy-preserving detection.