LiteVLA-H: Dual-Rate Vision-Language-Action Inference for Onboard Aerial Guidance and Semantic Perception

arXiv cs.CV / 5/5/2026

📰 NewsDeveloper Stack & InfrastructureTools & Practical UsageIndustry & Market MovesModels & Research

Key Points

  • LiteVLA-H is a compact 256M-parameter vision-language-action (VLA) model proposed for low-latency onboard drone deployment under strict compute and communication constraints.
  • The system uses dual-rate operation on an NVIDIA Jetson AGX Orin: a fast outer-loop for reactive guidance with short action-token outputs and a slower semantic mode for hazard/scene understanding and operator narration.
  • The authors find that, in the edge setting, end-to-end latency is largely dominated by multimodal pre-fill rather than by the additional decoding cost of a few more tokens, motivating their scheduling approach.
  • They report reactive action-token issuance at 50.65 ms (19.74 Hz) while still producing sentence-level semantic outputs at about 149.90–164.57 ms (6.08–6.67 Hz) on the same embedded platform.
  • A knowledge-preserving fine-tuning recipe mixes flight data, aerial semantic data, and generic caption/VQA supervision to specialize for aerial guidance without losing descriptive competence.

Abstract

Vision-language-action (VLA) models have shown strong semantic grounding and task generalization in manipulation, but aerial deployment remains difficult because drones require low-latency closed-loop guidance under strict onboard compute and communication constraints. We present LiteVLA-H, a compact 256M-parameter VLA system designed for dual-rate operation on an NVIDIA Jetson AGX Orin: a fast outer-loop guidance mode for short action-token outputs and a slower semantic mode for scene understanding, hazard description, and operator-facing narration. The central empirical observation is that, in this compact edge regime, end-to-end latency is dominated by multimodal pre-fill rather than by the marginal cost of decoding a few extra tokens. This motivates a scheduler that issues reactive action tokens at 50.65,ms (19.74,Hz) while still supporting sentence-level semantic outputs at 149.90--164.57\ms (6.08--6.67,Hz) on the same embedded platform. To specialize the model without collapsing its descriptive competence, we use a knowledge-preserving fine-tuning recipe that mixes reactive flight data, aerial semantic data, and generic caption/VQA supervision. Beyond reporting current latency measurements, we position the system against recent state-of-the-art architectures, including AnywhereVLA, FutureVLA, and ReMem-VLA, showing that the measured action branch reaches a higher edge inference rate under our deployment conditions while retaining periodic semantic awareness.