AI Navigate

PLUME: Building a Network-Native Foundation Model for Wireless Traces via Protocol-Aware Tokenization

arXiv cs.LG / 3/17/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • Plume is a new 140M-parameter foundation model designed specifically for learning from structured wireless traces (802.11) using PDML dissections.
  • It introduces a protocol-aware tokenizer that tokenizes along the dissector field tree, inserts timing gap tokens, and normalizes identifiers, reducing sequence length by about 6.2x and increasing per-token density.
  • In evaluation on real-world data, Plume attains 74-97% next-packet token accuracy across five failure categories and AUROC ≥ 0.99 for zero-shot anomaly detection.
  • Frontier LLMs (Claude Opus 4.6 and GPT-5.4) achieve comparable results given identical protocol context, but Plume uses >600x fewer parameters and can run on a single GPU with near-zero marginal cloud cost, enabling on-prem privacy-preserving root cause analysis.
  • The work demonstrates the viability of network-native foundation models and suggests potential downstream benefits for network debugging and security workflows.

Abstract

Foundation models succeed when they learn in the native structure of a modality, whether morphology-respecting tokens in language or pixels in vision. Wireless packet traces deserve the same treatment: meaning emerges from layered headers, typed fields, timing gaps, and cross-packet state machines, not flat strings. We present Plume (Protocol Language Understanding Model for Exchanges), a compact 140M-parameter foundation model for 802.11 traces that learns from structured PDML dissections. A protocol-aware tokenizer splits along the dissector field tree, emits gap tokens for timing, and normalizes identifiers, yielding 6.2x shorter sequences than BPE with higher per token information density. Trained on a curated corpus, Plume achieves 74-97% next-packet token accuracy across five real-world failure categories and AUROC >= 0.99 for zero-shot anomaly detection. On the same prediction task, frontier LLMs (Claude Opus 4.6, GPT-5.4) score comparably despite receiving identical protocol context, yet Plume does so with > 600x fewer parameters, fitting on a single GPU at effectively zero marginal cost vs. cloud API pricing, enabling on-prem, privacy-preserving root cause analysis.