AI Navigate

Data-Local Autonomous LLM-Guided Neural Architecture Search for Multiclass Multimodal Time-Series Classification

arXiv cs.LG / 3/18/2026

📰 NewsDeveloper Stack & InfrastructureIdeas & Deep AnalysisModels & Research

Key Points

  • The article introduces a data-local, LLM-guided NAS framework that runs all training and evaluation on-premises, ensuring raw data never leaves the facility while guiding search remotely through trial-level summaries.
  • It uses a multiclass, multimodal setup with one-vs-rest binary experts per class and modality-specific preprocessing plus a lightweight fusion MLP, enabling joint search over architectures and preprocessing steps.
  • Evaluation on two datasets (UEA30 and SleepEDFx) shows the method can achieve performance within published ranges while reducing manual intervention by enabling unattended architecture search.
  • Importantly, the controller observes only trial-level descriptors, metrics, learning-curve statistics, and failure logs, never accessing raw samples or intermediate feature representations, addressing data-privacy constraints in healthcare and similar domains.

Abstract

Applying machine learning to sensitive time-series data is often bottlenecked by the iteration loop: Performance depends strongly on preprocessing and architecture, yet training often has to run on-premise under strict data-local constraints. This is a common problem in healthcare and other privacy-constrained domains (e.g., a hospital developing deep learning models on patient EEG). This bottleneck is particularly challenging in multimodal fusion, where sensor modalities must be individually preprocessed and then combined. LLM-guided neural architecture search (NAS) can automate this exploration, but most existing workflows assume cloud execution or access to data-derived artifacts that cannot be exposed. We present a novel data-local, LLM-guided search framework that handles candidate pipelines remotely while executing all training and evaluation locally under a fixed protocol. The controller observes only trial-level summaries, such as pipeline descriptors, metrics, learning-curve statistics, and failure logs, without ever accessing raw samples or intermediate feature representations. Our framework targets multiclass, multimodal learning via one-vs-rest binary experts per class and modality, a lightweight fusion MLP, and joint search over expert architectures and modality-specific preprocessing. We evaluate our method on two regimes: UEA30 (public multivariate time-series classification dataset) and SleepEDFx sleep staging (heterogeneous clinical modalities such as EEG, EOG, and EMG). The results show that the modular baseline model is strong, and the LLM-guided NAS further improves it. Notably, our method finds models that perform within published ranges across most benchmark datasets. Across both settings, our method reduces manual intervention by enabling unattended architecture search while keeping sensitive data on-premise.