Space Network of Experts: Architecture and Expert Placement

arXiv cs.AI / 5/4/2026

💬 OpinionDeveloper Stack & InfrastructureIdeas & Deep AnalysisModels & Research

Key Points

  • The paper proposes a satellite-based approach to run energy-intensive LLMs using high-efficiency continuous solar energy harvesting envisioned for space data centers.
  • It highlights a key bottleneck: deploying a large LLM across satellites is difficult due to constrained onboard compute and communication, turning it into a model-component “placement” problem.
  • The authors introduce the Space Network of Experts (Space-XNet) framework for distributed execution of a mixture-of-experts (MoE) model in space.
  • Space-XNet uses a two-level placement strategy: assigning MoE layers to ring-arranged satellite subnets and then mapping individual experts within each layer/subnet to satellites based on activation probability and expected routing latency.
  • Experiments on a constellation of over 1,000 satellites show at least a threefold latency reduction versus random and ablation placement baselines.

Abstract

Leveraging continuous solar energy harvesting at high efficiency, space data centers are envisioned as a promising platform for executing energy-intensive large language models (LLMs). Recognizing this advantage, space and AI conglomerates (e.g., SpaceX, Google) are actively investing in this vision. One key challenge, however, is the efficient distributed deployment of a large-scale LLM in a satellite network due to the limited onboard computing and communication resources. This gives rise to a placement problem that involves partitioning and mapping model components to satellites such that the fundamentally different model architecture and network topology can be reconciled to ensure low-latency token generation. To address this problem, we present the Space Network of Experts (Space-XNet) framework targeting the distributed execution of a popular mixture-of-experts (MoE) model in space. The proposed placement strategies are two-level: (1) layer placement, which assigns MoE layers to satellite subnets; and (2) intra-layer expert placement, which assigns individual experts to satellites associated with the same layer/subnet. For layer placement, we exploit the ring-like communication pattern of autoregressive inference to partition the satellite constellation along the orbiting direction into subnets arranged on a ring, each hosting one MoE layer. Based on this architecture, we formulate and solve an optimization problem for intra-layer expert placement to map experts with heterogeneous activation probabilities onto satellites. The derived strategy reveals an intuitive principle: a frequently activated expert should be mapped to a satellite on a routing path with low expected latency. Experiments over a thousand-satellite constellation show that Space-XNet achieves at least a threefold latency reduction compared with conventional random and ablation-based placement strategies.