Communication-Efficient and Robust Multi-Modal Federated Learning via Latent-Space Consensus
arXiv cs.LG / 3/20/2026
📰 NewsModels & Research
Key Points
- Introduces CoMFed, a communication-efficient multi-modal federated learning framework that uses learnable projection matrices to create compressed latent representations.
- A latent-space regularizer aligns representations across clients to improve cross-modal consistency and robustness to outliers.
- The approach addresses heterogeneity in modalities and model architectures while preserving privacy and reducing communication overhead.
- Experimental results on human activity recognition benchmarks show competitive accuracy with minimal overhead.
Related Articles
[R] Combining Identity Anchors + Permission Hierarchies achieves 100% refusal in abliterated LLMs — system prompt only, no fine-tuning
Reddit r/MachineLearning
[P] Vibecoded on a home PC: building a ~2700 Elo browser-playable neural chess engine with a Karpathy-inspired AI-assisted research loop
Reddit r/MachineLearning
Meet DuckLLM 1.0 My First Model!
Reddit r/LocalLLaMA
Since FastFlowLM added support for Linux, I decided to benchmark all the models they support, here are some results
Reddit r/LocalLLaMA
What measure do I use to compare nested models and non nested models in high dimensional survival analysis [D]
Reddit r/MachineLearning