Generalizing the Geometry of Model Merging Through Frechet Averages
arXiv cs.LG / 5/1/2026
📰 NewsIdeas & Deep AnalysisModels & Research
Key Points
- The paper argues that effective model merging without extra training requires symmetry-aware methods, because naive parameter-space averaging can break under architectural symmetries.
- It proposes a general framework: perform merging as Fréchet averaging by choosing parameters that minimize a sum of geodesic distances on a suitably chosen manifold.
- The authors emphasize that the critical design choice is the overall geometry—specifically the metric, manifold, and distance approximation—which defines what it means for two models to be “close.”
- Under simplifying assumptions, the paper shows that Fréchet averaging can subsume and generalize Fisher merging.
- For low-rank adapters (LoRA), the paper identifies a distinct quotient-manifold geometry, reviews limitations of existing LoRA merging methods, and introduces a practical algorithm with comparisons to other approaches.
Related Articles

Why Autonomous Coding Agents Keep Failing — And What Actually Works
Dev.to

Why Enterprise AI Pilots Fail
Dev.to

The PDF Feature Nobody Asked For (That I Use Every Day)
Dev.to

How to Fix OpenClaw Tool Calling Issues
Dev.to

Mistral's new flagship Medium 3.5 folds chat, reasoning, and code into one model
THE DECODER