← Back to Paper List

An Empirical Study and Theoretical Explanation on Task-Level Model-Merging Collapse

Yuan Cao, Dezhi Ran, Yuzhe Guo, Mengzhou Wu, Simin Chen, Linyi Li, Wei Yang, Tao Xie
arXiv (2026)
Pretraining Benchmark

📝 Paper Summary

Model Merging Multi-Task Learning
Merging collapse is driven by representational incompatibility between tasks, identifiable via hidden-state distance, rather than parameter-level conflicts, and is theoretically bounded by the diameter of representation clusters.
Core Problem
Combining independently fine-tuned models often results in merging collapse, where the unified model suffers catastrophic performance degradation (up to -32.8% loss) compared to its constituents.
Why it matters:
  • Current model merging techniques fail unpredictably when scaled to multiple tasks, hindering the efficient reuse of specialized LLMs without retraining
  • Existing literature incorrectly attributes merging failure to parameter-space conflicts (e.g., sign disagreements), leading to ineffective mitigation strategies
  • Developers lack principled metrics to determine beforehand which task combinations will successfully merge versus which will collapse
Concrete Example: When merging Qwen2.5-3B models fine-tuned on GLUE tasks, combining the QQP and WNLI tasks leads to severe performance degradation across all merging methods (LA, TIES, etc.), while other pairs merge successfully. Parameter conflict metrics fail to flag this pair as problematic.
Key Novelty
Representation-Theoretic Merging Compatibility
  • Applies rate-distortion theory to model merging, proving that merging distortion is lower-bounded by the diameter of task-specific hidden state clusters
  • Demonstrates that task-level representational conflicts (geometry of hidden states) are the true predictor of collapse, unlike parameter update conflicts
  • Introduces Hidden-state Distance Similarity and Merging Difficulty Score (MDS) to quantitatively predict mergeability before combining models
Evaluation Highlights
  • Parameter conflict metrics (sign/magnitude change) show no statistically significant correlation with merging collapse (p-values > 0.05) across all experiments, refuting common assumptions
  • Proposed Hidden-state Distance Similarity strongly correlates with merging performance (p-values < 0.05) across 5 merging methods and 64 checkpoints
  • 2/3 of the 25 tested Lots-of-LoRAs task groups suffered >30% performance loss, confirming collapse is a widespread normative failure mode
Breakthrough Assessment
8/10
Significantly challenges the prevailing wisdom that parameter conflicts cause merging failure. Provides a solid theoretical proof (rate-distortion) and a practical metric that actually correlates with empirical results.
×