← Back to Paper List

A Stability Analysis of Fine-Tuning a Pre-Trained Model

Zihao Fu, Anthony Man-Cho So, Nigel Collier
arXiv (2023)
Pretraining Benchmark

📝 Paper Summary

Fine-tuning stability Learning theory for NLP
The paper derives theoretical stability bounds for full fine-tuning and head tuning of pre-trained models, identifying that larger sample sizes, smaller learning rates, and closer initialization distances mathematically guarantee more stable training.
Core Problem
Fine-tuning pre-trained models suffers from significant instability, where tuning the same model under the same settings results in widely varying performance.
Why it matters:
  • Instability impairs overall model performance and reliability in deployment
  • It makes different fine-tuned models incomparable, as performance variance swamps method differences
  • Existing solutions (smaller learning rates, noise regularization) lack a unified theoretical understanding of why they work
Concrete Example: When fine-tuning a model like BERT on a downstream task, running the exact same training configuration multiple times can yield classifiers with significantly different accuracy, purely due to the stochastic nature of the optimization.
Key Novelty
Unified Stability Analysis Framework & Derived Stabilization Strategies
  • Approximates full fine-tuning via second-order Taylor expansion to prove that stability is bounded by sample size, Lipschitz constants, and weight distance
  • Models head tuning as training a linear classifier on separable data, proving that stability improves with more iterations, smaller learning rates, and larger margins
  • Proposes three methods based on theory: Maximal Margin Regularizer (MMR) to increase feature margin, Multi-Head Loss (MHLoss) to accelerate convergence, and Self Unsupervised Re-Training (SURT) to minimize weight distance
Breakthrough Assessment
7/10
Provides a strong theoretical foundation for widely observed empirical phenomena in NLP fine-tuning, though the provided text lacks the quantitative results to confirm the efficacy of the proposed solutions.
×