← Back to Paper List

Parameter-Efficient Fine-Tuning with Discrete Fourier Transform

Ziqi Gao, Qichao Wang, Aochuan Chen, Zijing Liu, Bingzhe Wu, Liang Chen, Jia Li
Not explicitly reported in the paper
International Conference on Machine Learning (2024)
Pretraining MM Benchmark

📝 Paper Summary

Parameter-Efficient Fine-Tuning (PEFT) Model Compression
FourierFT compresses fine-tuning updates by learning only a tiny fraction of spectral coefficients in the Fourier domain rather than using low-rank matrices in the spatial domain.
Core Problem
Fine-tuning Large Foundation Models (LFMs) requires storing massive weight updates; even efficient methods like LoRA face storage challenges when scaling to many customized tasks or larger base models.
Why it matters:
  • Storing fine-tuned checkpoints for many users creates high storage and bandwidth costs for model hubs (e.g., Civitai)
  • Mobile applications need aggressively compressed weights to load customized models into limited RAM
  • Current methods like LoRA still require significant parameters (e.g., 33.5M for LLaMA2-7B instruction tuning), limiting scalability
Concrete Example: A LoRA adapter for a specific Stable Diffusion style requires ~40MB. Storing thousands of user-customized styles becomes prohibitively expensive for community platforms.
Key Novelty
FourierFT (Fourier Transform for Fine-Tuning)
  • Instead of factorizing weight changes into low-rank matrices (like LoRA), treats the weight change matrix as a spatial signal
  • Learns a very sparse set of coefficients in the frequency domain (Fourier basis) to reconstruct the spatial weight update
  • Uses a fixed random mask to select which frequencies to learn, significantly reducing trainable parameters compared to rank-decomposition methods
Evaluation Highlights
  • Surpasses LoRA on LLaMA2-7B instruction tuning with only 0.064M parameters (vs. LoRA's 33.5M), a ~500x reduction
  • Achieves comparable performance to Full Fine-Tuning on LLaMA2-7B with only 128K parameters
  • Matches LoRA performance on GLUE benchmark (RoBERTa) using only ~6-9% of LoRA's parameter count
Breakthrough Assessment
7/10
Offers drastic parameter reduction (orders of magnitude) over LoRA while maintaining performance. The approach of using Fourier basis for weight updates is a novel divergence from the dominant low-rank paradigm.
×