OOD: Out-of-Distribution—data that differs from the training data in style, environment, or other covariates while sharing the same task
Weight Averaging: Averaging the weights of multiple fine-tuned models (e.g., Model Soups) to approximate an ensemble, often improving robustness
Sparsity Bias: The tendency of optimization algorithms like SGD to rely on the smallest subset of features necessary to solve the training task, often discarding redundant information
Weakly Relevant Features: Features that are redundant/unnecessary for the training distribution (due to correlations) but may become necessary if the primary features are corrupted or missing in the test distribution
Three-distributions setup: A transfer learning framework involving a pre-training distribution, a fine-tuning distribution, and a distinct testing distribution
ERM: Empirical Risk Minimization—standard training that minimizes the average loss on the training data
Linear Connectivity: The property where linearly interpolating between the weights of two neural networks yields models with low loss, enabling weight averaging
Penultimate Layer: The layer of a neural network immediately preceding the final classification head (linear layer)