RankMe: A metric measuring the effective rank of a matrix based on the Von Neumann entropy of its singular values; higher values indicate higher-dimensional, more isotropic representations
alpha-ReQ: The power-law decay rate of the eigenvalues of the representation covariance matrix; higher alpha indicates faster decay and more compressed, anisotropic representations
RLVR: Reinforcement Learning from Verifiable Rewards—optimizing a policy to maximize rewards based on objective verification (e.g., math correctness) rather than a learned reward model
DPO: Direct Preference Optimization—a method to align language models to preferences by optimizing the relative log-probability of preferred vs. dispreferred responses
SFT: Supervised Fine-Tuning—adapting a pretrained model using labeled instruction-response pairs via maximum likelihood estimation
pass@k: An evaluation metric that estimates the probability of generating at least one correct solution given k independent samples
distributional memorization: The correlation between an LLM's output probabilities and the n-gram frequencies in its pretraining corpus, measured using an infinite-gram model
echolalia: A failure mode where the model repeats inputs or generates repetitive, non-contextual text, observed during the initial representational collapse
isotropic: Uniformity in all directions; in this context, a representation space where variance is spread out across many dimensions (high RankMe)
anisotropic: Directionally dependent; in this context, a representation space where information is compressed along specific principal axes (low RankMe)