NLL: Negative Log-Likelihood—the standard loss function for training language models, equivalent to Cross-Entropy
Trust Gate: A term in the gradient equation that scales the update magnitude based on the model's current confidence
Focus Index (alpha): A parameter in the deformed-log loss family; alpha=0 recovers NLL, alpha=1 recovers linear probability loss
Cayley Transform: A conformal mapping used here to smoothly interpolate the focus index from 0 to 1 based on the model's uncertainty radius
Rényi-2 Entropy: A measure of distribution concentration (collision entropy) used as a proxy for the model's predictive state
Tsallis Entropy: A generalized entropy family that includes Shannon entropy as a limit case; used to define the optimization geometry
Confident Conflicts: Situations where the pretrained model is confident in a prediction that disagrees with the SFT target (often due to noise in SFT data)
Sharpening: The process of pushing a model's output probabilities from 'mostly correct' (e.g., 0.8) to 'highly certain' (e.g., 0.99)
Deformed-log family: A generalization of the logarithm function parameterized by q (or alpha), allowing for tunable sensitivity to probabilities