CFT: Critical Token Fine-tuning—the proposed method that updates only tokens essential for correctness
Counterfactual perturbation: The process of replacing a token with an alternative prediction to see if the outcome (final answer) changes
SFT: Supervised Fine-Tuning—training a pre-trained model on labeled data using maximum likelihood estimation
Critical token: A token whose substitution with a plausible alternative results in an incorrect final answer
Pass@N: A metric measuring the probability that at least one correct solution is found among N generated samples
GRPO: Group Relative Policy Optimization—a reinforcement learning algorithm used here to test the fine-tuned models as initializations
Greedy decoding: Selecting the highest-probability token at each step during text generation
Parallel decoding: Evaluating multiple counterfactual paths for different positions simultaneously to speed up processing
Entropy: A measure of randomness or diversity in the model's output distribution