Confidence Deficit: A phenomenon where a model doubts its own correct intermediate steps due to low internal confidence, triggering unnecessary reflection
Termination Delay: A phenomenon where a model continues generating text (reflections) even after reaching a correct and confident answer
SimPO: Simple Preference Optimization—a training method that aligns models with preferences (here, conciseness) without a reference model, often more memory efficient than DPO
SFT: Supervised Fine-Tuning—training a model on high-quality target outputs (here, the concise chains generated by ConCISE)
Confidence Injection: The technique of inserting affirmative phrases (e.g., 'It is clear that...') into the context to artificially raise the model's confidence and prevent it from generating reflection steps
Early Stopping: Terminating the generation process when a detector indicates the model's internal confidence in the answer exceeds a threshold
Probing Prompt: A short text appended to the context to elicit a probability distribution over specific tokens (e.g., 'Wait', 'Great') that serves as a proxy for the model's internal state