Implicit CoT: Reasoning performed in the model's latent continuous space (hidden states) rather than by generating explicit natural language tokens
Explicit CoT: Standard Chain-of-Thought reasoning where the model generates step-by-step natural language explanations before the answer
Latent Space: The high-dimensional vector space where the model represents data internally, as opposed to the discrete vocabulary space of tokens
Self-Distillation: A training process where a model acts as both teacher and student, transferring knowledge (e.g., reasoning patterns) from one task configuration to another within the same network
CoT Shift: The phenomenon where explicit reasoning tokens change the hidden activation values of the final query token compared to a sequence without reasoning
Curriculum Learning: A training strategy where the model learns from easy to hard examples or gradually changes the task difficulty; used by prior baselines like Coconut but prone to forgetting
Stop-gradient: An operation during training that prevents error gradients from backpropagating through a specific part of the network, used here to freeze the teacher's signals