In-Context Learning (ICL): The ability of language models to learn tasks from a few examples in the prompt without parameter updates.
Chain-of-Thought (CoT): Prompting strategy where the model generates intermediate reasoning steps before the final answer.
Composite Task: A task that requires applying a sequence of simple functions (skills) to an input (e.g., first find the antonym, then swap the words).
Simple Task: A basic functional mapping (e.g., finding the antonym of a word).
Naïve CoT: Standard Chain-of-Thought where composite examples are broken into steps, but simple task examples remain in their original input-output format.
ExpCoT: Expanded Chain-of-Thought—a method that formats simple task examples as composite chains with missing steps marked by placeholders to align them structurally.
Inner Attention: The internal attention weights of the Transformer model, analyzed here to see which examples the model focuses on.