Saliency Score: A metric computed by multiplying attention weights by their gradients (w.r.t loss), representing the intensity of information flow between tokens
Information Flow: The transfer of semantic information between different parts of the input (Question, Prompt) and output (Rationale) across model layers
Zero-shot CoT: Prompting an LLM to reason (e.g., 'Let's think step by step') without providing any example demonstrations
IAP: Instance-Adaptive Prompting—the proposed method to select prompts based on saliency scores
GSM8K: Grade School Math 8K—a benchmark dataset of high quality grade school math word problems
MMLU: Massive Multitask Language Understanding—a benchmark covering 57 subjects across STEM, the humanities, and social sciences
BBH: BIG-Bench Hard—a subset of the BIG-Bench benchmark focusing on tasks where LLMs struggle
Plan-and-Solve: A prompting strategy that asks the model to devise a plan before solving the problem
Self-Discover: A method where the model selects and composes reasoning modules to solve a task