CoT: Chain of Thought—a prompting technique where the model is encouraged to generate intermediate reasoning steps before the final answer.
Zero-shot-CoT: The paper's proposed method; eliciting chain of thought reasoning using only a template (e.g., 'Let's think step by step') without example QA pairs.
Few-shot-CoT: Prior method (Wei et al., 2022) requiring manual creation of question-answer pairs with reasoning steps to guide the model.
System-1 vs System-2: Cognitive science distinction; System-1 is fast/intuitive (standard prompting), System-2 is slow/analytical (CoT prompting).
Greedy decoding: A generation strategy where the model selects the highest probability token at each step, making outputs deterministic.
InstructGPT: OpenAI's GPT-3 models fine-tuned with human feedback to follow instructions (e.g., text-davinci-002).
PaLM: A 540-billion parameter dense language model developed by Google.
Self-consistency: A decoding strategy (Wang et al., 2022) that samples multiple reasoning paths and takes a majority vote for the final answer.
Answer extraction: A secondary prompting step used to parse the final numerical or multiple-choice answer from the model's free-text reasoning.