PAL: Program-Aided Language modelsโa prompting technique where the reasoning chain is executable code (e.g., Python) rather than natural language.
CoT: Chain-of-Thoughtโa prompting method that encourages the model to generate intermediate reasoning steps before the final answer.
Backward Process: A synthesis step where the LLM generates a reasoning chain first, then generates a question that fits that chain.
Forward Process: A synthesis step where the LLM takes a generated question and produces a new, high-quality reasoning chain for it.
In-Cluster Complexity: A selection strategy where examples are grouped by semantic similarity, and the example with the most reasoning steps is chosen from each group.
Target Complexity: A constraint used during synthesis where the model is instructed to generate a reasoning chain with a specific number of steps/lines.
Self-Consistency: A technique (often called majority voting) where multiple reasoning paths are sampled, and the most common answer is selected; used here to validate synthetic examples.