Sal: Self-supervised Analogical Learning—the proposed framework that trains models to transfer reasoning patterns from easy/familiar questions to hard/rare ones via symbolic programs
conceptualization: A process in Sal that finds easier, structurally identical questions to a target question, solves them, and transfers their symbolic solution back to the target
simplification: A process in Sal for math problems that decomposes complex questions into iterative sub-questions to reduce cognitive load and generate reliable symbolic solutions
symbolic solution: A Python program generated by the model that, when executed, produces the final answer (as opposed to generating free text)
ask_llm: A helper function within the symbolic programs that allows the code to query an LLM for knowledge retrieval or soft comparisons
chain-of-thought: A prompting technique where the model generates intermediate reasoning steps before the final answer
silver labels: Labels generated by a model (e.g., via majority voting or high-confidence predictions) rather than human annotation, used for training
base language model: The underlying LLM (e.g., Mixtral-8x7B-Instruct) used to generate data and then fine-tuned on that data