CoTD: Chain-of-Thought Distillation—teaching small models to reason by fine-tuning them on rationales generated by large models
SLM: Small Language Model—models with significantly fewer parameters (e.g., <1B) compared to LLMs
Self-Ask-Self-Ans: A prompting strategy where an LLM iteratively asks and answers its own sub-questions to solve a complex problem
BM25: Best Matching 25—a probabilistic information retrieval algorithm based on term frequency and inverse document frequency
Decomposer: The student model responsible for breaking down the main question into sub-questions or deciding the final answer
Responser: The student model responsible for answering the sub-questions generated by the Decomposer, using retrieved documents
F1 score: A metric measuring the overlap between the predicted answer and the ground truth answer
EM: Exact Match—a metric checking if the predicted answer is identical to the ground truth
Hallucination: When a model generates plausible-sounding but factually incorrect information