SLM: Small Language Model—models with fewer parameters (typically <13B) that are cheaper to run and fine-tune
LLM: Large Language Model—models with massive parameter counts (e.g., GPT-3, GPT-4) trained on vast datasets
LoRA: Low-Rank Adaptation—a parameter-efficient fine-tuning technique that freezes pre-trained weights and injects trainable rank decomposition matrices
Chain-of-Thought: A prompting technique where the model generates intermediate reasoning steps before the final answer
AutoTrain: A Hugging Face tool that simplifies the process of training and fine-tuning machine learning models without extensive coding
Gretel.ai: A platform for generating synthetic data, used here to create mathematical QA pairs
MMLU: Massive Multitask Language Understanding—a benchmark designed to measure knowledge acquired during pretraining
Hallucination: When a language model generates plausible-sounding but factually incorrect information