CPT: Continual Pre-training—updating the model on new large-scale corpora (self-supervised) to learn new facts, domains, or languages
CIT: Continual Instruction Tuning—fine-tuning the model on a stream of supervised instruction-following tasks to improve response to commands
CA: Continual Alignment—updating the model to adhere to evolving human values, ethical standards, and preferences (often via RLHF)
Catastrophic Forgetting: The tendency of neural networks to abruptly lose previously learned information upon learning new information
Experience Replay: A continual learning strategy that stores a small subset of old data to mix with new data during training
PLMs: Pre-trained Language Models—often referring to smaller predecessors of LLMs (like BERT or RoBERTa) which had simpler adaptation strategies
Domain-incremental: A setting where the task structure remains the same but the input distribution (domain) changes over time (e.g., medical text vs. legal text)
Task-incremental: A setting where the model encounters entirely new types of tasks or classes over time
RAG: Retrieval-Augmented Generation—fetching external data at inference time rather than updating model weights
Model Editing: Directly modifying specific model parameters to fix specific factual errors, distinct from general continual learning