LLM: Large Language Model—a deep learning algorithm that can recognize, summarize, translate, predict, and generate text
PLM: Pre-trained Language Model—a model trained on a vast corpus of general data before domain-specific adaptation
FLM: Fine-tuning Language Model—the resulting model after adapting a PLM to a specific task or domain
PEFT: Parameter-Efficient Fine-Tuning—methods to adapt LLMs by freezing most parameters and training only a small subset or added adapters
LoRA: Low-Rank Adaptation—a PEFT technique that injects trainable rank decomposition matrices into transformer layers while freezing pre-trained weights
QLoRA: Quantized LoRA—an efficient fine-tuning approach that quantizes the base model to 4-bit precision to reduce memory usage
RAG: Retrieval-Augmented Generation—a technique that optimizes LLM output by referencing an authoritative knowledge base outside its training data
Hallucination: A phenomenon where an LLM generates plausible-sounding but factually incorrect or nonsensical information
Transformer: A deep learning architecture relying on self-attention mechanisms, serving as the backbone for modern LLMs like BERT and GPT