_comment: REQUIRED: Define ALL technical terms, acronyms, and method names used ANYWHERE in the entire summary. After drafting the summary, perform a MANDATORY POST-DRAFT SCAN: check every section individually (Core.one_sentence_thesis, evaluation_highlights, core_problem, Technical_details, Experiments.key_results notes, Figures descriptions and key_insights). HIGH-VISIBILITY RULE: Terms appearing in one_sentence_thesis, evaluation_highlights, or figure key_insights MUST be defined—these are the first things readers see. COMMONLY MISSED: PPO, DPO, MARL, dense retrieval, silver labels, cosine schedule, clipped surrogate objective, Top-k, greedy decoding, beam search, logit, ViT, CLIP, Pareto improvement, BLEU, ROUGE, perplexity, attention heads, parameter sharing, warm start, convex combination, sawtooth profile, length-normalized attention ratio, NTP. If in doubt, define it.
DP-SGD: Differentially Private Stochastic Gradient Descent—the standard algorithm for private training that adds noise to clipped gradients.
Zeroth-Order Optimization: Optimization methods that estimate gradients using only function values (losses) rather than explicit derivatives.
SPSA: Simultaneous Perturbation Stochastic Approximation—a specific zeroth-order method that estimates gradients using random perturbations.
LoRA: Low-Rank Adaptation—a parameter-efficient fine-tuning technique that freezes pre-trained weights and trains rank-decomposition matrices.
Epsilon-DP: Pure Differential Privacy, a stronger guarantee where the privacy loss is strictly bounded by epsilon (delta=0).
Laplace Mechanism: A DP mechanism that adds noise drawn from a Laplace distribution, typically used for epsilon-DP.
Gaussian Mechanism: A DP mechanism that adds noise drawn from a Gaussian distribution, typically used for (epsilon, delta)-DP.
Per-example gradient clipping: The process in DP-SGD of scaling down individual sample gradients to bound their norm (sensitivity) before aggregation.
Sensitivity: The maximum amount by which a single individual's data can change the function output (in this case, the loss difference scalar).
Membership Inference Attack: An attack that attempts to determine whether a specific data point was used to train a machine learning model.