LoRA: Low-Rank Adaptation—a PEFT method that approximates weight updates as the product of two low-rank matrices
PEFT: Parameter-Efficient Fine-Tuning—techniques to adapt large models by training only a small subset of parameters
Quantum Circuit: A model of computation where a state is manipulated by a sequence of unitary gates; used here as an analogy for tensor operations
Einsum: Einstein Summation—a notation for expressing tensor operations (contractions) concisely
Universality Theorem: A theorem stating that any matrix can be decomposed into a sequence of two-axis tensors (analogous to quantum gates)
Composition Openness: The property that the composition of two matrices from a set may fall outside that set; unlike low-rank matrices, QuanTA satisfies this, allowing expressivity to grow with depth
Intrinsic Rank: The minimum rank required to effectively represent the weight update necessary for a specific downstream task
DoRA: Weight-Decomposed Low-Rank Adaptation—a variant of LoRA that decomposes weights into magnitude and direction