PTQ: Post-Training Quantization—quantizing a model after it has been trained, without requiring a full re-training process.
QAT: Quantization-Aware Training—simulating quantization during the training process to adapt weights, usually requiring the full dataset.
Rounding-to-nearest (RTN): The standard approach of rounding a continuous value to the closest integer grid point.
AdaRound: A state-of-the-art PTQ method that learns whether to round weights up or down (using addition), but is restricted to the two nearest grids.
Element-wise division: Dividing each element of a matrix by the corresponding element of another matrix (or a broadcasted scalar).
Frobenius norm: The square root of the sum of the absolute squares of the elements of a matrix, used here as a distance metric for reconstruction error.
Block-wise reconstruction: Optimizing quantization parameters to minimize the error between the output of a quantized block of layers and the original full-precision block.