Residual Stream: The primary vector pathway in a Transformer where information is added by attention and MLP layers, preserving a linear additive structure throughout the network
OV Circuit: Output-Value circuit; the component of an attention head (W_V * W_O) responsible for writing information to the residual stream based on what was attended to
QK Circuit: Query-Key circuit; the component of an attention head (W_Q * W_K^T) responsible for determining which previous tokens to attend to (computing attention patterns)
Privileged Basis: A specific basis in the activation space (usually aligned with neurons) where individual dimensions carry semantic meaning, enforced by element-wise nonlinearities
LayerNorm: A normalization operation that centers and scales representations; geometrically interpreted as projecting inputs onto a hyperplane and then mapping to a hypersphere
Induction Heads: Specific attention mechanisms that copy the token following a previous occurrence of the current token, enabling in-context learning
Virtual Attention Heads: Effective attention operations formed by the composition of attention heads across different layers, allowing later heads to attend to information moved by earlier heads