PPO: Proximal Policy Optimization—a popular policy gradient method that constrains updates using a clipped surrogate objective to ensure a trust region
feature rank collapse: A phenomenon where the neural network's internal representations lose dimensionality, becoming less expressive and unable to distinguish states effectively
plasticity: The ability of a neural network to continue learning and adapting to new data distributions over time
capacity loss: A metric measuring the decrease in a network's ability to fit random target labels, indicating a loss of learning capability
trust region: A constraint in optimization (like PPO's clipping) that prevents the new policy from moving too far from the old policy to ensure safe updates
non-stationarity: The condition where the data distribution (states and rewards) changes over time, which is inherent in RL as the agent's policy changes
pre-activations: The values in a neural network layer before the non-linear activation function (e.g., ReLU) is applied
GAE: Generalized Advantage Estimator—a method to estimate the advantage function (how good an action is) by balancing bias and variance
PFO: Proximal Feature Optimization—the authors' proposed auxiliary loss to regularize changes in feature representations