Gradual Magnitude Pruning: A technique where network weights are slowly set to zero during training based on their absolute value, following a schedule (e.g., polynomial decay).
Impala Architecture: A specific ResNet-based deep neural network architecture commonly used in RL, consisting of residual blocks.
Sparsity: The percentage of parameters in a neural network that are set to zero (inactive).
IQM: Interquartile Mean—a robust aggregate metric that calculates the mean of the middle 50% of scores, reducing the impact of outliers.
Replay Ratio: The number of gradient updates performed per environment step collected.
Offline RL: Training RL agents using a fixed dataset of previously collected interactions without further environment interaction.
Dormant Neurons: Neurons in a neural network that become inactive (zero output) during training and stop contributing to the network's function.