Strong Lottery Tickets (SLTs): Subnetworks within randomly initialized neural networks that achieve high accuracy without any weight training, only by pruning
Weak Lottery Tickets: Subnetworks that require further training of weights to achieve optimal performance
Edge-Popup: The primary existing algorithm for finding SLTs, which assigns a score to each weight and selects the top-k scores
Continuously Relaxed Bernoulli Gates: A method to approximate binary (0/1) selection gates with continuous variables (0 to 1) to make the selection process differentiable
L0 regularization: A penalty term that counts the number of non-zero parameters in a model; usually non-differentiable, but approximated here
Concrete distribution: A continuous distribution that approximates discrete random variables (like Bernoulli) to allow for backpropagation
Straight-Through Estimator: A heuristic often used to estimate gradients for discrete variables by passing gradients through a threshold function as if it were an identity function
Sparsity: The percentage of weights in a neural network that are set to zero
Mask: A binary matrix applied to the weights of a network to select which connections are active
Scaled Kaiming Normal: A specific weight initialization strategy that has been found to be critical for the existence of strong lottery tickets