Strong Lottery Ticket Hypothesis with $\epilon$–perturbation

Abstract
The strong Lottery Ticket Hypothesis (LTH) (Ramanujan et al., 2019; Zhou et al., 2019) claims the existence of a subnetwork in a sufficiently large, randomly initialized neural network that approximates some target neural network without the need of training. We extend the theoretical guarantee of the strong LTH literature to a scenario more similar to the original LTH, by generalizing the weight change in the pretraining step to some perturbation around initialization. In particular, we focus on the following open questions: By allowing an ε-scale perturbation on the random initial weights, can we reduce the over-parameterization requirement for the candidate network in the strong LTH? Furthermore, does the weight change by SGD coincide with a good set of such perturbation? We answer the first question by first extending the theoretical result on the subset sum problem (Lueker, 1998) to allow perturbation on the candidates. Applying this result to the neural network setting, we show that by allowing ε-scale perturbation, we can reduce the overparameterization requirement of the strong LTH by a factor of O(1/(1 + ε)). To answer the second question, we show via experiments that the perturbed weight achieved by the projected SGD shows better performance under the strong LTH pruning.
Type
Publication
International Conference on Artificial Intelligence and Statistics, 2023