PruneTrain: Gradual Structured Pruning from Scratch for Faster Neural Network Training

01/26/2019
by   Sangkug Lym, et al.
0

Model pruning is a popular mechanism to make a network more efficient for inference. In this paper, we explore the use of pruning to also make the training of such neural networks more efficient. Unlike all prior model pruning methods that sparsify a pre-trained model and then prune it, we train the network from scratch, while gradually and structurally pruning parameters during the training. We build on our key observations: 1) once parameters are sparsified via regularization, they rarely re-appear in later steps, and 2) setting the appropriate regularization penalty at the beginning of training effectively converges the loss. We train ResNet and VGG networks on CIFAR10/100 and ImageNet datasets from scratch, and achieve 30-50 FLOPs and 20-30

READ FULL TEXT

Please sign up or login with your details

Forgot password? Click here to reset