Continuous Pruning of Deep Convolutional Networks Using Selective Weight Decay

11/20/2020
by   Hugo Tessier, et al.
0

During the last decade, deep convolutional networks have become the reference for many machine learning tasks, especially in computer vision. However, large computational needs make them hard to deploy on resource-constrained hardware. Pruning has emerged as a standard way to compress such large networks. Yet, the severe perturbation caused by most pruning approaches is thought to hinder their efficacy. Drawing inspiration from Lagrangian Smoothing, we introduce a new technique, Selective Weight Decay (SWD), which achieves continuous pruning throughout training. Our theoretically-grounded approach is versatile and can be applied to any problem, network or pruning structure. We show that SWD compares favorably to other approaches in terms of performance/parameters ratio on the CIFAR-10 and ImageNet ILSVRC2012 datasets. On CIFAR-10 and unstructured pruning, for a target rate of 0.1 accuracy of 81.32 and structured pruning, for a target rate of 2.5 reference technique drops at 10 accuracy at 93.22 pruning and the same target rate of 2.5 instead of the 77.07

READ FULL TEXT

Please sign up or login with your details

Forgot password? Click here to reset