Rate Distortion For Model Compression: From Theory To Practice

10/09/2018
by   Weihao Gao, et al.
0

As the size of neural network models increases dramatically today, study of model compression algorithms becomes important. Despite many practically successful compression methods, the fundamental limit of model compression remains unknown. In this paper, we study the fundamental limit for model compression via rate distortion theory. We bring the rate distortion function from data compression to model compression to quantify the fundamental limit. We prove a lower bound for the rate distortion function and prove its achievability for linear models. Motivated by our theory, we further present a pruning algorithm which takes consideration of the structure of neural networks and demonstrate its good performance for both synthetic and real neural network models.

READ FULL TEXT

Please sign up or login with your details

Forgot password? Click here to reset