PCAS: Pruning Channels with Attention Statistics

06/14/2018
by   Kohei Yamamoto, et al.
0

To implement deep neural networks on small embedded devices, conventional techniques use channel pruning looking considering manual compression rate per layer to reduce parameters. Besides it is difficult to consider the relationships between layers and it takes a lot of time for deeper models. For addressing these issues, we propose a new channel pruning technique based on attention that can evaluate the importance of channels. We improved the method with the criterion to allow the automatic channel selection using a single compression rate for the entire model. Experimental results showed that a parameter reduction of 90.8 accuracy degradation of around 1 CIFAR-10 benchmark.

READ FULL TEXT

Please sign up or login with your details

Forgot password? Click here to reset