Keywords :
channel pruning; coupled channels; information bottleneck; neural network compression; variational pruning; vib; Channel pruning; Coupled-channels; Information bottleneck; Network compression; Neural network compression; Neural-networks; Stochastic nature; Structured sparsities; Variational pruning; Vib; Artificial Intelligence; Computer Science Applications; Computer Vision and Pattern Recognition
Abstract :
[en] Variational channel pruning approaches have obtained impressive results thanks to their stochastic nature, well established foundation in information theory, and the practically appealing structured sparsity pattern they offer. Despite their success in pruning Plain Networks (PlainNets), their application has faced certain limitations in networks with structurally coupled channels such as ResNets. In such scenarios, not only is it required to prune structurally coupled channels together, but it is also necessary to ensure that the whole coupled group is irrelevant before pruning is applied. This is an under-investigated problem as most existing methods are designed without taking these couplings into account. In this paper, we propose a novel approach based on Information Theoretic Pruning of structurally Coupled Channels (ITPCC) in neural networks. IT-PCC allows for learning the probabilistic distribution of coupled channel set importance and prunes the ones with the least relevant information to the task at hand. Experimental results for image classification on CIFAR10, CI-FAR100, and ImageNet datasets show that the proposed method outperforms the state-of-the-art, more significantly at high compression rates.
Research center :
See outside Interdisciplinary Centre for Security, Reliability and Trust (SnT), University of Luxembourg
Scopus citations®
without self-citations
0