Trained Rank Pruning For Efficient Deep Neural Networks
EMC2 Workshop @ NeurIPS 2019 1
Efficient Deep Neural Networks EMC2 Workshop @ NeurIPS 2019 1 - - PowerPoint PPT Presentation
Trained Rank Pruning For Efficient Deep Neural Networks EMC2 Workshop @ NeurIPS 2019 1 Outline Low Rank (LR) Models Methods on obtaining LR models Decompose a pre-trained model Retrain a LR decomposed model Challenges on
EMC2 Workshop @ NeurIPS 2019 1
EMC2 Workshop @ NeurIPS 2019 2
prediction loss. Fine-tuning is required to recover some accuracy loss.
achieve good balance of model capacity and compression
EMC2 Workshop @ NeurIPS 2019 3
Our trained rank pruning method has 2 interleaved steps: (A) Conventional SGD training with nuclear norm regularization and sub-gradient, conditioning the network to be LR compatible
๐๐๐ ๐ ๐ฆ; ๐ฅ + ๐ เท
๐=1 ๐
||๐||โ
๐๐ก๐ฃ๐ = โ๐ + ๐๐๐ข๐ ๐ฃ๐
๐ข๐ ๐ฃ ๐
where ๐ = ๐โ๐๐ is the SVD decomposition and ๐๐ข๐ ๐ฃ, ๐
๐ข๐ ๐ฃ are truncated ๐, ๐ with ๐ ๐๐๐(๐).
(B) Training with LR decomposition, obtaining the LR network with rank pruning
EMC2 Workshop @ NeurIPS 2019 4 [1] H. Avron, S. Kale, S. P. Kasiviswanathan, and V. Sindhwani. Efficient and practical stochastic subgradient descent for nuclear norm regularization. In ICML, 2012.
EMC2 Workshop @ NeurIPS 2019 5
SGD with Nuclear Norm regularization SGD with Nuclear Norm regularization Training with low-rank decomposition
m SGD iterations
All comparison decomposition and pruning results here are finetuned to improve accuracy, while our methods results are from direct decomposition after training.
On both CIFAR-10 and ImageNet datasets, it shows that our TRP methods can outperform other existing methods both in channel-wise decomposition and spatial-wise decomposition formats. It achieves better balance of accuracy and complexity.
EMC2 Workshop @ NeurIPS 2019 6