Rui Lin
Paper download is intended for registered attendees only, and is
subjected to the IEEE Copyright Policy. Any other use is strongly forbidden.
Papers from this author
Exploiting Elasticity in Tensor Ranks for Compressing Neural Networks
Jie Ran, Rui Lin, Hayden Kwok-Hay So, Graziano Chesi, Ngai Wong
Auto-TLDR; Nuclear-Norm Rank Minimization Factorization for Deep Neural Networks
Abstract Slides Poster Similar
Elasticities in depth, width, kernel size and resolution have been explored in compressing deep neural networks (DNNs). Recognizing that the kernels in a convolutional neural network (CNN) are 4-way tensors, we further exploit a new elasticity dimension along the input-output channels. Specifically, a novel nuclear-norm rank minimization factorization (NRMF) approach is proposed to dynamically and globally search for the reduced tensor ranks during training. Correlation between tensor ranks across multiple layers is revealed, and a graceful tradeoff between model size and accuracy is obtained. Experiments then show the superiority of NRMF over the previous non-elastic variational Bayesian matrix factorization (VBMF) scheme.