PERSONAL Sign in with your SPIE account to access your personal subscriptions or to use specific features such as save to my library, sign up for alerts, save searches, etc.
Low rank approximation is an effective method in deep neural network (DNN) compression. In view of the fact that the redundancy information content of different network layers is different, a novel iterative low-rank approximation method based on the redundancy of each network layer is proposed. By giving priority to the network layer with higher redundancy, the loss of intrinsic information in each network layer is expected to be reduced and the performance of the compressed model is improved. Experimental results show that the performance of compressed model obtained by this method is improved with a slight reduction in compression ratio. It can be concluded that the proposed method can better retain intrinsic information in the pre-training network.
Fan Yang,Weirong Liu,Jie Liu,Chaorong Liu,Yanchun Mi, andHaowen Song
"Iterative low-rank approximation based on the redundancy of each network layer", Proc. SPIE 11720, Twelfth International Conference on Graphics and Image Processing (ICGIP 2020), 117202G (27 January 2021); https://doi.org/10.1117/12.2589425
ACCESS THE FULL ARTICLE
INSTITUTIONAL Select your institution to access the SPIE Digital Library.
PERSONAL Sign in with your SPIE account to access your personal subscriptions or to use specific features such as save to my library, sign up for alerts, save searches, etc.
The alert did not successfully save. Please try again later.
Fan Yang, Weirong Liu, Jie Liu, Chaorong Liu, Yanchun Mi, Haowen Song, "Iterative low-rank approximation based on the redundancy of each network layer," Proc. SPIE 11720, Twelfth International Conference on Graphics and Image Processing (ICGIP 2020), 117202G (27 January 2021); https://doi.org/10.1117/12.2589425