scholarly journals TRP: Trained Rank Pruning for Efficient Deep Neural Networks

Author(s):  
Yuhui Xu ◽  
Yuxi Li ◽  
Shuai Zhang ◽  
Wei Wen ◽  
Botao Wang ◽  
...  

To enable DNNs on edge devices like mobile phones, low-rank approximation has been widely adopted because of its solid theoretical rationale and efficient implementations. Several previous works attempted to directly approximate a pre-trained model by low-rank decomposition; however, small approximation errors in parameters can ripple over a large prediction loss. As a result, performance usually drops significantly and a sophisticated effort on fine-tuning is required to recover accuracy. Apparently, it is not optimal to separate low-rank approximation from training. Unlike previous works, this paper integrates low rank approximation and regularization into the training process. We propose Trained Rank Pruning (TRP), which alternates between low rank approximation and training. TRP maintains the capacity of the original network while imposing low-rank constraints during training. A nuclear regularization optimized by stochastic sub-gradient descent is utilized to further promote low rank in TRP. The TRP trained network inherently has a low-rank structure, and is approximated with negligible performance loss, thus eliminating the fine-tuning process after low rank decomposition. The proposed method is comprehensively evaluated on CIFAR-10 and ImageNet, outperforming previous compression methods using low rank approximation.

2014 ◽  
Vol 2014 ◽  
pp. 1-11
Author(s):  
Jinjiang Li ◽  
Mengjun Li ◽  
Hui Fan

Existing image inpainting algorithm based on low-rank matrix approximation cannot be suitable for complex, large-scale, damaged texture image. An inpainting algorithm based on low-rank approximation and texture direction is proposed in the paper. At first, we decompose the image using low-rank approximation method. Then the area to be repaired is interpolated by level set algorithm, and we can reconstruct a new image by the boundary values of level set. In order to obtain a better restoration effect, we make iteration for low-rank decomposition and level set interpolation. Taking into account the impact of texture direction, we segment the texture and make low-rank decomposition at texture direction. Experimental results show that the new algorithm is suitable for texture recovery and maintaining the overall consistency of the structure, which can be used to repair large-scale damaged image.


2021 ◽  
Vol 54 (7) ◽  
pp. 103-107
Author(s):  
Antonio Fazzi ◽  
Nicola Guglielmi ◽  
Ivan Markovsky ◽  
Konstantin Usevich

Author(s):  
Mei Sun ◽  
Jinxu Tao ◽  
Zhongfu Ye ◽  
Bensheng Qiu ◽  
Jinzhang Xu ◽  
...  

Background: In order to overcome the limitation of long scanning time, compressive sensing (CS) technology exploits the sparsity of image in some transform domain to reduce the amount of acquired data. Therefore, CS has been widely used in magnetic resonance imaging (MRI) reconstruction. </P><P> Discussion: Blind compressed sensing enables to recover the image successfully from highly under- sampled measurements, because of the data-driven adaption of the unknown transform basis priori. Moreover, analysis-based blind compressed sensing often leads to more efficient signal reconstruction with less time than synthesis-based blind compressed sensing. Recently, some experiments have shown that nonlocal low-rank property has the ability to preserve the details of the image for MRI reconstruction. Methods: Here, we focus on analysis-based blind compressed sensing, and combine it with additional nonlocal low-rank constraint to achieve better MR images from fewer measurements. Instead of nuclear norm, we exploit non-convex Schatten p-functionals for the rank approximation. </P><P> Results & Conclusion: Simulation results indicate that the proposed approach performs better than the previous state-of-the-art algorithms.


2020 ◽  
Vol 14 (12) ◽  
pp. 2791-2798
Author(s):  
Xiaoqun Qiu ◽  
Zhen Chen ◽  
Saifullah Adnan ◽  
Hongwei He

2020 ◽  
Vol 6 ◽  
pp. 922-933
Author(s):  
M. Amine Hadj-Youcef ◽  
Francois Orieux ◽  
Alain Abergel ◽  
Aurelia Fraysse

2021 ◽  
Vol 11 (10) ◽  
pp. 4582
Author(s):  
Kensuke Tanioka ◽  
Satoru Hiwa

In the domain of functional magnetic resonance imaging (fMRI) data analysis, given two correlation matrices between regions of interest (ROIs) for the same subject, it is important to reveal relatively large differences to ensure accurate interpretation. However, clustering results based only on differences tend to be unsatisfactory and interpreting the features tends to be difficult because the differences likely suffer from noise. Therefore, to overcome these problems, we propose a new approach for dimensional reduction clustering. Methods: Our proposed dimensional reduction clustering approach consists of low-rank approximation and a clustering algorithm. The low-rank matrix, which reflects the difference, is estimated from the inner product of the difference matrix, not only from the difference. In addition, the low-rank matrix is calculated based on the majorize–minimization (MM) algorithm such that the difference is bounded within the range −1 to 1. For the clustering process, ordinal k-means is applied to the estimated low-rank matrix, which emphasizes the clustering structure. Results: Numerical simulations show that, compared with other approaches that are based only on differences, the proposed method provides superior performance in recovering the true clustering structure. Moreover, as demonstrated through a real-data example of brain activity measured via fMRI during the performance of a working memory task, the proposed method can visually provide interpretable community structures consisting of well-known brain functional networks, which can be associated with the human working memory system. Conclusions: The proposed dimensional reduction clustering approach is a very useful tool for revealing and interpreting the differences between correlation matrices, even when the true differences tend to be relatively small.


2018 ◽  
Vol 20 (10) ◽  
pp. 2659-2669 ◽  
Author(s):  
Jiandong Tian ◽  
Zhi Han ◽  
Weihong Ren ◽  
Xiai Chen ◽  
Yandong Tang

Sign in / Sign up

Export Citation Format

Share Document