Timezone: »
Poster
Learning Best Combination for Efficient N:M Sparsity
Yuxin Zhang · Mingbao Lin · ZhiHang Lin · Yiting Luo · Ke Li · Fei Chao · Yongjian Wu · Rongrong Ji
By forcing N out of M consecutive weights to be non-zero, the recent N:M fine-grained network sparsity has received increasing attention with its two attractive advantages over traditional irregular network sparsity methods: 1) Promising performance at a high sparsity. 2) Significant speedups when performed on NVIDIA A100 GPUs. Current implementation on N:M sparsity requires a tedious pre-training phase or computationally heavy from-scratch training. To circumvent these problems, this paper presents an efficient solution for achieving N:M fine-grained sparsity from scratch. Specifically, we first make a re-formulation to convert the N:M fine-grained sparsity into a combinatorial problem, in which, the object falls into choosing the best weight combination among $C_M^N$ candidates. Then, we equip each combination with a learnable importance score, which can be jointly optimized along with its associated weights. Through rigorous proof, we demonstrate that the magnitude of the optimized score well reflects the importance of its corresponding weights combination to the training loss. Therefore, by gradually removing combinations with smaller scores till the best one is left, N:M fine-grained sparsity can be efficiently optimized during the normal training phase without any extra expenditure. Comprehensive experimental results have demonstrated that our proposed method for learning best combination, dubbed as LBC, consistently increases the efficacy of the off-the-shelf N:M methods across varying networks and datasets. Our project is released at https://github.com/zyxxmu/LBC.
Author Information
Yuxin Zhang (Xiamen University)
Mingbao Lin (Xiamen University)
ZhiHang Lin (Xiamen University)
Yiting Luo (Xiamen University)
Ke Li (Tencent)
Fei Chao (Aberystwyth University)
Yongjian Wu (Tencent Technology (Shanghai) Co.,Ltd)
Rongrong Ji (Xiamen University, China)
More from the Same Authors
-
2022 Poster: Make Sharpness-Aware Minimization Stronger: A Sparsified Perturbation Approach »
Peng Mi · Li Shen · Tianhe Ren · Yiyi Zhou · Xiaoshuai Sun · Rongrong Ji · Dacheng Tao -
2022 Poster: PyramidCLIP: Hierarchical Feature Alignment for Vision-language Model Pretraining »
Yuting Gao · Jinfeng Liu · Zihan Xu · Jun Zhang · Ke Li · Rongrong Ji · Chunhua Shen -
2021 Poster: Revisiting Discriminator in GAN Compression: A Generator-discriminator Cooperative Compression Scheme »
Shaojie Li · Jie Wu · Xuefeng Xiao · Fei Chao · Xudong Mao · Rongrong Ji -
2020 Poster: Rotated Binary Neural Network »
Mingbao Lin · Rongrong Ji · Zihan Xu · Baochang Zhang · Yan Wang · Yongjian Wu · Feiyue Huang · Chia-Wen Lin -
2020 Poster: UWSOD: Toward Fully-Supervised-Level Capacity Weakly Supervised Object Detection »
Yunhang Shen · Rongrong Ji · Zhiwei Chen · Yongjian Wu · Feiyue Huang -
2019 Poster: Variational Structured Semantic Inference for Diverse Image Captioning »
Fuhai Chen · Rongrong Ji · Jiayi Ji · Xiaoshuai Sun · Baochang Zhang · Xuri Ge · Yongjian Wu · Feiyue Huang · Yan Wang -
2019 Poster: FreeAnchor: Learning to Match Anchors for Visual Object Detection »
Xiaosong Zhang · Fang Wan · Chang Liu · Rongrong Ji · Qixiang Ye -
2019 Poster: Information Competing Process for Learning Diversified Representations »
Jie Hu · Rongrong Ji · ShengChuan Zhang · Xiaoshuai Sun · Qixiang Ye · Chia-Wen Lin · Qi Tian