Block-sparse CNN: towards a fast and memory-efficient framework for convolutional neural networks

作者:Nu Wen, Renzhong Guo, Biao He, Yong Fan, Ding Ma

摘要

In this paper, we propose a block-sparse convolutional neural network (BSCNN) architecture that converts a dense convolution kernel into a sparse one. Traditional convolutional neural networks (CNNs) face the problem that an increase in the number of network parameters will lead to more model and floating-point computations, and a higher risk of network overfitting. The block-sparse convolution uses sparse factor pairs to randomize a sparse convolution kernel, which can introduce mixed information and thereby enabling the extraction of more diverse features. In the meantime, a SUMMA-based parallel computing method is adopted to achieve a lightweight storage and a fast calculation of the convolution kernel. Experimental results show that, compared with current sparse networks, the proposed framework achieves better prediction accuracy than the classical backbone networks in terms of faster floating-point operation and less storage space requirements.

论文关键词:Block-sparse kernel, Convolutional neural network, Sparse matrix multiplication

论文评审过程:

论文官网地址:https://doi.org/10.1007/s10489-020-01815-z