MobyDeep: A lightweight CNN architecture to configure models for text classification

作者:

Highlights:

摘要

Nowadays, trends in deep learning for text classification are addressed to create complex models to deal with huge datasets. Deeper models are usually based on cutting edge neural network architectures, achieving good results in general but demanding better hardware than shallow ones. In this work, a new Convolutional Neural Network (CNN) architecture (MobyDeep) for text classification tasks is proposed. Designed as a configurable tool, resultant models (MobyNets) are able to manage big corpora sizes under low computational costs. To achieve those milestones, the architecture was conceived to produce lightweight models, having their internal layers based on a new proposed convolutional block. That block was designed and customized by adapting ideas from image to text processing, helping to squeezing model sizes and to reduce computational costs. The architecture was also designed as a residual network, covering complex functions by extending models up to 28 layers. Moreover, middle layers were optimized by residual connections, helping to remove fully connected layers on top and resulting in Fully CNN. Corpus were chosen from the recent literature, aiming to define real scenarios when comparing configured MobyDeep models with other state-of the-art works. Thus, three models were configured in 8, 16 and 28 layers respectively, offering competitive accuracy results.

论文关键词:Convolutional lightweight architecture,Network layer optimization,Text classification

论文评审过程:Received 21 March 2022, Revised 9 September 2022, Accepted 14 September 2022, Available online 24 September 2022, Version of Record 12 October 2022.

论文官网地址:https://doi.org/10.1016/j.knosys.2022.109914