Deep compression of probabilistic graphical networks

作者:

Highlights:

• Deep compression of these deterministic models has been proposed in last few years to reduce the number of connections and nodes, while remaining the classification accuracy of these models. This paper is the first attempt to combine deep probabilistic graphical networks (PGNs) and deep compression techniques together to derive sparse versions of the deep probabilistic models.

• The developed pruning approach is a layer-by-layer and pre-pruning one. That is, the final PGNs grow in a layer-by-layer and compressed manner. It is very efficient in redundancy reduction. The whole pruning only takes a few seconds to conduct, while other complex methods take much more time to select sparse architectures or conduct model compression, such as regularization and low-rank factorization approaches. These model compression techniques are more time-consuming compared with pruning due to the complexity of regularization and factorization.

• The correctness and efficiency of the proposed compression approach are evaluated by a number of PGNs with different datasets. It can be used to compress trained PGNs to achieve a lightweight network which can be embedded in on-chip memory of small mobile devices. It is also flexible to implement and can be generalized to other deep PGNs.

摘要

•Deep compression of these deterministic models has been proposed in last few years to reduce the number of connections and nodes, while remaining the classification accuracy of these models. This paper is the first attempt to combine deep probabilistic graphical networks (PGNs) and deep compression techniques together to derive sparse versions of the deep probabilistic models.•The developed pruning approach is a layer-by-layer and pre-pruning one. That is, the final PGNs grow in a layer-by-layer and compressed manner. It is very efficient in redundancy reduction. The whole pruning only takes a few seconds to conduct, while other complex methods take much more time to select sparse architectures or conduct model compression, such as regularization and low-rank factorization approaches. These model compression techniques are more time-consuming compared with pruning due to the complexity of regularization and factorization.•The correctness and efficiency of the proposed compression approach are evaluated by a number of PGNs with different datasets. It can be used to compress trained PGNs to achieve a lightweight network which can be embedded in on-chip memory of small mobile devices. It is also flexible to implement and can be generalized to other deep PGNs.

论文关键词:Deep compression,Probabilistic graphical models,Probabilistic graphical networks,Deep learning

论文评审过程:Received 10 November 2018, Revised 24 May 2019, Accepted 19 July 2019, Available online 20 July 2019, Version of Record 24 July 2019.

论文官网地址:https://doi.org/10.1016/j.patcog.2019.106979