Deep multimodal fusion for semantic image segmentation: A survey

作者:

Highlights:

摘要

Recent advances in deep learning have shown excellent performance in various scene understanding tasks. However, in some complex environments or under challenging conditions, it is necessary to employ multiple modalities that provide complementary information on the same scene. A variety of studies have demonstrated that deep multimodal fusion for semantic image segmentation achieves significant performance improvement. These fusion approaches take the benefits of multiple information sources and generate an optimal joint prediction automatically. This paper describes the essential background concepts of deep multimodal fusion and the relevant applications in computer vision. In particular, we provide a systematic survey of multimodal fusion methodologies, multimodal segmentation datasets, and quantitative evaluations on the benchmark datasets. Existing fusion methods are summarized according to a common taxonomy: early fusion, late fusion, and hybrid fusion. Based on their performance, we analyze the strengths and weaknesses of different fusion strategies. Current challenges and design choices are discussed, aiming to provide the reader with a comprehensive and heuristic view of deep multimodal image segmentation.

论文关键词:Image fusion,Multi-modal,Deep learning,Semantic segmentation

论文评审过程:Received 25 September 2020, Accepted 29 September 2020, Available online 7 October 2020, Version of Record 12 January 2021.

论文官网地址:https://doi.org/10.1016/j.imavis.2020.104042