Leveraging multimodal information for event summarization and concept-level sentiment analysis

作者:

Highlights:

摘要

The rapid growth in the amount of user-generated content (UGCs) online necessitates for social media companies to automatically extract knowledge structures (concepts) from photos and videos to provide diverse multimedia-related services. However, real-world photos and videos are complex and noisy, and extracting semantics and sentics from the multimedia content alone is a very difficult task because suitable concepts may be exhibited in different representations. Hence, it is desirable to analyze UGCs from multiple modalities for a better understanding. To this end, we first present the EventBuilder system that deals with semantics understanding and automatically generates a multimedia summary for a given event in real-time by leveraging different social media such as Wikipedia and Flickr. Subsequently, we present the EventSensor system that aims to address sentics understanding and produces a multimedia summary for a given mood. It extracts concepts and mood tags from visual content and textual metadata of UGCs, and exploits them in supporting several significant multimedia-related services such as a musical multimedia summary. Moreover, EventSensor supports sentics-based event summarization by leveraging EventBuilder as its semantics engine component. Experimental results confirm that both EventBuilder and EventSensor outperform their baselines and efficiently summarize knowledge structures on the YFCC100M dataset.

论文关键词:Multimedia summarization,Semantics analysis,Sentics analysis,Multimodal analysis,Multimedia-related services,00-01,99-00,

论文评审过程:Received 15 November 2015, Revised 7 May 2016, Accepted 10 May 2016, Available online 11 May 2016, Version of Record 12 August 2016.

论文官网地址:https://doi.org/10.1016/j.knosys.2016.05.022