Integrating multi-modal content analysis and hyperbolic visualization for large-scale news video retrieval and exploration

作者:

Highlights:

摘要

In this paper, we have developed a novel scheme to achieve more effective analysis, retrieval and exploration of large-scale news video collections by performing multi-modal video content analysis and synchronization. First, automatic keyword extraction is performed on news closed captions and audio channels to detect the most interesting news topics (i.e., keywords for news topic interpretation), and the associations among these news topics (i.e., contextual relationships among the news topics) are further determined according to their co-occurrence probabilities. Second, visual semantic items, such as human faces, text captions, video concepts, are extracted automatically by using our semantic video analysis techniques. The news topics are automatically synchronized with the most relevant visual semantic items. In addition, an interestingness weight is assigned for each news topic to characterize its importance. Finally, a novel hyperbolic visualization scheme is incorporated to visualize large-scale news topics according to their associations and interestingness. With a better global overview of large-scale news video collections, users can specify their queries more precisely and explore large-scale news video collections interactively. Our experiments on large-scale news video collections have provided very positive results.

论文关键词:Multi-modal content analysis,Interestingness assignment,Association determination,Hyperbolic visualization

论文评审过程:Received 15 April 2008, Accepted 29 April 2008, Available online 15 May 2008.

论文官网地址:https://doi.org/10.1016/j.image.2008.04.014