A system for the semantic multi-modal analysis of news audio-visual content
News related content is nowadays among the most popular types of content for users in everyday applications. Although the generation and distribution of news content has become commonplace, due to the availability of inexpensive media capturing devices and the development of media sharing services targeting both professional and user-generated news content, the automatic analysis and annotation that is required for supporting intelligent search and delivery of this content remains an open issue. In this paper, a complete architecture for knowledge-assisted multi-modal analysis of news-related multimedia content is presented, along with its constituent components. The proposed analysis architecture employs state-of-theart methods for the analysis of each individual modality (visual, audio, text) separately, and proposes a novel fusion technique based on the particular characteristics of news-related content for the combination of the individual modality analysis results. Experimental results on news broadcast video illustrate the usefulness of the proposed techniques in the automatic generation of semantic annotations.