Interactive exploration of journalistic video footage through multimodal semantic matching
Sarah Ibrahimi (Universiteit van Amsterdam)
Shuo Chen (Universiteit van Amsterdam)
Devanshu Arya (Universiteit van Amsterdam)
Arthur Câmara (TU Delft - Web Information Systems)
Yunlu Chen (Universiteit van Amsterdam)
Tanja Crijns (RTL Nederland)
Maurits Van Der Goes (RTL Nederland)
Thomas Mensink (Google LLC)
Emiel Van Miltenburg (Tilburg University)
undefined More Authors (External organisation)
More Info
expand_more
Other than for strictly personal use, it is not permitted to download, forward or distribute the text or part of it, without the consent of the author(s) and/or copyright holder(s), unless the work is under an open content license such as Creative Commons.
Abstract
This demo presents a system for journalists to explore video footage for broadcasts. Daily news broadcasts contain multiple news items that consist of many video shots and searching for relevant footage is a labor intensive task. Without the need for annotated video shots, our system extracts semantics from footage and automatically matches these semantics to query terms from the journalist. The journalist can then indicate which aspects of the query term need to be emphasized, e.g. the title or its thematic meaning. The goal of this system is to support the journalists in their search process by encouraging interaction and exploration with the system.