University of Amsterdam and Renmin University at TRECVID 2016: Searching Video, Detecting Events and Describing Video

Open Access
Authors
Publication date 11-2016
Event TRECVID workshop 2016
Number of pages 5
Organisations
  • Faculty of Science (FNWI) - Informatics Institute (IVI)
Abstract
In this paper we summarize our TRECVID 2016 video recognition experiments. We participated in three tasks: video search, event detection and video description. Here we describe the tasks on event detection and video descrip-tion. For event detection we explore semantic representa-tions based on VideoStory and an ImageNet Shuffle for both zero-shot and few-example regimes. For the showcase task on video description we experiment with a deep network that predicts a visual representation from a natural language de-scription, and use this space for the sentence matching. For generative description we enhance a neural image caption-ing model with Early Embedding and Late Reranking. The 2016 edition of the TRECVID benchmark has been a fruitful participation for our joint-team, resulting in the best overall result for zero- and few-example event detection as well as video description by matching and in generative mode.
Document type Paper
Language English
Published at https://www-nlpir.nist.gov/projects/tvpubs/tv16.papers/mediamill.pdf https://www.semanticscholar.org/paper/University-of-Amsterdam-and-Renmin-University-at-Snoek-Dong/d7bbd75e9471dbcb20a04043f8156cb967567f3f
Other links https://ivi.fnwi.uva.nl/isis/publications/2016/SnoekPTRECVID2016
Downloads
mediamill (Final published version)
Permalink to this page
Back