University of Amsterdam and Renmin University at TRECVID 2016: Searching Video, Detecting Events and Describing Video
| Authors |
|
|---|---|
| Publication date | 11-2016 |
| Event | TRECVID workshop 2016 |
| Number of pages | 5 |
| Organisations |
|
| Abstract |
In this paper we summarize our TRECVID 2016 video recognition experiments. We participated in three tasks: video search, event detection and video description. Here we describe the tasks on event detection and video descrip-tion. For event detection we explore semantic representa-tions based on VideoStory and an ImageNet Shuffle for both zero-shot and few-example regimes. For the showcase task on video description we experiment with a deep network that predicts a visual representation from a natural language de-scription, and use this space for the sentence matching. For generative description we enhance a neural image caption-ing model with Early Embedding and Late Reranking. The 2016 edition of the TRECVID benchmark has been a fruitful participation for our joint-team, resulting in the best overall result for zero- and few-example event detection as well as video description by matching and in generative mode.
|
| Document type | Paper |
| Language | English |
| Published at | https://www-nlpir.nist.gov/projects/tvpubs/tv16.papers/mediamill.pdf https://www.semanticscholar.org/paper/University-of-Amsterdam-and-Renmin-University-at-Snoek-Dong/d7bbd75e9471dbcb20a04043f8156cb967567f3f |
| Other links | https://ivi.fnwi.uva.nl/isis/publications/2016/SnoekPTRECVID2016 |
| Downloads |
mediamill
(Final published version)
|
| Permalink to this page | |
