Unified Embedding and Metric Learning for Zero-Exemplar Event Detection
| Authors | |
|---|---|
| Publication date | 2017 |
| Book title | 30th IEEE Conference on Computer Vision and Pattern Recognition |
| Book subtitle | CVPR 2017 : 21-26 July 2016, Honolulu, Hawaii : proceedings |
| ISBN |
|
| ISBN (electronic) |
|
| Event | 2017 IEEE Conference on Computer Vision and Pattern Recognition |
| Pages (from-to) | 2087-2096 |
| Publisher | Piscataway, NJ: IEEE |
| Organisations |
|
| Abstract |
Event detection in unconstrained videos is conceived as a content-based video retrieval with two modalities: textual and visual. Given a text describing a novel event, the goal is to rank related videos accordingly. This task is zero-exemplar, no video examples are given to the novel event.
Related works train a bank of concept detectors on external data sources. These detectors predict confidence scores for test videos, which are ranked and retrieved accordingly. In contrast, we learn a joint space in which the visual and textual representations are embedded. The space casts a novel event as a probability of pre-defined events. Also, it learns to measure the distance between an event and its related videos. Our model is trained end-to-end on publicly available EventNet. When applied to TRECVID Multimedia Event Detection dataset, it outperforms the state-of-the-art by a considerable margin. |
| Document type | Conference contribution |
| Language | English |
| Published at | https://doi.org/10.1109/CVPR.2017.225 |
| Published at | https://arxiv.org/abs/1705.02148 |
| Other links | https://ivi.fnwi.uva.nl/isis/publications/2017/HusseinCVPR2017 |
| Downloads |
Hussein_Unified_Embedding_and_CVPR_2017_paper
(Accepted author manuscript)
|
| Permalink to this page | |
