Image2Emoji: Zero-shot Emoji Prediction for Visual Media
| Authors | |
|---|---|
| Publication date | 2015 |
| Book title | MM'15 |
| Book subtitle | proceedings of the 2015 ACM Multimedia Conference: October 26-30, 2015, Brisbane, Australia |
| ISBN (electronic) |
|
| Event | 2015 ACM International Conference on Multimedia |
| Pages (from-to) | 1311-1314 |
| Publisher | New York, NY: Association for Computing Machinery |
| Organisations |
|
| Abstract |
We present Image2Emoji, a multi-modal approach for generating emoji labels for an image in a zero-shot manner. Different from existing zero-shot image-to-text approaches, we exploit both image and textual media to learn a semantic embedding for the new task of emoji prediction. We propose that the widespread adoption of emoji suggests a semantic universality which is well-suited for interaction with visual media. We quantify the efficacy of our proposed model on the MSCOCO dataset, and demonstrate the value of visual, textual and multi-modal prediction of emoji. We conclude the paper with three examples of the application potential of emoji in the context of multimedia retrieval.
|
| Document type | Conference contribution |
| Language | English |
| Published at | https://doi.org/10.1145/2733373.2806335 |
| Downloads |
p1311-cappallo
(Final published version)
|
| Permalink to this page | |
