Translating visual art into music
| Authors |
|
|---|---|
| Publication date | 2019 |
| Book title | 2019 International Conference on Computer Vision, Workshops |
| Book subtitle | proceedings : 27 October-2 November 2019, Seoul, Korea |
| ISBN |
|
| ISBN (electronic) |
|
| Event | 17th IEEE/CVF International Conference on Computer Vision Workshop, ICCVW 2019 |
| Pages (from-to) | 3117-3120 |
| Number of pages | 4 |
| Publisher | Los Alamitos, California: IEEE Computer Society |
| Organisations |
|
| Abstract |
The Synesthetic Variational Autoencoder (SynVAE) introduced in this research is able to learn a consistent mapping between visual and auditive sensory modalities in the absence of paired datasets. A quantitative evaluation on MNIST as well as the Behance Artistic Media dataset (BAM) shows that SynVAE is capable of retaining sufficient information content during the translation while maintaining cross-modal latent space consistency. In a qualitative evaluation trial, human evaluators were furthermore able to match musical samples with the images which generated them with accuracies of up to 73%. |
| Document type | Conference contribution |
| Language | English |
| Published at | https://doi.org/10.1109/ICCVW.2019.00378 |
| Other links | http://www.proceedings.com/52964.html https://www.scopus.com/pages/publications/85082484301 |
| Permalink to this page | |
