Annotating images by harnessing worldwide user-tagged photos

Authors
Publication date 2009
Book title 2009 IEEE International Conference on Acoustics, Speech, and Signal Processing: Proceedings: April 19—24, 2009, Taipei International Convention Center, Taipei, Taiwan
ISBN
  • 9781424423538
Event 2009 IEEE International Conference on Acoustics, Speech, and Signal Processing (ICASSP 2009), Taipei, Taiwan
Pages (from-to) 3717-3720
Publisher Piscataway, NJ: IEEE
Organisations
  • Faculty of Science (FNWI) - Informatics Institute (IVI)
Abstract
Automatic image tagging is important yet challenging due to the semantic gap and the lack of learning examples to model a tag's visual diversity. Meanwhile, social user tagging is creating rich multimedia content on the Web. In this paper, we propose to combine the two tagging approaches in a search-based framework. For an unlabeled image, we first retrieve its visual neighbors from a large user-tagged image database. We then select relevant tags from the result images to annotate the unlabeled image. To tackle the unreliability and sparsity of user tagging, we introduce a joint-modality tag relevance estimation method which efficiently addresses both textual and visual clues. Experiments on 1.5 million Flickr photos and 10 000 Corel images verify the proposed method.
Document type Conference contribution
Published at https://doi.org/10.1109/ICASSP.2009.4960434
Permalink to this page
Back