Share your Model instead of your Data: Privacy Preserving Mimic Learning for Ranking

Open Access
Authors
Publication date 2017
Book title Neu-IR: Workshop on Neural Information Retrieval
Book subtitle accepted papers
Event SIGIR 2017 Workshop on Neural Information Retrieval (Neu-IR'17)
Number of pages 5
Publisher Ithaca, NY: ArXiv
Organisations
  • Faculty of Humanities (FGw)
  • Faculty of Science (FNWI) - Informatics Institute (IVI)
  • Interfacultary Research - Institute for Logic, Language and Computation (ILLC)
Abstract
Deep neural networks have become a primary tool for solving problems in many fields. They are also used for addressing information retrieval problems and show strong performance in several tasks. Training these models requires large, representative datasets and for most IR tasks, such data contains sensitive information from users. Privacy and confidentiality concerns prevent many data owners from sharing the data, thus today the research community can only benefit from research on large-scale datasets in a limited manner. In this paper, we discuss privacy preserving mimic learning, i.e., using predictions from a privacy preserving trained model instead of labels from the original sensitive training data as a supervision signal. We present the results of preliminary experiments in which we apply the idea of mimic learning and privacy preserving mimic learning for the task of document re-ranking as one of the core IR tasks. This research is a step toward laying the ground for enabling researchers from data-rich environments to share knowledge learned from actual users' data, which should facilitate research collaborations.
Document type Conference contribution
Note Workshop at SIGIR 2017. All accepted papers published on arXiv.org.
Language English
Published at https://arxiv.org/abs/1707.07605
Other links https://neu-ir.weebly.com/
Downloads
1707.07605 (Accepted author manuscript)
Permalink to this page
Back