Learning Expressive Meta-Representations with Mixture of Expert Neural Processes

Open Access
Authors
Publication date 2023
Host editors
  • S. Koyejo
  • S. Mohamed
  • A. Agarwal
  • D. Belgrave
  • K. Cho
  • A. Oh
Book title 36th Conference on Neural Information Processing Systems (NeurIPS 2022)
Book subtitle New Orleans, Louisiana, USA, 28 November-9 December 2022
ISBN
  • 9781713871088
ISBN (electronic)
  • 9781713873129
Series Advances in Neural Information Processing Systems
Event Thirty-sixth Conference on Neural Information Processing Systems
Volume | Issue number 34
Pages (from-to) 26242-26255
Publisher San Diego, CA: Neural Information Processing Systems Foundation
Organisations
  • Faculty of Science (FNWI) - Informatics Institute (IVI)
Abstract
Neural processes (NPs) formulate exchangeable stochastic processes and are promising models for meta learning that do not require gradient updates during the testing phase. However, most NP variants place a strong emphasis on a global latent variable. This weakens the approximation power and restricts the scope of applications using NP variants, especially when data generative processes are complicated.To resolve these issues, we propose to combine the Mixture of Expert models with Neural Processes to develop more expressive exchangeable stochastic processes, referred to as Mixture of Expert Neural Processes (MoE-NPs). Then we apply MoE-NPs to both few-shot supervised learning and meta reinforcement learning tasks. Empirical results demonstrate MoE-NPs' strong generalization capability to unseen tasks in these benchmarks.
Document type Conference contribution
Note With supplemental file
Language English
Published at https://papers.nips.cc/paper_files/paper/2022/hash/a815fe7cad6af20a6c118f2072a881d2-Abstract-Conference.html https://openreview.net/forum?id=ju38DG3sbg6
Other links https://www.proceedings.com/68431.html
Downloads
Supplementary materials
Permalink to this page
Back