Recursive Neural Networks with Bottlenecks Diagnose (Non-)Compositionality

Open Access
Authors
Publication date 2022
Host editors
  • Y. Goldberg
  • Z. Kozareva
  • Y. Zhang
Book title Findings of the Association for Computational Linguistics: EMNLP 2022
Book subtitle Conference on Empirical Methods in Natural Language Processing (EMNLP), Abu Dhabi, United Arab Emirates, 7-11 December 2022
Event The 2022 Conference on Empirical Methods in Natural Language Processing
Pages (from-to) 4361–4378
Publisher Stroudsburg, PA: Association for Computational Linguistics
Organisations
  • Interfacultary Research - Institute for Logic, Language and Computation (ILLC)
Abstract

A recent line of work in NLP focuses on the (dis)ability of models to generalise compositionally for artificial languages. However, when considering natural language tasks, the data involved is not strictly, or locally, compositional. Quantifying the compositionality of data is a challenging task, which has been investigated primarily for short utterances. We use recursive neural models (Tree-LSTMs) with bottlenecks that limit the transfer of information between nodes. We illustrate that comparing data's representations in models with and without the bottleneck can be used to produce a compositionality metric. The procedure is applied to the evaluation of arithmetic expressions using synthetic data, and sentiment classification using natural language data. We demonstrate that compression through a bottleneck impacts non-compositional examples disproportionately and then use the bottleneck compositionality metric (BCM) to distinguish compositional from non-compositional samples, yielding a compositionality ranking over a dataset.

Document type Conference contribution
Note With supplementary video
Language English
Published at https://doi.org/10.18653/v1/2022.findings-emnlp.320
Other links https://www.scopus.com/pages/publications/85149895244
Downloads
2022.findings-emnlp.320 (Final published version)
Supplementary materials
Permalink to this page
Back