Analysing Neural Language Models: Contextual Decomposition Reveals Default Reasoning in Number and Gender Assignment

Open Access
Authors
Publication date 2019
Host editors
  • M. Bansal
  • A. Villavicencio
Book title The 23rd Conference on Computational Natural Language Learning
Book subtitle CoNLL 2019 : proceedings of the conference : November 3-4, 2019, Hong Kong, China
ISBN (electronic)
  • 9781950737727
Event 23rd Conference on Computational Natural Language Learning
Pages (from-to) 1-11
Publisher Stroudsburg, PA: The Association for Computational Linguistics
Organisations
  • Interfacultary Research - Institute for Logic, Language and Computation (ILLC)
Abstract
Extensive research has recently shown that recurrent neural language models are able to process a wide range of grammatical phenomena. How these models are able to perform these remarkable feats so well, however, is still an open question. To gain more insight into what information LSTMs base their decisions on, we propose a generalisation of Contextual Decomposition (GCD). In particular, this setup enables us to accurately distil which part of a prediction stems from semantic heuristics, which part truly emanates from syntactic cues and which part arise from the model biases themselves instead. We investigate this technique on tasks pertaining to syntactic agreement and co-reference resolution and discover that the model strongly relies on a default reasoning effect to perform these tasks.
Document type Conference contribution
Note With attachment.
Language English
Published at https://doi.org/10.18653/v1/K19-1001
Downloads
K19-1001 (Final published version)
Supplementary materials
Permalink to this page
Back