Assessing incrementality in sequence-to-sequence models

Open Access
Authors
Publication date 2019
Host editors
  • I. Augenstein
  • S. Gella
  • S. Ruder
  • K. Kann
  • B. Can
  • J. Welbl
  • A. Conneau
  • X. Ren
  • M. Rei
Book title The 4th Workshop on Representation Learning for NLP (RepL4NLP-2019)
Book subtitle ACL 2019 : proceedings of the workshop : August 2, 2019, Florence, Italy
ISBN (electronic)
  • 9781950737352
Event 4th Workshop on Representation Learning for NLP (RepL4NLP)
Pages (from-to) 209–217
Publisher Stroudsburg, PA: The Association for Computational Linguistics
Organisations
  • Interfacultary Research - Institute for Logic, Language and Computation (ILLC)
Abstract
Since their inception, encoder-decoder models have successfully been applied to a wide array of problems in computational linguistics. The most recent successes are predominantly due to the use of different variations of attention mechanisms, but their cognitive plausibility is questionable. In particular, because past representations can be revisited at any point in time, attention-centric methods seem to lack an incentive to build up incrementally more informative representations of incoming sentences. This way of processing stands in stark contrast with the way in which humans are believed to process language: continuously and rapidly integrating new information as it is encountered. In this work, we propose three novel metrics to assess the behavior of RNNs with and without an attention mechanism and identify key differences in the way the different model types process sentences.
Document type Conference contribution
Language English
Published at https://doi.org/10.18653/v1/W19-4324
Downloads
W19-4324 (Final published version)
Permalink to this page
Back