A simple and accurate syntax-agnostic neural model for dependency-based semantic role labeling

Open Access
Authors
Publication date 2017
Host editors
  • R. Levy
  • L. Specia
Book title The 21st Conference on Computational Natural Language Learning
Book subtitle Proceedings of the Conference : CoNNL 2017 : August 2-august 4, 2017, Vancouver, Canada
ISBN (electronic)
  • 9781945626548
Event 21st Conference on Computational Natural Language Learning, CoNLL 2017
Pages (from-to) 411-420
Number of pages 10
Publisher Stroudsburg, PA: Association for Computational Linguistics
Organisations
  • Interfacultary Research - Institute for Logic, Language and Computation (ILLC)
Abstract

We introduce a simple and accurate neural model for dependency-based semantic role labeling. Our model predicts predicate-argument dependencies relying on states of a bidirectional LSTM encoder. The semantic role labeler achieves competitive performance on English, even without any kind of syntactic information and only using local inference. However, when automatically predicted part-of-speech tags are provided as input, it substantially outperforms all previous local models and approaches the best reported results on the English CoNLL-2009 dataset. We also consider Chinese, Czech and Spanish where our approach also achieves competitive results. Syntactic parsers are unreliable on out-of-domain data, so standard (i.e., syntactically-informed) SRL models are hindered when tested in this setting. Our syntax-agnostic model appears more robust, resulting in the best reported results on standard out-of-domain test sets.

Document type Conference contribution
Language English
Published at https://doi.org/10.18653/v1/k17-1041
Other links https://github.com/diegma/neural-dep-srl https://www.scopus.com/pages/publications/85062917879
Downloads
K17-1041 (Final published version)
Permalink to this page
Back