A syntactic language model based on incremental CCG parsing
| Authors |
|
|---|---|
| Publication date | 2008 |
| Book title | SLT 2008: 2008 IEEE Workshop on Spoken Language Technology: Proceedings |
| ISBN |
|
| Event | 2008 IEEE Workshop on Spoken Language Technology (SLT 2008), Goa, India |
| Pages (from-to) | 205-208 |
| Publisher | IEEE |
| Organisations |
|
| Abstract |
Syntactically-enriched language models (parsers) constitute a promising component in applications such as machine translation and speech-recognition. To maintain a useful level of accuracy, existing parsers are non-incremental and must span a combinatorially growing space of possible structures as every input word is processed. This prohibits their incorporation into standard linear-time decoders. In this paper, we present an incremental, linear-time dependency parser based on Combinatory Categorial Grammar (CCG) and classification techniques. We devise a deterministic transform of CCG-bank canonical derivations into incremental ones, and train our parser on this data. We discover that a cascaded, incremental version provides an appealing balance between efficiency and accuracy.
|
| Document type | Conference contribution |
| Published at | https://doi.org/10.1109/SLT.2008.4777876 |
| Permalink to this page | |