Big Generalizations with Small Data: Exploring the Role of Training Samples in Learning Adjectives of Size
| Authors | |
|---|---|
| Publication date | 2019 |
| Host editors |
|
| Book title | Beyond Vision and LANguage: inTEgrating Real-world kNowledge (LANTERN) |
| Book subtitle | EMNLP-IJCNLP 2019 : proceedings of the workshop : November 3, 2019, Hong Kong, China |
| ISBN (electronic) |
|
| Event | 1st Workshop on Beyond Vision and LANguage: inTEgrating Real-World kNowledge, LANTERN@EMNLP-IJCNLP 2019 |
| Pages (from-to) | 18-23 |
| Publisher | Stroudsburg, PA: The Association for Computational Linguistics |
| Organisations |
|
| Abstract |
In this paper, we experiment with a recently proposed visual reasoning task dealing with quantities – modeling the multimodal, contextually-dependent meaning of size adjectives (‘big’, ‘small’) – and explore the impact of varying the training data on the learning behavior of a state-of-art system. In previous work, models have been shown to fail in generalizing to unseen adjective-noun combinations. Here, we investigate whether, and to what extent, seeing some of these cases during training helps a model understand the rule subtending the task, i.e., that being big implies being not small, and vice versa. We show that relatively few examples are enough to understand this relationship, and that developing a specific, mutually exclusive representation of size adjectives is beneficial to the task.
|
| Document type | Conference contribution |
| Language | English |
| Related dataset | MALeViC |
| Published at | https://doi.org/10.18653/v1/D19-6403 |
| Downloads |
D19-6403
(Final published version)
|
| Permalink to this page | |
