Abstract
Image captioning is a challenging vision-to-language task, which has garnered a lot of attention over the past decade. The introduction of Encoder-Decoder based architectures expedited the research in this area and provided the backbone of the most recent systems. Moreover, leveraging relationships between objects for holistic scene understanding, which in turn improves captioning, has recently sparked interest among researchers. Our proposed model encodes the spatial and semantic proximity of object pairs into linguistically-aware relationship embeddings. Moreover, it captures the global semantics of the image using NASNet. This way, true semantic relations that are not apparent in visual content of an image can be learned, such that the decoder can attend to the most relevant object relations and visual features to generate more semantically-meaningful captions. Our experiments highlight the usefulness of linguistically-aware object relations as well as NASNet visual features for image captioning.
Original language | English |
---|---|
Title of host publication | 2020 35th International Conference on Image and Vision Computing New Zealand (IVCNZ) |
Publisher | IEEE, Institute of Electrical and Electronics Engineers |
Pages | 1-6 |
Number of pages | 6 |
ISBN (Electronic) | 9781728185798 |
DOIs | |
Publication status | Published - 25 Nov 2020 |
Event | 35th International Conference on Image and Vision Computing New Zealand, IVCNZ - Wellington, New Zealand Duration: 25 Nov 2020 → 27 Nov 2020 Conference number: 35 |
Publication series
Name | International Conference Image and Vision Computing New Zealand |
---|---|
Volume | 2020-November |
ISSN (Print) | 2151-2191 |
ISSN (Electronic) | 2151-2205 |
Conference
Conference | 35th International Conference on Image and Vision Computing New Zealand, IVCNZ |
---|---|
Abbreviated title | IVCNZ 2020 |
Country/Territory | New Zealand |
City | Wellington |
Period | 25/11/20 → 27/11/20 |