@inproceedings{abfed7a4821545f385aff451ac7dcabe,
title = "Bi-SAN-CAP: Bi-Directional Self-Attention for Image Captioning",
abstract = "In a typical image captioning pipeline, a Convolutional Neural Network (CNN) is used as the image encoder and Long Short-Term Memory (LSTM) as the language decoder. LSTM with attention mechanism has shown remarkable performance on sequential data including image captioning. LSTM can retain long-range dependency of sequential data. However, it is hard to parallelize the computations of LSTM because of its inherent sequential characteristics. In order to address this issue, recent works have shown benefits in using self-attention, which is highly parallelizable without requiring any temporal dependencies. However, existing techniques apply attention only in one direction to compute the context of the words. We propose an attention mechanism called Bi-directional Self-Attention (Bi-SAN) for image captioning. It computes attention both in forward and backward directions. It achieves high performance comparable to state-of-the-art methods.",
keywords = "Bi-directional Self-Attention, Deep Learning, Image Captioning, Self-Attention",
author = "Hossain, {Md Zakir} and Ferdous Sohel and Shiratuddin, {Mohd Fairuz} and Hamid Laga and Mohammed Bennamoun",
year = "2019",
month = dec,
day = "1",
doi = "10.1109/DICTA47822.2019.8946003",
language = "English",
series = "2019 Digital Image Computing: Techniques and Applications, DICTA 2019",
publisher = "IEEE, Institute of Electrical and Electronics Engineers",
booktitle = "2019 Digital Image Computing",
address = "United States",
note = "2019 International Conference on Digital Image Computing: Techniques and Applications, DICTA 2019 ; Conference date: 02-12-2019 Through 04-12-2019",
}