self-critical training
fecc3a370a23d13b1cf91ac3c1e1ca92-AuthorFeedback.pdf
R1: Cut down on some sections (3.2.1, 3.2.2 and 3.2.5) to spare space for the qualitative examples. We will revise our paper according to the suggestion in the final version. We added experiments on MS-COCO and Flicker30k using single-head attention, Table 1. R2: The base attention model performs better than up-down and GCN-LSTM. In addition, our experimental results showed that increasing the number of min.
Reviews: Adaptively Aligned Image Captioning via Adaptive Attention Time
Although the two techniques have been well explored individually, this is the first work combining it for attention for image captioning. This should make reproducing the results easier. The base attention model already is doing much better than up-down attention and recent methods like GCN-LSTM and so it's not clear where the gains are coming from. It'd be good to see AAT applied to traditional single-head attention instead of multi-head attention to convincingly show that AAT helps. For instance, how does the attention time steps vary with word position in the caption?
Reviews: Image Captioning: Transforming Objects into Words
Summary - The proposed approach to image captioning extends two prior works, object-based Up-Down method of [2] and Transformer of [22] (already used for image captioning in [21]). Specifically, the authors integrate spatial relations between objects in the captioning Transformer model, proposing the Object Relation Transformer. The modification amounts to introducing an object relation module [9] into the encoding layer of the Transformer model. Tests of statistical significance show that the proposed model outperforms the standard Transformer in terms of CIDEr-D, BLEU-1 and ROUGE-L, while SPICE-attribute breakdown shows improvement for Relation and Count categories. Qualitative results include examples where Object Relation Transformer leads to more correct spatial Relation and Count predictions.