The search functionality is under construction.

IEICE TRANSACTIONS on Information

Effectively Utilizing the Category Labels for Image Captioning

Junlong FENG, Jianping ZHAO

  • Full Text Views

    2

  • Cite this

Summary :

As a further investigation of the image captioning task, some works extended the vision-text dataset for specific subtasks, such as the stylized caption generating. The corpus in such dataset is usually composed of obvious sentiment-bearing words. While, in some special cases, the captions are classified depending on image category. This will result in a latent problem: the generated sentences are in close semantic meaning but belong to different or even opposite categories. It is a worthy issue to explore an effective way to utilize the image category label to boost the caption difference. Therefore, we proposed an image captioning network with the label control mechanism (LCNET) in this paper. First, to further improve the caption difference, LCNET employs a semantic enhancement module to provide the decoder with global semantic vectors. Then, through the proposed label control LSTM, LCNET can dynamically modulate the caption generation depending on the image category labels. Finally, the decoder integrates the spatial image features with global semantic vectors to output the caption. Using all the standard evaluation metrics shows that our model outperforms the compared models. Caption analysis demonstrates our approach can improve the performance of semantic representation. Compared with other label control mechanisms, our model is capable of boosting the caption difference according to the labels and keeping a better consistent with image content as well.

Publication
IEICE TRANSACTIONS on Information Vol.E106-D No.5 pp.617-624
Publication Date
2023/05/01
Publicized
2021/12/13
Online ISSN
1745-1361
DOI
10.1587/transinf.2022DLP0013
Type of Manuscript
Special Section PAPER (Special Section on Deep Learning Technologies: Architecture, Optimization, Techniques, and Applications)
Category
Core Methods

Authors

Junlong FENG
  Changchun University of Science and Technology
Jianping ZHAO
  Changchun University of Science and Technology

Keyword