Image Caption Generation using Object Attention Mechanism 


Vol. 46,  No. 4, pp. 369-375, Apr.  2019
10.5626/JOK.2019.46.4.369


PDF

  Abstract

Explosive increases in image data have led studies investigating the role of image caption generation in image expression of natural language. The current technologies for generating Korean image captions contain errors associated with object concurrence attributed to dataset translation from English datasets. In this paper, we propose a model of image caption generation employing attention as a new loss function using the extracted nouns of image references. The proposed method displayed BLEU1 0.686, BLEU2 0.557, BLEU3 0.456, BLEU4 0.372, which proves that the proposed model facilitates the resolution of high-frequency word-pair errors. We also showed that it enhances the performance compared with previous studies and reduces redundancies in the sentences. As a result, the proposed method can be used to generate a caption corpus effectively.


  Statistics
Cumulative Counts from November, 2022
Multiple requests among the same browser session are counted as one view. If you mouse over a chart, the values of data points will be shown.


  Cite this article

[IEEE Style]

D. Park and J. Cha, "Image Caption Generation using Object Attention Mechanism," Journal of KIISE, JOK, vol. 46, no. 4, pp. 369-375, 2019. DOI: 10.5626/JOK.2019.46.4.369.


[ACM Style]

Da-Sol Park and Jeong-Won Cha. 2019. Image Caption Generation using Object Attention Mechanism. Journal of KIISE, JOK, 46, 4, (2019), 369-375. DOI: 10.5626/JOK.2019.46.4.369.


[KCI Style]

박다솔, 차정원, "객체 Attention을 이용한 이미지 캡션 생성," 한국정보과학회 논문지, 제46권, 제4호, 369~375쪽, 2019. DOI: 10.5626/JOK.2019.46.4.369.


[Endnote/Zotero/Mendeley (RIS)]  Download


[BibTeX]  Download



Search




Journal of KIISE

  • ISSN : 2383-630X(Print)
  • ISSN : 2383-6296(Electronic)
  • KCI Accredited Journal

Editorial Office

  • Tel. +82-2-588-9240
  • Fax. +82-2-521-1352
  • E-mail. chwoo@kiise.or.kr