Novel Object Captioning with Semantic Match from External Knowledge
-
Published:2023-07-04
Issue:13
Volume:13
Page:7868
-
ISSN:2076-3417
-
Container-title:Applied Sciences
-
language:en
-
Short-container-title:Applied Sciences
Author:
Du Sen1ORCID, Zhu Hong1, Lin Guangfeng2ORCID, Wang Dong1, Shi Jing1
Affiliation:
1. School of Automation and Information Engineering, Xi’an University of Technology, Xi’an 710048, China 2. School of Printing, Packaging and Digital Media, Xi’an University of Technology, Xi’an 710054, China
Abstract
Automatically describing the content of an image is a challenging task that is on the edge between natural language and computer vision. The current image caption models can describe the objects that are frequently seen in the training set very well, but they fail to describe the novel objects that are rarely seen or never seen in the training set. Despite describing novel objects being important for practical applications, only a few works investigate this issue. Furthermore, those works only investigate rarely seen objects, but ignore the never-seen objects. Meanwhile, the number of never-seen objects is more than the number of frequently seen and rarely seen objects. In this paper, we propose two blocks that incorporate external knowledge into the captioning model to solve this issue. Initially, in the encoding phase, the Semi-Fixed Word Embedding block is an improvement for the word embedding layer that enables the captioning model to understand the meaning of the arbitrary visual words rather than a fixed number of words. Furthermore, the Candidate Sentences Selection block chooses candidate sentences by semantic matching rather than probability, avoiding the influence of never-seen words. In experiments, we qualitatively analyze the proposed blocks and quantitatively evaluate several captioning models with the proposed blocks on the Nocaps dataset. The experimental results show the effectiveness of the proposed blocks for novel objects, especially when describing never-seen objects, CIDEr and SPICE improved by 13.1% and 12.0%, respectively.
Funder
NSFC the Research and development of manufacturing information system platform supporting product lifecycle management the Key Research and Development Program of Shaanxi Doctoral Research Fund of Xi’an University of Technology, China Xi’an Science and Technology Foundation Natural Science Foundation of Shaanxi Province
Subject
Fluid Flow and Transfer Processes,Computer Science Applications,Process Chemistry and Technology,General Engineering,Instrumentation,General Materials Science
Reference55 articles.
1. Hubert: Self-supervised speech representation learning by masked prediction of hidden units;Hsu;IEEE/ACM Trans. Audio Speech Lang. Process.,2021 2. Kenton, J.D.M.W.C., and Toutanova, L.K. (2019, January 2–7). Bert: Pre-training of deep bidirectional transformers for language understanding. Proceedings of the NAACL-HLT, Minneapolis, MN, USA. 3. Language models are few-shot learners;Brown;Adv. Neural Inf. Process. Syst.,2020 4. Pan, Y., Yao, T., Li, Y., and Mei, T. (2020, January 14–19). X-linear attention networks for image captioning. Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA. 5. A comprehensive survey of deep learning for image captioning;Hossain;ACM Comput. Surv. (CsUR),2019
|
|