Abstract
AbstractZero-shot learning (ZSL) aims at recognizing classes for which no visual sample is available at training time. To address this issue, one can rely on a semantic description of each class. A typical ZSL model learns a mapping between the visual samples of seen classes and the corresponding semantic descriptions, in order to do the same on unseen classes at test time. State of the art approaches rely on generative models that synthesize visual features from the prototype of a class, such that a classifier can then be learned in a supervised manner. However, these approaches are usually biased towards seen classes whose visual instances are the only one that can be matched to a given class prototype. We propose a regularization method that can be applied to any conditional generative-based ZSL method, by leveraging only the semantic class prototypes. It learns to synthesize discriminative features for possible semantic description that are not available at training time, that is the unseen ones. The approach is evaluated for ZSL and GZSL on four datasets commonly used in the literature, either in inductive or transductive settings, with results on-par or above state of the art approaches. The code is available at https://github.com/hanouticelina/lsa-zsl.
Funder
Horizon 2020 Framework Programme
Agence NAtionale de la Recherche
Publisher
Springer Science and Business Media LLC
Subject
Computer Networks and Communications,Hardware and Architecture,Media Technology,Software
Reference34 articles.
1. Adjali O, Besancon R, Ferret O et al (2020) Multimodal entity linking for tweets. In: European conference on information retrieval, Lisbon, Portugal
2. Arjovsky M, Chintala S, Bottou L (2017) Wasserstein generative adversarial networks. In: Proceedings of the 34th international conference on machine learning. JMLR.org, ICML’17, vol 70, pp 214–223
3. Arora G, Verma VK, Mishra A et al (2017) Generalized zero-shot learning via synthesized examples. CoRR: http://arxiv.org/1712.03878
4. Bucher M, Herbin S, Jurie F (2017) Generating Visual Representations for Zero-Shot Classification. In: International conference on computer vision (ICCV) workshops: TASK-CV: transferring and adapting source knowledge in computer vision, Venise, Italy
5. Chami I, Tamaazousti Y, Le Borgne H (2017) Amecon: abstract meta-concept features for text-illustration. In: ACM international conference on multimedia retrieval. ICMR, Bucharest
Cited by
1 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献