Prototype-Based Support Example Miner and Triplet Loss for Deep Metric Learning
-
Published:2023-08-02
Issue:15
Volume:12
Page:3315
-
ISSN:2079-9292
-
Container-title:Electronics
-
language:en
-
Short-container-title:Electronics
Author:
Yang Shan1, Zhang Yongfei123, Zhao Qinghua4, Pu Yanglin1, Yang Hangyuan1
Affiliation:
1. Beijing Key Laboratory of Digital Media, School of Computer Science and Engineering, Beihang University, Beijing 100191, China 2. State Key Laboratory of Virtual Reality Technology and Systems, Beihang University, Beijing 100191, China 3. Pengcheng Laboratory, Shenzhen 518055, China 4. State Key Laboratory of Software Development Environment, School of Computer Science and Engineering, Beihang University, Beijing 100191, China
Abstract
Deep metric learning aims to learn a mapping function that projects input data into a high-dimensional embedding space, facilitating the clustering of similar data points while ensuring dissimilar ones are far apart. The most recent studies focus on designing a batch sampler and mining online triplets to achieve this purpose. Conventionally, hard negative mining schemes serve as the preferred batch sampler. However, most hard negative mining schemes search for hard examples in randomly selected mini-batches at each epoch, which often results in less-optimal hard examples and thus sub-optimal performances. Furthermore, Triplet Loss is commonly adopted to perform online triplet mining by pulling the hard positives close to and pushing the negatives away from the anchor. However, when the anchor in a triplet is an outlier, the positive example will be pulled away from the centroid of the cluster, thus resulting in a loose cluster and inferior performance. To address the above challenges, we propose the Prototype-based Support Example Miner (pSEM) and Triplet Loss (pTriplet Loss). First, we present a support example miner designed to mine the support classes on the prototype-based nearest neighbor graph of classes. Following this, we locate the support examples by searching for instances at the intersection between clusters of these support classes. Second, we develop a variant of Triplet Loss, referred to as a Prototype-based Triplet Loss. In our approach, a dynamically updated prototype is used to rectify outlier anchors, thus reducing their detrimental effects and facilitating a more robust formulation for Triplet Loss. Extensive experiments on typical Computer Vision (CV) and Natural Language Processing (NLP) tasks, namely person re-identification and few-shot relation extraction, demonstrated the effectiveness and generalizability of the proposed scheme, which consistently outperforms the state-of-the-art models.
Funder
National Natural Science Foundation of China the Fundamental Research Funds for the Central Universities
Subject
Electrical and Electronic Engineering,Computer Networks and Communications,Hardware and Architecture,Signal Processing,Control and Systems Engineering
Reference69 articles.
1. Deep Learning for Person Re-Identification: A Survey and Outlook;Ye;IEEE Trans. Pattern Anal. Mach. Intell.,2020 2. Liao, S., and Shao, L. (2021, January 18–24). Graph Sampling Based Deep Metric Learning for Generalizable Person Re-Identification. Proceedings of the 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA. 3. Harwood, B., Kumar BG, V., Carneiro, G., Reid, I., and Drummond, T. (2017, January 22–29). Smart mining for deep metric learning. Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy. 4. Manmatha, R., Wu, C., Smola, A., and Krähenbühl, P. (2017, January 22–29). Sampling Matters in Deep Embedding Learning. Proceedings of the 2017 IEEE International Conference on Computer Vision (ICCV), Venice, Italy. 5. Schroff, F., Kalenichenko, D., and Philbin, J. (2015, January 7–12). Facenet: A unified embedding for face recognition and clustering. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA.
Cited by
3 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献
|
|