Author:
Abhishek Kumar,Brown Colin J.,Hamarneh Ghassan
Abstract
AbstractModern deep learning training procedures rely on model regularization techniques such as data augmentation methods, which generate training samples that increase the diversity of data and richness of label information. A popular recent method, mixup, uses convex combinations of pairs of original samples to generate new samples. However, as we show in our experiments, mixup can produce undesirable synthetic samples, where the data is sampled off the manifold and can contain incorrect labels. We propose $$\zeta $$
ζ
-mixup, a generalization of mixup with provably and demonstrably desirable properties that allows convex combinations of $${T} \ge 2$$
T
≥
2
samples, leading to more realistic and diverse outputs that incorporate information from $${T}$$
T
original samples by using a p-series interpolant. We show that, compared to mixup, $$\zeta $$
ζ
-mixup better preserves the intrinsic dimensionality of the original datasets, which is a desirable property for training generalizable models. Furthermore, we show that our implementation of $$\zeta $$
ζ
-mixup is faster than mixup, and extensive evaluation on controlled synthetic and 26 diverse real-world natural and medical image classification datasets shows that $$\zeta $$
ζ
-mixup outperforms mixup, CutMix, and traditional data augmentation techniques. The code will be released at https://github.com/kakumarabhishek/zeta-mixup.
Funder
Natural Sciences and Engineering Research Council of Canada
British Columbia Cancer Foundation
Collaborative Health Research Projects
Simon Fraser University
Publisher
Springer Science and Business Media LLC
Reference110 articles.
1. Schmidhuber J. Deep learning in neural networks: an overview. Neural Netw. 2015;61:85–117.
2. LeCun Y, Bengio Y, Hinton G. Deep learning. Nature. 2015;521(7553):436–44.
3. Alom MZ, Taha TM, Yakopcic C, Westberg S, Sidike P, Nasrin MS, Van Esesn BC, Awwal AAS, Asari VK. The history began from AlexNet: A comprehensive survey on deep learning approaches. arXiv preprint arXiv:1803.01164. 2018.
4. Wu Z, Pan S, Chen F, Long G, Zhang C, Philip SY. A comprehensive survey on graph neural networks. IEEE Trans Neural Netw Learn Syst. 2020;32(1):4–24.
5. Deng J, Dong W, Socher R, Li L-J, Li K, Fei-Fei L. ImageNet: a large-scale hierarchical image database. In: 2009 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2009; pp. 248–255. IEEE.
Cited by
1 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献