Abstract
AbstractOver-fitting is a significant threat to the integrity and reliability of deep neural networks with generous parameters. One problem is that the model learned more specific features than general features in the training process. To solve the problem, we propose an adversarial training method to assist the model in strengthening general representation learning. In this method, we make a classification model as a generator G and introduce an unsupervised discriminator D to distinguish the hidden feature of the classification model from real images to limit their spatial distance. Notably, the D will fall into the trap of a perfect discriminator resulting in the gradient of confrontation loss of 0 after overtraining. To avoid this situation, we train the D with a probability $$P_{c}$$
P
c
. Our proposed method is easy to incorporate into existing frameworks. It has been evaluated under various network architectures over different fields of datasets. Experiments show that this method, under low computational cost, outperforms the benchmark by 1.5–2 points on different datasets. For semantic segmentation on VOC, our proposed method achieves 2.2 points higher mAP.
Funder
the Science and Technology Program of Quzhou
the Science and Technology Program of Zhejiang
Publisher
Springer Science and Business Media LLC
Subject
Computational Mathematics,General Computer Science
Reference33 articles.
1. Alexey, D., Philipp, F., et al.: Discriminative unsupervised feature learning with exemplar convolutional neural networks. IEEE Trans. Pattern Anal. Mach. Intell. 38(9), 1734–1747 (2016). https://doi.org/10.1109/TPAMI.2015.2496141
2. Arjovsky, M., Bottou, L.: Towards principled methods for training generative adversarial networks. In: International Conference on Learning Representations, Toulon, France (2017)
3. Cook, J.A., Ranstam, J.: Overfitting. BJS Stat. Nugget 103(13), 1814 (2016). https://doi.org/10.1002/bjs.10244
4. Dai, Z., Yang, Z., et al.: Good semi-supervised learning that requires a bad gan. Adv. Neural Inf. Process. Syst. 30, 6510–6520 (2017)
5. Dan, H., Thomas, D.: Benchmarking neural network robustness to common corruptions and perturbations. In: International Conference on Learning Representations (ICLR) (2019)