DRD-GAN: A Novel Distributed Conditional Wasserstein Deep Convolutional Relativistic Discriminator GAN with Improved Convergence

Author:

Roy Arunava1ORCID,Dasgupta Dipankar1ORCID

Affiliation:

1. The University of Memphis, USA

Abstract

Generative Adversarial Network (GAN) exhibited significant capabilities in many applications including image enhancement and manipulation, language translation, generating images/videos from text, creating art and music, and so on. However, train ing GANs using large datasets remains highly computationally intensive for most of the standalone systems. Additionally, standalone GANs often exhibit poor synchronization between their generator and discriminator with unstable training , poor convergence along with a large number of mode collapse s and vanishing / exploding gradients. Standalone GANs also failed to learn in a decentralized environment, where the data is distributed among several client machines. Some researchers have lately used the most prevalent decentralized setting available today, called Federated Learning (FL) to develop distributed -GAN strategies as the possible solutions, although their implementations mostly failed to address the above issues mainly because of: the training instability within the distributed component s, which eventually leads to the poor synchronization among the generator s and discriminator s scattered over several machines. In this work, we developed a computationally inexpensive Wasserstein conditional Distributed Relativistic Discriminator-GAN or DRD-GAN to alleviate the above issues. DRD-GAN stabilizes its train ing (with non-convex losses) by keeping a global generator in the central server and relativistic discriminator s in the local client s (one discriminator per client ), and uses Wasserstein-1 for computing local and global losses. It eventually avoids mode collapse s, vanishing/exploding gradient s (both in the presence of iid and non-iid samples) and helps DRD-GAN to produce high-quality fake images. Apart from that, the sheer unavailability of a capable conditional distributed -GAN model has become another motivation behind the current work. Essentially, we revisited the FL paradigms, locating one discriminator per client , and a generator in the central server that aggregates the updates coming from multiple discriminator s. Relativistic discriminator s in the client s are train ed on both iid and non-iid private data. We presented a detailed mathematical formulation of DRD-GAN and empirically evaluated our claims using CIFAR-10, MNIST, EuroSAT, and CelebA datasets.

Publisher

Association for Computing Machinery (ACM)

Reference62 articles.

1. Martin Arjovsky, Soumith Chintala, and Léon Bottou. 2017. Wasserstein generative adversarial networks. In International conference on machine learning. PMLR, 214–223.

2. PerFED-GAN: Personalized federated learning via generative adversarial networks;Cao Xingjian;IEEE Internet of Things Journal,2022

3. Qi Chang, Hui Qu, Yikai Zhang, Mert Sabuncu, Chao Chen, Tong Zhang, and Dimitris N Metaxas. 2020. Synthetic learning: Learn from distributed asynchronized discriminator gan without sharing medical image data. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 13856–13866.

4. Qi Chang, Zhennan Yan, Lohendran Baskaran, Hui Qu, Yikai Zhang, Tong Zhang, Shaoting Zhang, and Dimitris N Metaxas. 2020. Multi-modal AsynDGAN: Learn from distributed medical image data without sharing private information. arXiv preprint arXiv:2012.08604 (2020).

5. Casey Chu, Andrey Zhmoginov, and Mark Sandler. 2017. Cyclegan, a master of steganography. arXiv preprint arXiv:1712.02950 (2017).

同舟云学术

1.学者识别学者识别

2.学术分析学术分析

3.人才评估人才评估

"同舟云学术"是以全球学者为主线,采集、加工和组织学术论文而形成的新型学术文献查询和分析系统,可以对全球学者进行文献检索和人才价值评估。用户可以通过关注某些学科领域的顶尖人物而持续追踪该领域的学科进展和研究前沿。经过近期的数据扩容,当前同舟云学术共收录了国内外主流学术期刊6万余种,收集的期刊论文及会议论文总量共计约1.5亿篇,并以每天添加12000余篇中外论文的速度递增。我们也可以为用户提供个性化、定制化的学者数据。欢迎来电咨询!咨询电话:010-8811{复制后删除}0370

www.globalauthorid.com

TOP

Copyright © 2019-2024 北京同舟云网络信息技术有限公司
京公网安备11010802033243号  京ICP备18003416号-3