Affiliation:
1. School of Software, Dalian University of Technology, 116024, China
Abstract
Generating pictures from text is an interesting, classic, and challenging task. Benefited from the development of generative adversarial networks (GAN), the generation quality of this task has been greatly improved. Many excellent cross modal GAN models have been put forward. These models add extensive layers and constraints to get impressive generation pictures. However, complexity and computation of existing cross modal GANs are too high to be deployed in mobile terminal. To solve this problem, this paper designs a compact cross modal GAN based on canonical polyadic decomposition. We replace an original convolution layer with three small convolution layers and use an autoencoder to stabilize and speed up training. The experimental results show that our model achieves 20% times of compression in both parameters and FLOPs without loss of quality on generated images.
Funder
Fundamental Research Funds for the Central Universities
Subject
Electrical and Electronic Engineering,Computer Networks and Communications,Information Systems
Reference39 articles.
1. Generating images from captions with attention;E. Mansimov
2. Pixel recurrent neural networks;A. van den Oord
3. Conditional image generation with PixelCNN decoders;A. van den Oord;Advances in Neural Information Processing Systems,2016
4. Generative adversarial text to image synthesis;S. Reed;Proceedings of Machine Learning Research,2016
5. StackGAN++: Realistic Image Synthesis with Stacked Generative Adversarial Networks
Cited by
5 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献