Affiliation:
1. State Key Laboratory of Software Development Environment Beihang University Beijing P. R. China
2. Zhongguancun Laboratory Beijing P. R. China
Abstract
AbstractObjects in a real scene often occlude each other and inferring a complete appearance from the visible part is an important and challenging task. In this paper, the authors propose a self‐supervised generative adversarial network GIGAN (GAN for generating the invisible), which can generate the complete appearance of objects without labelled invisible part information. The authors build two cycle transformation networks CycleIncomplete (CycleI) and CycleComplete (CycleC) that share parameters to improve the accuracy of mask completion. This design does not require well‐matched training images and can make better use of the limited labelled samples. In addition, the authors propose a conditional normalization module and combine it with the inferred complete mask output. The combination not only enhances the content recovery ability and obtains more realistic outputs, but also improves the efficiency of the generation process. Experimental results show that compared with existing self‐supervised learning models, our method achieves l1 error, mean intersection‐over‐union (mIOU), and Fréchet inception distance (FID) improvements on the COCOA and KINS datasets.
Publisher
Institution of Engineering and Technology (IET)
Subject
Electrical and Electronic Engineering,Computer Vision and Pattern Recognition,Signal Processing,Software
Cited by
1 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献
1. VGGAN: Visual Grounding GAN Using Panoptic Transformers;2023 8th International Conference on Image, Vision and Computing (ICIVC);2023-07-27