Video Colorization Based on Variational Autoencoder
-
Published:2024-06-20
Issue:12
Volume:13
Page:2412
-
ISSN:2079-9292
-
Container-title:Electronics
-
language:en
-
Short-container-title:Electronics
Author:
Zhang Guangzi1ORCID, Hong Xiaolin1, Liu Yan1ORCID, Qian Yulin1ORCID, Cai Xingquan1ORCID
Affiliation:
1. School of Information Science and Technology, North China University of Technology, Beijing 100144, China
Abstract
This paper introduces a variational autoencoder network designed for video colorization using reference images, addressing the challenge of colorizing black-and-white videos. Although recent techniques perform well in some scenarios, they often struggle with color inconsistencies and artifacts in videos that feature complex scenes and long durations. To tackle this, we propose a variational autoencoder framework that incorporates spatio-temporal information for efficient video colorization. To improve temporal consistency, we unify semantic correspondence with color propagation, allowing for simultaneous guidance in colorizing grayscale video frames. Additionally, the variational autoencoder learns spatio-temporal feature representations by mapping video frames into a latent space through an encoder network. The decoder network then transforms these latent features back into color images. Compared to traditional coloring methods, our approach accurately captures temporal relationships between video frames, providing precise colorization while ensuring video consistency. To further enhance video quality, we apply a specialized loss function that constrains the generated output, ensuring that the colorized video remains spatio-temporally consistent and natural. Experimental results demonstrate that our method significantly improves the video colorization process.
Funder
Funding Project of Humanities and Social Sciences Foundation of the Ministry of Education in China
Reference33 articles.
1. Let There Be Color!;Iizuka;ACM Trans. Graph.,2016 2. Zhang, R., Isola, P., and Efros, A.A. (2016, January 11–14). Colorful Image Colorization. Proceedings of the Computer Vision—ECCV 2016, Amsterdam, The Netherlands. 3. Larsson, G., Maire, M., and Shakhnarovich, G. (2016, January 11–14). Learning Representations for Automatic Colorization. Proceedings of the Computer Vision—ECCV 2016, Amsterdam, The Netherlands. 4. Lai, W.S., Huang, J.B., Wang, O., Shechtman, E., Yumer, E., and Yang, M.H. (2018). Learning Blind Video Temporal Consistency. arXiv. 5. Zhao, J., Liu, L., Snoek, C.G.M., Han, J., and Shao, L. (2018). Pixel-Level Semantics Guided Image Colorization. arXiv.
|
|