A Masked-Pre-Training-Based Fast Deep Image Prior Denoising Model

Author:

Ji Shuichen1,Xu Shaoping2ORCID,Cheng Qiangqiang3,Xiao Nan2,Zhou Changfei2,Xiong Minghai2

Affiliation:

1. School of Information Engineering, Nanchang University, Nanchang 330031, China

2. School of Mathematics and Computer Sciences, Nanchang University, Nanchang 330031, China

3. School of Mechanical and Electronic Engineering, Gandong University, Fuzhou 344000, China

Abstract

Compared to supervised denoising models based on deep learning, the unsupervised Deep Image Prior (DIP) denoising approach offers greater flexibility and practicality by operating solely with the given noisy image. However, the random initialization of network input and network parameters in the DIP leads to a slow convergence during iterative training, affecting the execution efficiency heavily. To address this issue, we propose the Masked-Pre-training-Based Fast DIP (MPFDIP) Denoising Model in this paper. We enhance the classical Restormer framework by improving its Transformer core module and incorporating sampling, residual learning, and refinement techniques. This results in a fast network called FRformer (Fast Restormer). The FRformer model undergoes offline supervised training using the masked processing technique for pre-training. For a specific noisy image, the pre-trained FRformer network, with its learned parameters, replaces the UNet network used in the original DIP model. The online iterative training of the replaced model follows the DIP unsupervised training approach, utilizing multi-target images and an adaptive loss function. This strategy further improves the denoising effectiveness of the pre-trained model. Extensive experiments demonstrate that the MPFDIP model outperforms existing mainstream deep-learning-based denoising models in reducing Gaussian noise, mixed Gaussian–Poisson noise, and low-dose CT noise. It also significantly enhances the execution efficiency compared to the original DIP model. This improvement is mainly attributed to the FRformer network’s initialization parameters obtained through masked pre-training, which exhibit strong generalization capabilities for various types and intensities of noise and already provide some denoising effect. Using them as initialization parameters greatly improves the convergence speed of unsupervised iterative training in the DIP. Additionally, the techniques of multi-target images and the adaptive loss function further enhance the denoising process.

Funder

Natural Science Foundation of China

Publisher

MDPI AG

Reference51 articles.

1. Chen, Z., Kaushik, P., Shuangfei, Z., Alvin, W., Zhile, R., Alex, S., Alex, C., and Li, F. (2023, January 17–24). AutoFocusFormer: Image Segmentation off the Grid. Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada.

2. Jie, Q., Wu, J., Pengxiang, Y., Ming, L., Ren, Y., Xuefeng, X., Yitong, W., Rui, W., Shilei, W., and Xin, P. (2023, January 17–24). Freeseg: Unified, universal and open-vocabulary image segmentation. Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada.

3. Attentional Full-Relation Network for Few-Shot Image Classification;Li;Chin. J. Comput.,2023

4. SCS-Net: Sharpend cosine similarity based neural network for hyperspectral image classification;Ahmad;IEEE Geosci. Remote. Sens. Lett.,2024

5. Channel Attention Embedded Transformer for Image Super-Resolution Reconstruction;Xiong;J. Image Graph. China,2023

同舟云学术

1.学者识别学者识别

2.学术分析学术分析

3.人才评估人才评估

"同舟云学术"是以全球学者为主线,采集、加工和组织学术论文而形成的新型学术文献查询和分析系统,可以对全球学者进行文献检索和人才价值评估。用户可以通过关注某些学科领域的顶尖人物而持续追踪该领域的学科进展和研究前沿。经过近期的数据扩容,当前同舟云学术共收录了国内外主流学术期刊6万余种,收集的期刊论文及会议论文总量共计约1.5亿篇,并以每天添加12000余篇中外论文的速度递增。我们也可以为用户提供个性化、定制化的学者数据。欢迎来电咨询!咨询电话:010-8811{复制后删除}0370

www.globalauthorid.com

TOP

Copyright © 2019-2024 北京同舟云网络信息技术有限公司
京公网安备11010802033243号  京ICP备18003416号-3