An Unsupervised Fundus Image Enhancement Method with Multi-Scale Transformer and Unreferenced Loss
-
Published:2023-07-04
Issue:13
Volume:12
Page:2941
-
ISSN:2079-9292
-
Container-title:Electronics
-
language:en
-
Short-container-title:Electronics
Author:
Hu Yanzhe1ORCID, Li Yu1, Zou Hua2ORCID, Zhang Xuedong3
Affiliation:
1. School of Electronic and Electrical Engineering, Wuhan Textile University, Wuhan 430077, China 2. School of Computer Science, Wuhan University, Wuhan 430072, China 3. School of Information Engineering, Tarim University, Alaer 843300, China
Abstract
Color fundus images are now widely used in computer-aided analysis systems for ophthalmic diseases. However, fundus imaging can be affected by human, environmental, and equipment factors, which may result in low-quality images. Such quality fundus images will interfere with computer-aided diagnosis. Existing methods for enhancing low-quality fundus images focus more on the overall visualization of the image rather than capturing pathological and structural features at the finer scales of the fundus image sufficiently. In this paper, we design an unsupervised method that integrates a multi-scale feature fusion transformer and an unreferenced loss function. Due to the loss of microscale features caused by unpaired training, we construct the Global Feature Extraction Module (GFEM), a combination of convolution blocks and residual Swin Transformer modules, to achieve the extraction of feature information at different levels while reducing computational costs. To improve the blurring of image details caused by deep unsupervised networks, we define unreferenced loss functions that improve the model’s ability to suppress edge sharpness degradation. In addition, uneven light distribution can also affect image quality, so we use an a priori luminance-based attention mechanism to improve low-quality image illumination unevenness. On the public dataset, we achieve an improvement of 0.88 dB in PSNR and 0.024 in SSIM compared to the state-of-the-art methods. Experiment results show that our method outperforms other deep learning methods in terms of vascular continuity and preservation of fine pathological features. Such a framework may have potential medical applications.
Funder
Bingtuan Science and Technology Program
Subject
Electrical and Electronic Engineering,Computer Networks and Communications,Hardware and Architecture,Signal Processing,Control and Systems Engineering
Reference41 articles.
1. Wang, X., Peng, Y., Lu, L., Lu, Z., Bagheri, M., and Summers, R.M. (2017, January 21–26). ChestX-Ray8: Hospital-Scale Chest X-Ray Database and Benchmarks on Weakly-Supervised Classification and Localization of Common Thorax Diseases. Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA. 2. Color-Aware Two-Branch DCNN for Efficient Plant Disease Classification;Schuler;MENDEL,2022 3. Huck Yang, C.H., Liu, F., Huang, J.H., Tian, M., Lin, I.H., Liu, Y.C., Morikawa, H., Yang, H.H., and Tegner, J. (2018). Asian Conference on Computer Vision, Springer. 4. Xing, X., Liang, G., Blanton, H., Rafique, M.U., Wang, C., Lin, A.L., and Jacobs, N. (2020). European Conference on Computer Vision, Springer. 5. Ghani, A., See, C.H., Sudhakaran, V., Ahmad, J., and Abd-Alhameed, R. (2019). Accelerating retinal fundus image classification using artificial neural networks (ANNs) and reconfigurable hardware (FPGA). Electronics, 8.
|
|