Searching for just a few words should be enough to get started. If you need to make more complex queries, use the tips below to guide you.
Article type: Research Article
Authors: Zhou, Daxina; b | Qian, Yurongb; * | Ma, Yuanyuana; b | Fan, Yingyinga; b | Yang, Jianenga; b | Tan, Fuxianga; b
Affiliations: [a] School of Software, XinJiang University, Urumqi, China | [b] Key Laboratory of Software Engineering Technology, Xinjiang University, Urumqi, China
Correspondence: [*] Corresponding author. Yurong Qian, Key Laboratory of Software Engineering Technology, Xinjiang University, Urumqi, 830000, China. E-mail: liuhui@stu.xju.edu.cn.
Abstract: Low-illumination image restoration has been widely used in many fields. Aiming at the problem of low resolution and noise amplification in low light environment, this paper applies style transfer of CycleGAN(Cycle-Consistent Generative Adversarial Networks) to low illumination image enhancement. In the design network structure, different convolution kernels are used to extract the features from three paths, and the deep residual shrinkage network is designed to suppress the noise after convolution. The color deviation of the image can be resolved by the identity loss of CycleGAN. In the discriminator, different convolution kernels are used to extract image features from two paths. Compared with the training and testing results of Deep-Retinex network, GLAD network, KinD and other network methods on LOL-dataset and Brightening dataset, CycleGAN based on multi-scale depth residuals contraction proposed in this experiment on LOL-dataset results image quality evaluation indicators PSNR = 24.62, NIQE = 4.9856, SSIM = 0.8628, PSNR = 27.85, NIQE = 4.7652, SSIM = 0.8753. From the visual effect and objective index, it is proved that CycleGAN based on multi-scale depth residual shrinkage has excellent performance in low illumination enhancement, detail recovery and denoising.
Keywords: Style migration, cycle-consistent generative adversarial networks, depth residual shrinkage, image enhancement
DOI: 10.3233/JIFS-211664
Journal: Journal of Intelligent & Fuzzy Systems, vol. 42, no. 3, pp. 2383-2395, 2022
IOS Press, Inc.
6751 Tepper Drive
Clifton, VA 20124
USA
Tel: +1 703 830 6300
Fax: +1 703 830 2300
sales@iospress.com
For editorial issues, like the status of your submitted paper or proposals, write to editorial@iospress.nl
IOS Press
Nieuwe Hemweg 6B
1013 BG Amsterdam
The Netherlands
Tel: +31 20 688 3355
Fax: +31 20 687 0091
info@iospress.nl
For editorial issues, permissions, book requests, submissions and proceedings, contact the Amsterdam office info@iospress.nl
Inspirees International (China Office)
Ciyunsi Beili 207(CapitaLand), Bld 1, 7-901
100025, Beijing
China
Free service line: 400 661 8717
Fax: +86 10 8446 7947
china@iospress.cn
For editorial issues, like the status of your submitted paper or proposals, write to editorial@iospress.nl
如果您在出版方面需要帮助或有任何建, 件至: editorial@iospress.nl