广东工业大学学报 ›› 2024, Vol. 41 ›› Issue (01): 55-62.doi: 10.12052/gdutxb.220039
杨镇雄, 谭台哲
Yang Zhen-xiong, Tan Tai-zhe
摘要: 传统的基于深度学习的方法在低照度图像增强中已经有比较好的发挥,但是这些方法通常需要成对的数据集进行训练,而相对应的数据集正是目前难以收集的。目前的增强方法在真实的低照度图像增强中也会产生增强效果不完美和出现图像噪声等问题。针对这些问题,设计了无监督生成对抗网络,使其可以不用配对训练数据集进行训练,并且把网络分解为注意力机制网络和增强网络2个子网络。通过注意力机制网络把低照度图像中的低光区域和亮光区域区分开,并使用残差增强网络结合全局局部判别器,对图像进行增强。实验结果表明,本文的方法在低光照图像增强方面优于Enlighten-GAN、Cycle-GAN等方法。
中图分类号:
[1] ZHOU Z, FENG Z, LIU J, et al. Single-image low-light enhancement via generating and fusing multiple sources [J]. Neural Computing and Applications, 2020, 32: 6455-6465. [2] GHARBI M, CHEN J, BARRON J T, et al. Deep bilateral learning for real-time image enhancement [J]. ACM Transactions on Graphics (TOG), 2017, 36(4): 118. [3] GUO X, LI Y, LING H. LIME: Low-light image enhancement via illumination map estimation [J]. IEEE Transactions on Image Processing, 2016, 26(2): 982-993. [4] LORE K G, AKINTAYO A, SARKAR S. LLNet: a deep autoencoder approach to natural low-light image enhancement [J]. Pattern Recognition, 2017, 61: 650-662. [5] SHEN L , YUE Z , FENG F , et al. MSR-net: low-light image enhancement using deep convolutional network[EB/OL]. arXiv: 1711.02488 (2017-11-07) [2022-03-25]. https://doi.org/10.48550/arXiv.1711.02488. [6] WEI C , WANG W , YANG W, et al. Deep retinex decomposition for low-light enhancement[EB/OL]. arXiv: 1808.04560 (2018-08-14) [2022-03-25]. https://doi.org/10.48550/arXiv.1808.04560. [7] CHEN C , CHEN Q , XU J , et al. Learning to see in the dark[EB/OL]. arXiv: 1805.01934 (2018-05-04) [2022-03-25]. https://doi.org/10.48550/arXiv.1805.01934. [8] KALANTARI N K, RAMAMOORTHI R. Deep high dynamic range imaging of dynamic scenes[J]. ACM Transactions on Graphics, 2017, 36(4): 1-12. [9] WU X , SHAO J , GAO L , et al. Unpaired image-to-image translation from shared deep space[C]//2018 25th IEEE International Conference on Image Processing (ICIP) . Athens: IEEE, 2018: 2127-2131. [10] LIU M Y , BREUEL T , KAUTZ J . Unsupervised image-to-image translation networks[EB/OL]. arXiv: 1703.00848 (2018-07-23) [2022-03-25]. https://doi.org/10.48550/arXiv.1703.00848. [11] MADAM N T, KUMAR S, RAJAGOPALAN A N. Unsupervised class-specific deblurring[C]//Computer Vision–ECCV 2018: 15th European Conference. Munich: Springer International Publishing, 2018: 358-374. [12] HUANG X, LIU M Y, BELONGIE S, et al. Multimodal unsupervised image-to-image translation[EB/OL]. arXiv: 1804.04732 (2018-08-14) [2022-03-25]. https://doi.org/10.48550/arXiv.1804.04732. [13] CHOI Y, CHOI M, KIM M, et al. Stargan: unified generative adversarial networks for multi-domain image-to-image translation[EB/OL]. arXiv: 1711.09020 (2018-09-21) [2022-03-25]: https://doi.org/10.48550/arXiv.1711.09020. [14] HOFFMAN J, TZENG E, PARK T, et al. CyCADA: cycle-consistent adversarial domain adaptation[EB/OL].arXiv: 1711.03213 (2017-12-29) [2022-03-25]. https://doi.org/10.48550/arXiv.1711.03213. [15] ZHU J Y, PARK T, ISOLA P, et al. Unpaired image-to-image translation using cycle-consistent adversarial networks[EB/OL].arXiv: 1703.10593 (2020-08-24) [2022-03-25]. https://doi.org/10.48550/arXiv.1703.10593. [16] GOODFELLOW I, POUGET-ABADIE J, MIRZA M, et al. Generative adversarial networks [J]. Communications of the ACM, 2020, 63(11): 139-144. [17] GAWADE A, PANDHARKAR R, DEOLEKAR S. Gantoon: creative cartoons using generative adversarial network[C]//Information, Communication and Computing Technology: 5th International Conference, ICICCT 2020. Singapore: Springer, 2020: 222-230. [18] HU D. An introductory survey on attention mechanisms in NLP problems[EB/OL]. arXiv: 1811.05544 (2018-11-12) [2022-03-25]. https://doi.org/10.48550/arXiv.1811.05544. [19] WANG F, TAX D M J. Survey on the attention based RNN model and its applications in computer vision[EB/OL]. arXiv: 1601.06823 (2016-01-25) [2022-03-25]. https://doi.org/10.48550/arXiv.1601.06823. [20] JIANG X, ZHANG L, XU M, et al. Attention scaling for crowd counting[C]//2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR). Seattle, WA: IEEE. 2020: 4706-4715. [21] HAN Y S, YOO J, YE J C. Deep residual learning for compressed sensing CT reconstruction via persistent homology analysis[EB/OL]. arXiv: 1611.06391 (2016-11-25) [2022-03-25]. https://doi.org/10.48550/arXiv.1611.06391. [22] SHIBATA N, TANITO M, MITSUHASHI K, et al. Development of a deep residual learning algorithm to screen for glaucoma from fundus photography [J]. Scientific Reports, 2018, 8(1): 14665. [23] KUANG D Y. On reducing negative jacobian determinant of the deformation predicted by deep registration networks[EB/OL]. arXiv: 1907.00068 (2019-06-28) [2022-03-25]. https://doi.org/10.48550/arXiv.1907.00068. [24] XIONG W, LIU D, SHEN X, et al. Unsupervised real-world low-light image enhancement with decoupled networks. [EB/OL]. arXiv: 2005.02818 (2020-05-06) [2022-03-25]. https://doi.org/10.48550/arXiv.2005.02818. [25] RONNEBERGER O, FISCHER P, BROX T. U-net: convolutional networks for biomedical image segmentation[EB/OL]. arXiv: 1505.04597 (2015-05-18) [2022-03-25]. https://doi.org/10.48550/arXiv.1505.04597 [26] JIANG Y, GONG X, LIU D, et al. Enlightengan: deep light enhancement without paired supervision [J]. IEEE Transactions on Image Processing, 2021, 30: 2340-2349. [27] IIZUKA S, SIMO-SERRA E, ISHIKAWA H. Globally and locally consistent image completion [J]. ACM Transactions on Graphics (TOG), 2017, 36(4): 107. [28] MAO X, LI Q, XIE H, et al. Least squares generative adversarial networks[EB/OL]. arXiv: 1611.04076 (2017-04-05) [2022-03-25]. https://doi.org/10.48550/arXiv.1611.04076. [29] HUYNH-THU Q, GHANBARI M. Scope of validity of PSNR in image/video quality assessment [J]. Electronics letters, 2008, 44(13): 800-801. |
[1] | 邝永年, 王丰. 基于前景区域生成对抗网络的视频异常行为检测研究[J]. 广东工业大学学报, 2024, 41(01): 63-68,92. |
[2] | 赖志茂, 章云, 李东. 基于Transformer的人脸深度伪造检测技术综述[J]. 广东工业大学学报, 2023, 40(06): 155-167. |
[3] | 曾安, 陈旭宙, 姬玉柱, 潘丹, 徐小维. 基于自注意力和三维卷积的心脏多类分割方法[J]. 广东工业大学学报, 2023, 40(06): 168-175. |
[4] | 吴亚迪, 陈平华. 基于用户长短期偏好和音乐情感注意力的音乐推荐模型[J]. 广东工业大学学报, 2023, 40(04): 37-44. |
[5] | 曹智雄, 吴晓鸰, 骆晓伟, 凌捷. 融合迁移学习与YOLOv5的安全帽佩戴检测算法[J]. 广东工业大学学报, 2023, 40(04): 67-76. |
[6] | 赖东升, 冯开平, 罗立宏. 基于多特征融合的表情识别算法[J]. 广东工业大学学报, 2023, 40(03): 10-16. |
[7] | 吴俊贤, 何元烈. 基于通道注意力的自监督深度估计方法[J]. 广东工业大学学报, 2023, 40(02): 22-29. |
[8] | 刘洪伟, 林伟振, 温展明, 陈燕君, 易闽琦. 基于MABM的消费者情感倾向识别模型——以电影评论为例[J]. 广东工业大学学报, 2022, 39(06): 1-9. |
[9] | 滕少华, 董谱, 张巍. 融合语义结构的注意力文本摘要模型[J]. 广东工业大学学报, 2021, 38(03): 1-8. |
[10] | 梁观术, 曹江中, 戴青云, 黄云飞. 一种基于注意力机制的无监督商标检索方法[J]. 广东工业大学学报, 2020, 37(06): 41-49. |
[11] | 曾碧卿, 韩旭丽, 王盛玉, 徐如阳, 周武. 基于双注意力卷积神经网络模型的情感分析研究[J]. 广东工业大学学报, 2019, 36(04): 10-17. |
[12] | 高俊艳, 刘文印, 杨振国. 结合注意力与特征融合的目标跟踪[J]. 广东工业大学学报, 2019, 36(04): 18-23. |
|