论文中文题名: | 基于强化学习的图像恢复方法研究 |
姓名: | |
学号: | 19208208045 |
保密级别: | 保密(1年后开放) |
论文语种: | chi |
学科代码: | 085212 |
学科名称: | 工学 - 工程 - 软件工程 |
学生类型: | 硕士 |
学位级别: | 工程硕士 |
学位年度: | 2022 |
培养单位: | 西安科技大学 |
院系: | |
专业: | |
研究方向: | 媒体计算与可视化 |
第一导师姓名: | |
第一导师单位: | |
第二导师姓名: | |
论文提交日期: | 2022-06-19 |
论文答辩日期: | 2022-06-06 |
论文外文题名: | Research on Image Recovery Algorithm Based on Reinforcement Learning |
论文中文关键词: | |
论文外文关键词: | Image recovery ; image denoising ; image retouching ; Personalized image retouching ; reinforcement learning |
论文中文摘要: |
数字图像在传播过程中不可避免的会造成失真与退化,通常人类专家在恢复图像时会使用多种不同的图像恢复工具。受此启发,本文基于强化学习方法集成已经成熟的图像处理工具,通过合理的决策以寻找一种新的组合方式,完成对图像的全局或局部恢复。本文的主要工作和创新点如下: (1) 针对现有的图像去噪方法大多是基于结构多样的卷积神经网络,同时网络结构设计复杂,去噪过程可解释性不强的问题,提出了一种基于强化学习和生成对抗网络的图像去噪方法。该方法为了实现高质量、可解释的图像降噪效果,所以将每一个像素都被视为一个强化学习智能体,像素智能体通过选取高斯滤波、双边滤波器、盒滤波器和中值滤波器等具有可解释的计算过程与物理意义的去噪算法作为动作并逐步的进行像素级去噪。同时,为了提升最后的去噪效果,本文采用对抗学习的方式提升了模型对图像细节的恢复能力,同时设计了一种全卷积判别器来解决像素级操作所需的像素级奖励这一问题。实验结果表明:相比于 CBDnet 算法,提出方法在 SIDD Medium测试集上的 1280 张测试图像上平均 PSNR 值提高了 1.12,在 Nam 测试数据集上平均PSNR 值提高了 0.17,同时参数量与计算量比 DnCNN 减少了 33%。 (2) 针对现有的强化学习方法在进行图像修饰时,只能进行全局修饰的问题,提出了一种基于强化学习的像素级图像修饰方法。受人类专家在 PhotoShop 等软件中修饰图像过程的启发,本文对图像的不同区域使用不同的修饰工具进行调整。首先,本文通过对局部区域进行特定的曝光与对比度调整;然后,通过高动态范围中的曝光融合算法避免直接分区域修饰产生的拼接边界;最后,在实现对不同区域进行特定曝光对比度修饰的同时达到了较好的效果,实验结果表明:在 MIT-Adobe-5K 的 500 张测试数据集上,在 SSIM 值上取得了较好的成绩,平均 SSIM 值比 DeepUPE 要高 0.018,在 LOL 数据集上平均 PSNR 值比EnlightenGAN 高 0.14,平均 SSIM 值比 Zero-DCE要高 0.057。 (3) 针对基于强化学习的单风格图像修饰算法无法有效解决多风格的个性化图像修饰的问题,提出了一种基于强化学习和生成对抗网络的个性化图像修饰方法。首先,通过对比学习预训练一个风格编码器,提取图像的滤镜风格;然后,将待修饰图像和参考图像输入风格编码器获得各自的滤镜风格编码,并通过自适应实例归一化将混合后的特征输入策略网络;最后,通过控制曝光,对比度,高光、阴影等 12 种不可微的滤波器工具,使目标图像滤镜风格与参考图像更为接近。实验结果表明:在单风格修饰上提出方法在平均 PSNR 值上比 Pix2Pix 高 0.62,平均 SSIM 值则高 0.056,在多风格修饰上,PSNR 值和∆00∗ 值则分别比 WCT2 高 0.66 和低 3.797。 |
论文外文摘要: |
Digital image inevitably distorts and degrades in the process of transmission. Using a variety of different image restoration tools can restore the image. Inspired by this, this paper integrates mature image processing tools based on reinforcement learning, and finds a new combination way through reasonable decision-making, to complete the global or local restoration of the image. The main work and innovation of this paper are as follows: (1) Most of image denoising methods are based on convolution neural networks with various structures. While the network structure design is complex and the explanation of denoising process is not strong. An image denoising method based on reinforcement learning and generating countermeasure network is proposed. In order to achieve high-quality and interpretable image denoising, each pixel is regarded as a reinforcement learning agent. Pixel agents select a Gaussian filter, a bilateral filter, a box filter, and a median filter as actions and carry out pixel-level denoising step by step. At the same time, in order to improve the final denoising effect, this paper uses the method of adversarial learning to improve the ability of the model by recovering the image details, and designing a full convolution discriminator. The experimental results show that: compared with the CBDnet algorithm, the average PSNR of the proposed method on 1280 test images of the SIDD Medium test set is increased by 1.12. On the Nam test data set, the average PSNR is increased by 0.17, while the number of parameters and the amount of calculation are 33% that less than that of DnCNN. (2) In order to solve the problem that the reinforcement learning methods only modifies images globally, a pixel-level image modification method based on reinforcement learning is proposed. Inspired by the process of decorating the image in PhotoShop and other software, this paper uses different modification tools to adjust different areas of the image. First, this paper makes a specific exposure and contrast adjustment to the local area, and then avoids the splicing boundary caused by direct sub-region modification through the exposure fusion algorithm in high dynamic range. Second, good results are achieved by modifying the specific exposure contrast of different regions. The experimental results show that on the 500th test data set of MIT-Adobe-5K, the average SSIM is 0.018 which is higher than that of DeepUPE. The average PSNR of LOL data set is 0.14 which is higher than that of Enlighten GAN. And the average SSIM is 0.057 which is higher than Zero-DCE. (3) Single-style image modification algorithm based on reinforcement learning can not solve the problem of multi-style personalized image modification. A personalized image modification method based on reinforcement learning and generating countermeasure network is proposed. First, a style encoder is pre-trained by comparative learning to extract the filter style of the image. Then the filter style encoders are input to the modified image. The reference image obtains their respective filter style coding. The mixed feature input strategy network is normalized by adaptive examples. Second, the filter style of the target image is closer to that of the reference image by controlling 12 kinds of non-differentiable filter tools, such as exposure, contrast, highlight, shadow. The experimental results show that the average PSNR of the proposed method is 0.62 which is higher than that of Pix2Pix. The average SSIM is 0.056 which is higher than that of WCT2 in single style modification. In multi-style modification, the PSNR and ∆00∗ are 0.66 higher and 3.797 lower than that of SSIM,respectively.
|
参考文献: |
[1] 何龙军. 医学影像远程实时后处理方法研究[D].华中科技大学,2015. [3] 申 红 . 小 波 变 换 域 井 下 视 频 监 控 图 像 改 进 阈 值 去 噪 方 法 [J]. 金 属 矿 山,2017(07):151-154. [4] 刘帅奇,扈琪,刘彤,赵杰.合成孔径雷达图像去噪算法研究综述[J].兵器装备工程学报,2018,39(12):106-112+252. computer vision and pattern recognition. 2020: 9394-9402. [9] He K, Zhang X, Ren S, et al. Deep residual learning for image [36]Kosugi S, Yamasaki T. Unpaired image enhancement featuring [38]Bellman R. A Markovian decision process[J]. Journal of mathematics and mechanics, 1957: 679-684. [48]Watkins C J C H, Dayan P. Q-learning[J]. Machine learning, 1992, 8(3): 279-292. [70]Kim H U, Koh Y J, Kim C S. PieNet: Personalized image enhancement network[C]//European Conference on Computer Vision. Springer, Cham, 2020: 374-390. |
中图分类号: | TP391.4 |
开放日期: | 2023-06-28 |