论文中文题名: | 基于改进GAN与多尺度对齐融合的超分辨率重建算法研究 |
姓名: | |
学号: | 21208223051 |
保密级别: | 公开 |
论文语种: | chi |
学科代码: | 085400 |
学科名称: | 工学 - 电子信息 |
学生类型: | 硕士 |
学位级别: | 工程硕士 |
学位年度: | 2024 |
培养单位: | 西安科技大学 |
院系: | |
专业: | |
研究方向: | 图像处理 |
第一导师姓名: | |
第一导师单位: | |
论文提交日期: | 2024-06-18 |
论文答辩日期: | 2024-05-30 |
论文外文题名: | Research on Super-Resolution Reconstruction Algorithm Based on Improved GAN and Multi-Scale Alignment Fusion |
论文中文关键词: | |
论文外文关键词: | Super-resolution Reconstruction ; Deep Learning ; Dense Residual ; Multi-scale ; Feature Fusion |
论文中文摘要: |
超分辨率研究是计算视觉中的经典问题,随着成像技术的进步,对高清晰度图像和视频的需求激增。超分辨率技术能有效重建含丰富纹理细节的高分辨率图像和视频,并因其低成本和灵活性,应用范围日益广泛,吸引了众多学者关注。因此,本文对基于深度学习的超分辨率重建算法深入研究,论文主要的研究内容以及创新点总结如下: (1)针对单帧图像超分辨率重建算法中图像边缘平滑、伪影、以及高频信息提取不足等问题,提出了一种基于改进增强型生成对抗网络的图像超分辨率重建算法。首先,引入多尺度深度可分离特征提取模块,多尺度结构有助于捕捉不同尺度的图像特征,深度可分离卷积降低了模型参数量和计算量,提高了网络训练的稳定性。其次,引入了多尺度大内核注意力构建多尺度深度可分离密集连接模块,通过密集连接融合卷积层输入,更好的结合局部感知和远程依赖,充分提取图像特征。最后,利用多级残差网络结合大内核注意力尾部模块,进一步优化了高频细节和关键信息的整合,促进深层网络的训练,并显著提升了图像重建质量。实验验证了该算法的有效性与可行性。 (2)针对视频超分辨率重建算法中无法充分提取图像特征信息、特征对齐精度不高,以及特征融合中时序信息提取不足的问题,提出了一种基于多尺度融合和轴向可变形卷积的视频超分辨率重建算法。首先,采用多尺度特征对齐策略,对目标帧和相邻帧在不同尺度上执行对齐操作,有效提取局部和全局特征。其次,引入轴向可变形卷积对齐块,维持了局部和全局信息的平衡,优化了偏移量的预测,保证了不同尺度上目标帧和相邻帧的有效对齐。最后,采用多尺度区域关注特征融合策略,加强了对视频帧内复杂纹理区域的关注,并以不同尺度的对齐特征进行融合,增强了对齐帧的时序信息补充能力,从而提升了重建效果。实验验证了该算法的有效性与可行性。 关 键 词:超分辨率重建;深度学习;密集残差;多尺度;特征融合 研究类型:应用研究 |
论文外文摘要: |
Super-resolution research is a classic problem in computer vision, and with advancements in imaging technology, the demand for high-definition images and videos has surged. Super-resolution techniques can effectively reconstruct high-resolution images and videos with rich texture details, and due to their low cost and flexibility, their application scope is increasingly broad, attracting attention from numerous scholars. Therefore, this paper conducts an in-depth study on super-resolution reconstruction algorithms based on deep learning, summarizing the main research content and innovation points as follows: (1) To address issues in the single-frame image super-resolution reconstruction field, such as edge smoothing, artifacts, and insufficient high-frequency information extraction, a super-resolution reconstruction algorithm based on an improved enhanced generative adversarial network is proposed. Firstly, a multi-scale deep separable feature extraction module is introduced, where the multi-scale structure aids in capturing image features at different scales, and the use of deep separable convolution reduces model parameter count and computational load, enhancing network training stability. Secondly, multi-scale large kernel attention is introduced to construct a multi-scale deep separable densely connected module, merging convolutional layer inputs through dense connections to better combine local perception and distant dependencies, thoroughly extracting image features. Finally, utilizing a multi-level residual network combined with a large kernel attention tail module further optimizes the integration of high-frequency details and key information, facilitating deep network training, and significantly improving image reconstruction quality. Experimental results demonstrate significant improvements in both objective and subjective evaluation metrics, verifying the effectiveness and feasibility of the proposed algorithm. (2) To solve issues in video super-resolution reconstruction algorithms, such as insufficient feature extraction, low feature alignment accuracy, and inadequate temporal information extraction in feature fusion, a video super-resolution reconstruction algorithm based on multi-scale fusion and axial deformable convolution is proposed. The algorithm optimizes alignment and fusion modules, starting with a multi-scale feature alignment strategy that performs alignment operations on target frames and adjacent frames at different scales, effectively extracting local and global features. Then, an axial deformable convolution alignment block is introduced, maintaining a balance between local and global information, optimizing offset prediction, and ensuring effective alignment of target frames and adjacent frames at different scales. Finally, a multi-scale regional attention feature fusion strategy is employed, enhancing focus on complex texture regions within video frames and merging alignment features at different scales to bolster the temporal information supplementation capability of aligned frames, thereby improving reconstruction effects. Experimental results show significant enhancements in both objective and subjective evaluation metrics, validating the effectiveness and feasibility of the proposed algorithm. Key words: Super-resolution Reconstruction; Deep Learning; Dense Residual; Multi-scale; Feature Fusion Thesis: Application Research |
参考文献: |
[2] 钟梦圆, 姜麟. 超分辨率图像重建算法综述[J]. 计算机科学与探索, 2022, 16(5):972-990. [3] 李佳星, 赵勇先, 王京华. 基于深度学习的单幅图像超分辨率重建算法综述[J]. 自动化学报, 2021, 47(10): 2341-2363. [4] 黄健, 赵元元, 郭苹, 等. 深度学习的单幅图像超分辨率重建方法综述[J].计算机工程与应用, 2021, 57(18):13-23. [5] 吴靖,叶晓晶,黄峰,等.基于深度学习的单帧图像超分辨率重建综述[J].电子学报,2022,50(09):2265-2294. [6] 于亚龙, 穆远彪. 插值算法的研究[J]. 现代计算机: 中旬刊, 2014 (2): 32-35. [7] 李艳玲. 图像的最近邻缩放原理及实现[J]. 长治学院学报, 2016, 33(5): 31-32. [8] 王森, 杨克俭. 基于双线性插值的图像缩放算法的研究与实现[J]. 自动化技术与应用, 2008, 27(7): 44-45. [9] 王会鹏, 周利莉, 张杰. 一种基于区域的双三次图像插值算法[J]. 计算机工程, 2010,36(19): 216-218. [14] 张磊, 杨建峰, 薛彬, 等. 改进的最大后验概率估计法实现单幅图像超分辨率重建[J].激光与光电子学进展, 2011, 48(1): 82-27. [27] 陈洪刚, 李自强, 张永飞, 等. 基于迭代交替优化的图像盲超分辨率重建[J]. 电子与信息学报, 2022, 44(10): 3343-3352. [31] 江俊君, 程豪, 李震宇, 等. 深度学习视频超分辨率技术概述[J]. 中国图象图形学报, 2023, (7):1927-1964. [32] 何小海, 吴媛媛, 陈为龙, 等. 视频超分辨率重建技术综述[J]. 信息与电子工程, 2011, 9(1): 1-6. |
中图分类号: | TP391 |
开放日期: | 2024-06-18 |