基于信息蒸馏和共享注意力网络的轻量级图像超分辨率重建方法

A lightweight image super-resolution reconstruction method based on information distillation and shared attention network

  • 摘要: 当前基于卷积神经网络(Convolutional Neural Network, CNN)和基于Transformer框架的图像超分辨率重建方法已经取得了很大成功,但是这些方法还存在参数量过多、复杂度过高、内存需求较大等缺点。针对上述缺点,提出一种全新的基于信息蒸馏和共享注意力网络(Information Distillation and Shared Attention Network, IDSA-Net)的轻量级图像超分辨率重建方法,以在提升特征表达能力的同时降低计算负担。首先,引入注意力共享机制,构建注意力共享蒸馏模块,使后续模块无需重复计算空间注意力矩阵,该部分是自注意力计算中的主要开销,从而实现矩阵信息的跨层共享,减少计算量。此外,设计大核通道信息提纯模块,该模块结合通道混洗操作,并利用大核注意力对蒸馏后的特征进行权重重分配,从而增强有用特征并抑制冗余信息。同时,提出新的实例归一混合注意力模块,学习通道特征和空间特征。最后,利用亚像素卷积层实现上采样操作进行图像重建。在Set5、Set14、BSD100和Urban100四个公开的基准数据集,通过对比和消融实验的定量和可视化分析,所提方法在峰值信噪比(Peak Signal-to-Noise Ratio, PSNR)和结构相似性指数(Structural Similarity, SSIM)评价指标中优于十一种最先进的图像超分辨率重建方法,验证了其有效性和准确性。

     

    Abstract: Lightweight super-resolution reconstruction is an important technology in the field of image processing and has been widely applied in various domains. Although current image super-resolution methods, such as those based on convolutional neural networks and Transformer models, have achieved significant success in this field, they still face considerable challenges and limitations, including excessive parameter counts, high computational complexity, and substantial memory requirements. To address these issues, a novel lightweight image super-resolution reconstruction method named the Information Distillation and Shared Attention Network (IDSA-Net) is proposed, aiming to reduce computational burden while enhancing feature representation capability. First, an attention-sharing mechanism is introduced to construct an Attention-Sharing Distillation Module, allowing subsequent modules to avoid repeatedly computing the spatial attention matrix, which constitutes the main computational overhead in self-attention calculations. This enables cross-layer sharing of matrix information and reduces computational costs. The module integrates local feature extraction with convolutional operations, while self-attention calculations are paired with sequence modeling units, thereby achieving information purification during the feature extraction and distillation stages. This enhances the network's ability to capture effective features and suppresses noise introduced during distillation. Furthermore, a Large-Kernel Channel Information Purification Module is designed, which combines channel shuffle operations and utilizes large-kernel depth wise separable convolutions from large-kernel attention to expand the receptive field. This enhances the model's perception of multi-scale features and global context, allowing for reweighting of the distilled features to strengthen useful information and suppress redundancies. Additionally, a novel Instance Normalization-based Hybrid Attention Module is proposed to learn channel and spatial features. In the channel attention phase, average pooling is performed separately along the width and height directions of the input features to mitigate information loss caused by traditional global pooling. Channel features are extracted using one-dimensional depth wise separable convolutions, avoiding the dimensionality reduction and expansion operations commonly used in conventional methods. This better preserves the integrity of channel features and improves the precision of information representation. In the spatial attention phase, instance normalization is applied to each channel of every sample individually, enabling the model to focus more on local structural features within the image rather than relying on global statistical information. Finally, sub-pixel convolutional layers are employed for up sampling and image reconstruction. Quantitative and visual analyses through comparative and ablation experiments on four public benchmark datasets—Set5, Set14, BSD100, and Urban100—demonstrate that the proposed method outperforms eleven state-of-the-art image super-resolution methods in terms of Peak Signal-to-Noise Ratio (PSNR) and Structural Similarity Index (SSIM), validating its effectiveness and accuracy.

     

/

返回文章
返回