CSINet: A Cross-Scale Interaction Network for Lightweight Image Super-Resolution

Sensors (Basel). 2024 Feb 9;24(4):1135. doi: 10.3390/s24041135.

Abstract

In recent years, advancements in deep Convolutional Neural Networks (CNNs) have brought about a paradigm shift in the realm of image super-resolution (SR). While augmenting the depth and breadth of CNNs can indeed enhance network performance, it often comes at the expense of heightened computational demands and greater memory usage, which can restrict practical deployment. To mitigate this challenge, we have incorporated a technique called factorized convolution and introduced the efficient Cross-Scale Interaction Block (CSIB). CSIB employs a dual-branch structure, with one branch extracting local features and the other capturing global features. Interaction operations take place in the middle of this dual-branch structure, facilitating the integration of cross-scale contextual information. To further refine the aggregated contextual information, we designed an Efficient Large Kernel Attention (ELKA) using large convolutional kernels and a gating mechanism. By stacking CSIBs, we have created a lightweight cross-scale interaction network for image super-resolution named "CSINet". This innovative approach significantly reduces computational costs while maintaining performance, providing an efficient solution for practical applications. The experimental results convincingly demonstrate that our CSINet surpasses the majority of the state-of-the-art lightweight super-resolution techniques used on widely recognized benchmark datasets. Moreover, our smaller model, CSINet-S, shows an excellent performance record on lightweight super-resolution benchmarks with extremely low parameters and Multi-Adds (e.g., 33.82 dB@Set14 × 2 with only 248 K parameters).

Keywords: cross-scale interaction; efficient large convolutional kernel attention; factorized convolution; super-resolution.

Grants and funding

This research was funded by the Macau Science and Technology Development Funds [Grant number 0061/2020/A2]; this research was also funded by the Science and Technology of Social Development Program [Grant number 20211800904512,20231800935472], and Dongguan Sci-tech Commissoner Program [Grant number 20231800500352].