Video Super-Resolution Method Using Deformable Convolution-Based Alignment Network

Sensors (Basel). 2022 Nov 3;22(21):8476. doi: 10.3390/s22218476.

Abstract

With the advancement of sensors, image and video processing have developed for use in the visual sensing area. Among them, video super-resolution (VSR) aims to reconstruct high-resolution sequences from low-resolution sequences. To use consecutive contexts within a low-resolution sequence, VSR learns the spatial and temporal characteristics of multiple frames of the low-resolution sequence. As one of the convolutional neural network-based VSR methods, we propose a deformable convolution-based alignment network (DCAN) to generate scaled high-resolution sequences with quadruple the size of the low-resolution sequences. The proposed method consists of a feature extraction block, two different alignment blocks that use deformable convolution, and an up-sampling block. Experimental results show that the proposed DCAN achieved better performances in both the peak signal-to-noise ratio and structural similarity index measure than the compared methods. The proposed DCAN significantly reduces the network complexities, such as the number of network parameters, the total memory, and the inference speed, compared with the latest method.

Keywords: alignment network; channel attention; convolutional neural network; deformable convolution; dilated convolution; spatial attention; video super-resolution.

MeSH terms

  • Image Processing, Computer-Assisted* / methods
  • Neural Networks, Computer*
  • Signal-To-Noise Ratio

Grants and funding

This research received no external funding.