Video Compressive Sensing Reconstruction Using Unfolded LSTM

Sensors (Basel). 2022 Sep 21;22(19):7172. doi: 10.3390/s22197172.

Abstract

Video compression sensing can use a few measurements to obtain the original video by reconstruction algorithms. There is a natural correlation between video frames, and how to exploit this feature becomes the key to improving the reconstruction quality. More and more deep learning-based video compression sensing (VCS) methods are proposed. Some methods overlook interframe information, so they fail to achieve satisfactory reconstruction quality. Some use complex network structures to exploit the interframe information, but it increases the parameters and makes the training process more complicated. To overcome the limitations of existing VCS methods, we propose an efficient end-to-end VCS network, which integrates the measurement and reconstruction into one whole framework. In the measurement part, we train a measurement matrix rather than a pre-prepared random matrix, which fits the video reconstruction task better. An unfolded LSTM network is utilized in the reconstruction part, deeply fusing the intra- and interframe spatial-temporal information. The proposed method has higher reconstruction accuracy than existing video compression sensing networks and even performs well at measurement ratios as low as 0.01.

Keywords: end-to-end deep learning network; measurement matrix training; unfolded LSTM; video compressing sensing.

MeSH terms

  • Algorithms
  • Data Compression* / methods
  • Physical Phenomena

Grants and funding

This research received no external funding.