Network-based H.264/AVC whole frame loss visibility model and frame dropping methods

IEEE Trans Image Process. 2012 Aug;21(8):3353-63. doi: 10.1109/TIP.2012.2191567. Epub 2012 Mar 21.

Abstract

We examine the visual effect of whole frame loss by different decoders. Whole frame losses are introduced in H.264/AVC compressed videos which are then decoded by two different decoders with different common concealment effects: frame copy and frame interpolation. The videos are seen by human observers who respond to each glitch they spot. We found that about 39% of whole frame losses of B frames are not observed by any of the subjects, and over 58% of the B frame losses are observed by 20% or fewer of the subjects. Using simple predictive features which can be calculated inside a network node with no access to the original video and no pixel level reconstruction of the frame, we developed models which can predict the visibility of whole B frame losses. The models are then used in a router to predict the visual impact of a frame loss and perform intelligent frame dropping to relieve network congestion. Dropping frames based on their visual scores proves superior to random dropping of B frames.

MeSH terms

  • Algorithms
  • Computer Communication Networks*
  • Data Compression / methods*
  • Image Enhancement / methods*
  • Imaging, Three-Dimensional / methods*
  • Pattern Recognition, Automated / methods*
  • Reproducibility of Results
  • Sample Size
  • Sensitivity and Specificity
  • Signal Processing, Computer-Assisted*
  • Video Recording / methods*