Exploring visual and motion saliency for automatic video object extraction

IEEE Trans Image Process. 2013 Jul;22(7):2600-10. doi: 10.1109/TIP.2013.2253483. Epub 2013 Mar 20.

Abstract

This paper presents a saliency-based video object extraction (VOE) framework. The proposed framework aims to automatically extract foreground objects of interest without any user interaction or the use of any training data (i.e., not limited to any particular type of object). To separate foreground and background regions within and across video frames, the proposed method utilizes visual and motion saliency information extracted from the input video. A conditional random field is applied to effectively combine the saliency induced features, which allows us to deal with unknown pose and scale variations of the foreground object (and its articulated parts). Based on the ability to preserve both spatial continuity and temporal consistency in the proposed VOE framework, experiments on a variety of videos verify that our method is able to produce quantitatively and qualitatively satisfactory VOE results.

Publication types

  • Research Support, Non-U.S. Gov't