Intrinsic dimensionality predicts the saliency of natural dynamic scenes

IEEE Trans Pattern Anal Mach Intell. 2012 Jun;34(6):1080-91. doi: 10.1109/TPAMI.2011.198.

Abstract

Since visual attention-based computer vision applications have gained popularity, ever more complex, biologically inspired models seem to be needed to predict salient locations (or interest points) in naturalistic scenes. In this paper, we explore how far one can go in predicting eye movements by using only basic signal processing, such as image representations derived from efficient coding principles, and machine learning. To this end, we gradually increase the complexity of a model from simple single-scale saliency maps computed on grayscale videos to spatiotemporal multiscale and multispectral representations. Using a large collection of eye movements on high-resolution videos, supervised learning techniques fine-tune the free parameters whose addition is inevitable with increasing complexity. The proposed model, although very simple, demonstrates significant improvement in predicting salient locations in naturalistic videos over four selected baseline models and two distinct data labeling scenarios.

Publication types

  • Research Support, Non-U.S. Gov't

MeSH terms

  • Algorithms*
  • Eye Movements / physiology
  • Humans
  • Pattern Recognition, Visual
  • Principal Component Analysis
  • Video Recording
  • Vision, Ocular / physiology*
  • Visual Perception