Depth perception in disparity-defined objects: finding the balance between averaging and segregation

Philos Trans R Soc Lond B Biol Sci. 2016 Jun 19;371(1697):20150258. doi: 10.1098/rstb.2015.0258.

Abstract

Deciding what constitutes an object, and what background, is an essential task for the visual system. This presents a conundrum: averaging over the visual scene is required to obtain a precise signal for object segregation, but segregation is required to define the region over which averaging should take place. Depth, obtained via binocular disparity (the differences between two eyes' views), could help with segregation by enabling identification of object and background via differences in depth. Here, we explore depth perception in disparity-defined objects. We show that a simple object segregation rule, followed by averaging over that segregated area, can account for depth estimation errors. To do this, we compared objects with smoothly varying depth edges to those with sharp depth edges, and found that perceived peak depth was reduced for the former. A computational model used a rule based on object shape to segregate and average over a central portion of the object, and was able to emulate the reduction in perceived depth. We also demonstrated that the segregated area is not predefined but is dependent on the object shape. We discuss how this segregation strategy could be employed by animals seeking to deter binocular predators.This article is part of the themed issue 'Vision in our three-dimensional world'.

Keywords: binocular disparity; depth perception; disparity averaging; object segregation; psychophysics; stereopsis.

Publication types

  • Research Support, Non-U.S. Gov't

MeSH terms

  • Depth Perception*
  • Humans
  • Vision Disparity
  • Vision, Binocular*