Why does the world appear stable while our eyes move?

February 15, 2008

Whenever we shift our gaze, attention is directed to a new target. This shift in attention causes a brief compression of visual space, according to a study published February 15 in the open-access journal PLoS Computational Biology.

The team of researchers from the University of Münster, Germany, describes a model of brain function in which eye movement signals are used to boost the neural representation of objects located at the future eye position. This boost comes at the expense of a temporary loss of spatial accuracy. This research shows a direct correlation between visual perception and eye movement control.

Humans move their eyes 2-3 times a second without noticing. Each gaze shift triggers a host of internal brain processes with very delicate timing. The gaze shift is preceded by a brief shift of attention towards the new gaze target so that visual processing at the target area improves some 50 milliseconds before the eye itself looks at the target. This preceding improvement increases the sensitivity of visual neurons in many brain areas, which then respond more strongly to stimuli near the gaze target just prior to the gaze movement.

Using a detailed neuro-computational model of the representation of the visual world in cortical maps, the researchers investigated the consequences of these sensitivity changes to the perception of spatial location. Their results showed that objects presented just before the eye movement appear to lie at the gaze target rather than at their true spatial location, akin to a compression of visual space. Moreover, this model explains a peculiar finding that neurons in some brain areas appear to move their receptive field, i.e. the visual direction to which they respond, prior to eye movement. Analysis of the net effect of all receptive field changes in the model shows that the brain dynamically recruits cells for processing visual information around the target. This increase in processing capacity presumably allows one to perceive details of the object before looking at it, therefore making the world appear stable while we move our eyes.

This new model prompts many predictions that can guide experimental research – one step towards theory driven brain research. The model also paves the way to develop novel concepts for artificial vision systems.

Citation: Hamker FH, Zirnsak M, Calow D, Lappe M (2008) The peri-saccadic perception of objects and space. PLoS Comput Biol 4(2): e31. doi:10.1371/journal.pcbi.0040031 (compbiol.plosjournals.org/perlserv/?request=get-document&doi=10.1371/journal.pcbi.0040031)

Source: Public Library of Science

Explore further: Sandia teams with industry to improve human-data interaction

Related Stories

Sandia teams with industry to improve human-data interaction

August 13, 2015

Intelligence analysts working to identify national security threats in warzones or airports or elsewhere often flip through multiple images to create a video-like effect. They also may toggle between images at lightning speed, ...

Chameleons' eyes are not so independent

July 8, 2015

Famed for their ability to change colour, chameleons have yet another mind-boggling talent: their eyes appear to swivel completely independently. This means that they can simultaneously track two completely different views ...

Robot eyes will benefit from insect vision

June 10, 2015

The way insects see and track their prey is being applied to a new robot under development at the University of Adelaide, in the hopes of improving robot visual systems.

Goshawk hunt and prey-evasion strategies revealed

January 21, 2015

Stealth is the goshawk's greatest asset. Plummeting out of the air, the raptors fix their gaze on the oblivious victim below. Intrigued by the birds' attack tactics, Suzanne Amador Kane from Haverford College, USA, decided ...

Camera-carrying falcons reveal mystery of raptor pursuit

January 15, 2014

Raptors are the masters of the aerial dogfight. Intercepting prey on the wing, falcons lock their victims in their gaze before engaging in battle. Intrigued by the raptor's attack strategy, Suzanne Amador Kane recruited falconers ...

Visual nudge improves accuracy of mammogram readings

January 26, 2012

In 2011 -- to the consternation of women everywhere -- a systematic review of randomized clinical trials showed that routine mammography was of little value to younger women at average or low risk of breast cancer.

Recommended for you

New nanomaterial maintains conductivity in 3-D

September 4, 2015

An international team of scientists has developed what may be the first one-step process for making seamless carbon-based nanomaterials that possess superior thermal, electrical and mechanical properties in three dimensions.

Astronomers detect the farthest galaxy yet with Keck telescope

September 4, 2015

A team of Caltech researchers that has spent years searching for the earliest objects in the universe now reports the detection of what may be the most distant galaxy ever found. In an article published August 28, 2015 in Astrophysical ...

1 comment

Adjust slider to filter visible comments by rank

Display comments: newest first

Ashibayai
5 / 5 (1) Feb 15, 2008
Very interesting emergence in the complexity of sensory perception working with the brain's need for efficiency.

Please sign in to add a comment. Registration is free, and takes less than a minute. Read more

Click here to reset your password.
Sign in to get notified via email when new comments are made.