Would you notice if all the objects around you simultaneously moved farther away? In a virtual-reality laboratory in Oxford, researchers have obtained the puzzling finding that humans can be "blind" to changes like this.
Computer-generated vision has shown that viewing a scene with two eyes, or walking around it, provides enough information to calculate its 3D structure. To find out how far away things are by this method, however, requires knowledge of the separation of the eyes or the distance walked. There is good evidence that the human visual system uses both these pieces of information when making judgments of 3D size, shape, and distance.
In the new work, performed at the University of Oxford, Dr. Andrew Glennerster and colleagues use an immersive virtual-reality display to show that the human visual system cannot be carrying out the same type of 3D reconstruction that is used in computer vision.
People experiencing the virtual-reality display failed to notice when the virtual scene around them quadrupled in size as they walked around, and, as a result, they made gross errors in judging the size of objects. Intriguingly, these results imply that observers are more willing to adjust their estimate of the separation between the eyes or the distance walked than to accept that the scene around them has changed in size. More broadly, these findings mark a significant shift in the debate about the way in which the brain forms a stable representation of the world--that is, the world as it is perceived to exist independent of head and eye movements.
Source: Cell Press
Explore further: Science magazine retracts study on voters' gay-rights views