(Phys.org) —Marketers partnering with software innovators are thinking up new ways to know customers better, to take the guesswork out of product targeting. Technologies hold out hope that companies will minimize the toll that product duds can take on advertising budgets and name-brands. Eye-tracking is seen as one of the technology tools with potential for selling smart. Where does the consumer look? What catches the most attention? What is that person's viewing behavior? A prototype system introduced at a Paris show this week takes the comment, "Oh, I'm just looking" and turns it into good news for shopkeepers and retail giants. "I'm just looking" customers may reveal how to target brands and, fundamentally, sell.
The eye-tracking system shown at the Conference on Human Factors in Computing Systems in Paris this week is called SideWays, by Andreas Bulling of the Max Planck Institute for Informatics in Saarbrücken, Germany, and Yanxia Zhang and Hans Gellersen at Lancaster University in the UK.
The system can track what the customer is looking at; it detects the faces of people walking by and calculates where the eyes are relative to the eye corners. Potentially, the store could readjust ads and displays for optimal positioning of products that appear to grab the greatest interest.
The device uses a single camera positioned close to the screen. In their paper, "SideWays: A Gaze Interface for SpontaneousInteraction with Situated Displays," the researchers wrote, "SideWays robustly detects whether users attend to the center of the display or cast glances to the left or right."
The viewer can control the screen, using eyes to control content such as scrolling through items on a list. Attention to the central display is the default state, but sidelong glances can trigger input or actions.
The authors think they have made a contribution in their work by validating that "SideWays enables eye gaze as input for interactive displays, without the need of prior calibration or specialist hardware." They said the significance lay in the achievement of "robust gaze control, albeit coarse-grained, without need for calibration means that our system is person-independent." Any user can walk up to a display fitted with their system, and interact with it using their eyes only.
"Person-independence and interaction without preparation are critical steps toward genuinely spontaneous interaction with displays we encounter in public environment."
Explore further: Brain signals turn into drone commands in Lisbon presentation
More information: SideWays: A Gaze Interface for Spontaneous Interaction with Situated Displays: www.andreas-bulling.de/fileadm… docs/zhang13_chi.pdf