(Phys.org) -- A research team from Pierre and Marie Curie University in Paris have ported technology originally developed to help robots maneuver in real world environments to Braille enabled devices that help vision impaired people do the same.
The new technology uses a pair of glasses affixed with sensors and cameras connected to hand-held devices that allow the blind wearer to feel the 3D environment around them. The team from the university's Institute of Intelligent Systems and Robotics will be presenting their findings at this months IEEE International Conference on Robotics and Automation.
The new technology incorporates cameras and sensors initially developed for use with robotics technology, along with image processing hardware and software, to gather information about the surrounding environment. From there though, the two technologies diverge. Instead of the data collected being routed to AI systems for use by a robot, it moves instead to a processor that converts 3D imagery to Braille signals that are passed along to real-time Braille devices in the hands of the person using the device. The result, the team says, is situational awareness unlike anything else currently available to assist the blind in moving around without benefit of a cane, guide dog or other assistive device.
Two create true 3D imagery, the system uses two cameras on each side of the glasses, which are connected to an image processor that picks out objects, edges and other pertinent details of the view ahead. From that, a map is generated to represent the scene from the users perspective. That map, along with data from a gyroscope and accelerometers that provide information about speed and current location, is then converted by another processor to a series of signals that are sent to the Braille devices in the hand. Each is a flat pad with pins that can be raised or lowered using heat, to create a virtual tactile real-time image of the surrounding environment. And because the system generates new maps at a rate of about ten per second, the user is able to move, the developers say, at walking speed through a real world environment.
Only time will tell of course, just how useful the new technology will be, as those that use it will likely develop a connection with the system over time akin to the way others grow accustomed to keyboards, canes and guide dogs, and only they will be able to judge whether new systems such as this are as useful as they appear in demonstrations.
Explore further: An eel-lectrifying future for autonomous underwater robots