Posted Mar 22, 2013
Original, 01 May 2012 by Helen Knight
Magazine issue 2862.
ROBOTS need help navigating their surroundings and sophisticated location systems to keep track of their position. Now the same technologies are being adapted to help blind people navigate indoor and outdoor spaces independently.
One such system, being developed by Edwige Pissaloux and colleagues at the Institute of Intelligent Systems and Robotics at the Pierre and Marie Curie University in Paris, France, consists of a pair of glasses equipped with cameras and sensors like those used in robot exploration. The system, unveiled at a talk at the Massachusetts Institute of Technology this month, produces a 3D map of the wearer’s environment and their position within it that is constantly updated and displayed in a simplified form on a handheld electronic Braille device. It could eventually allow blind people to make their way, unaided, wherever they want to go, says Pissaloux. “Navigation for me means not only being able to move around by avoiding nearby obstacles, but also to understand how the space is socially organised – for example, where you are in relation to the pharmacy, library or intersection,” she says.
Two cameras on either side of the glasses generate a 3D image of the scene. A processor analyses the image, picking out the edges of walls or objects, which it uses to create a 3D map. The system’s collection of accelerometers and gyroscopes – like those used in robots to monitor their position – keeps track of the user’s location and speed. This information is combined with the image to determine the user’s position in relation to other objects.
The system generates almost 10 maps per second which are transmitted to the handheld Braille device to be displayed as a dynamic tactile map. The Braille pad consists of an 8-centimetre-square grid of 64 taxels – pins with a shape memory alloy spring in the middle. When heat is applied to the springs, they expand, raising the pins to represent boundaries. The Braille version of the map is updated fast enough for a visually-impaired wearer to pass through an area at walking speed, says Pissaloux. Seth Teller, who develops assistive technologies at MIT, calls the work exciting and ambitious.
This is not the only robotics project to be re-purposed. Software that predicts how far a robot has travelled based on information from its on-board sensors is being modified to track a person’s movements based on their stride length. The low-cost system, being developed by Eelke Folmer and Kostas Bekris at the University of Nevada in Reno would help blind people navigate around buildings using just a smartphone.
The new system uses freely available 2D digital indoor maps and the smartphone’s built-in accelerometer and compass. Directions are provided using synthetic speech. To help the smartphone calibrate and adjust to a user’s individual stride length, the user must initially use touch to detect the landmarks in their environment, such as corridor intersections, doors and elevators. The system will be presented at the IEEE International Conference on Robotics and Automation in St Paul, Minnesota, in May.
David Ross at the Atlanta Vision Loss Center in Decatur, Georgia, says that the sensing problems faced by robots and blind people are similar but there are big differences. “Sensing systems developed for mobile robots may have some application, but must be adapted considerably to suit a wide variety of human needs and situations,” he says.
The all-seeing ring
A virtual assistant can help blind people explore their surroundings. Developed by Suranga Nanayakkara at the MIT’s Media Lab, EyeRing consists of a ring equipped with a camera, and headphones. The user points the ring at an object they are holding and uses voice commands to say what they need to know – the colour of an item of clothing, say, or the denomination of paper money. The ring takes a picture of the object, which is transmitted wirelessly to a cellphone, where software analyses the image. The required information is then read out by a synthesised voice. It is being presented at the Conference on Human Factors in Computing Systems in Austin, Texas, in M