PDA

View Full Version : Camera-toting EyeRing could help blind people to "see" objects


(‘“*JiĢäR*”’)
05-02-2013, 12:31 PM
Camera-toting EyeRing
could help blind people to "see" objects

http://2.bp.blogspot.com/-0TftLOsKO7I/UC4iZgxDq-I/AAAAAAAA7Dk/wpqu4bCvFtg/s640/funwithnet.png

Generally speaking, the vast majority of augmented reality applications that enhance the world around us by overlaying digital content on images displayed on smartphone, tablet or computer screens are aimed squarely at the sighted user. A team from the Fluid Interfaces Group at MIT's Media Lab has developed a chunky finger-worn device called EyeRing that translates images of objects captured through a camera lens into aural feedback to aid the blind.

http://images.gizmag.com/gallery_lrg/eyering-3.JPG

http://images.gizmag.com/gallery_lrg/eyering.JPG

http://images.gizmag.com/gallery_lrg/eyering-4.JPG

The EyeRing system was created by Suranga Nanayakkara (who currently directs the Augmented Senses Research Group at Singapore University of Technology and Design), PhD student Roy Shilkrot and associate professor and founder of the Fluid Interfaces Group, Pattie Maes. It features a 3D-printed ABS nylon outer housing containing a small VGA camera unit, a 16 MHz AVR processor, a Bluetooth radio module and a 3.7V Li-ion battery. There's also a mini-USB port for charging the battery and reprogramming the unit, a power on/off switch and a thumb-activated button for confirming commands.

http://images.gizmag.com/inline/eyering-3.JPG

The user points the ring's camera at an object, tells the system what kind of information is needed via a microphone on the earphone cord and then clicks the button on the side of the unit. The camera snaps a photo and sends it to a Bluetooth-linked smartphone. A specially-developed Android app processes the image using computer-vision algorithms according to the preset mode selected by the user and uses a text-to-speech module to announce the appropriate results through earphones plugged into the smartphone.

The proof-of-concept prototype EyeRing system can currently be used to identify currency, text, pricing information on tags, and colors, and the preset mode can be changed by double-clicking the button and speaking a different command into the microphone. The information is also displayed on the screen of the smartphone in text form.

Other applications for the system include cane-free navigation. The current version determines the distance from objects by comparing two images taken by the camera, and can even create a 3D map of the surrounding environment. Future implementations might include the ability to shoot live video and provide all sorts of useful information to the wearer.

The EyeRing project is still very much a work in progress, with Maes confirming that any moves toward commercialization are likely at least two years away. The group is already working on an iOS flavor, though, and size and shape refinements are certain to be made (think Ringbow).

Single-user tests with a volunteer have confirmed the existing system's assistance potential, but the addition of more sensors (such as an infrared light source or a laser module, a second camera, a depth sensor or inertial sensors) to the finger-worn device could open it up to whole new levels of usefulness.

Source: MIT

SHAYAN
07-05-2013, 01:54 AM
Good Sharing

(‘“*JiĢäR*”’)
07-05-2013, 08:38 AM
Thanks

Copyright ©2008
Powered by vBulletin Copyright ©2000 - 2009, Jelsoft Enterprises Ltd.