Health & Wellbeing

Camera-toting EyeRing could help blind people to "see" objects

Camera-toting EyeRing could he...
The EyeRing system features a finger-worn camera module and wirelessly connected smartphone that provide a visually-impaired user with aural feedback on objects captured by the lens
The EyeRing system features a finger-worn camera module and wirelessly connected smartphone that provide a visually-impaired user with aural feedback on objects captured by the lens
View 4 Images
The EyeRing features a 3D-printed ABS nylon outer housing containing a small VGA camera unit, a 16 MHz AVR processor, a Bluetooth radio module and a 3.7V Li-ion battery
1/4
The EyeRing features a 3D-printed ABS nylon outer housing containing a small VGA camera unit, a 16 MHz AVR processor, a Bluetooth radio module and a 3.7V Li-ion battery
The user points the ring at an object, tells the system what kind of information is needed via a microphone on the earphone cord and then clicks the button on the side of the ring unit to activate the camera and send a photo to a Bluetooth-linked smartphone for processing
2/4
The user points the ring at an object, tells the system what kind of information is needed via a microphone on the earphone cord and then clicks the button on the side of the ring unit to activate the camera and send a photo to a Bluetooth-linked smartphone for processing
The EyeRing system features a finger-worn camera module and wirelessly connected smartphone that provide a visually-impaired user with aural feedback on objects captured by the lens
3/4
The EyeRing system features a finger-worn camera module and wirelessly connected smartphone that provide a visually-impaired user with aural feedback on objects captured by the lens
The EyeRing's could also allow visually-impaired users to "read" text in non-braille books
4/4
The EyeRing's could also allow visually-impaired users to "read" text in non-braille books

Generally speaking, the vast majority of augmented reality applications that enhance the world around us by overlaying digital content on images displayed on smartphone, tablet or computer screens are aimed squarely at the sighted user. A team from the Fluid Interfaces Group at MIT's Media Lab has developed a chunky finger-worn device called EyeRing that translates images of objects captured through a camera lens into aural feedback to aid the blind.

The EyeRing system was created by Suranga Nanayakkara (who currently directs the Augmented Senses Research Group at Singapore University of Technology and Design), PhD student Roy Shilkrot and associate professor and founder of the Fluid Interfaces Group, Pattie Maes. It features a 3D-printed ABS nylon outer housing containing a small VGA camera unit, a 16 MHz AVR processor, a Bluetooth radio module and a 3.7V Li-ion battery. There's also a mini-USB port for charging the battery and reprogramming the unit, a power on/off switch and a thumb-activated button for confirming commands.

The user points the ring at an object, tells the system what kind of information is needed via a microphone on the earphone cord and then clicks the button on the side of the ring unit to activate the camera and send a photo to a Bluetooth-linked smartphone for processing
The user points the ring at an object, tells the system what kind of information is needed via a microphone on the earphone cord and then clicks the button on the side of the ring unit to activate the camera and send a photo to a Bluetooth-linked smartphone for processing

The user points the ring's camera at an object, tells the system what kind of information is needed via a microphone on the earphone cord and then clicks the button on the side of the unit. The camera snaps a photo and sends it to a Bluetooth-linked smartphone. A specially-developed Android app processes the image using computer-vision algorithms according to the preset mode selected by the user and uses a text-to-speech module to announce the appropriate results through earphones plugged into the smartphone.

The proof-of-concept prototype EyeRing system can currently be used to identify currency, text, pricing information on tags, and colors, and the preset mode can be changed by double-clicking the button and speaking a different command into the microphone. The information is also displayed on the screen of the smartphone in text form.

Other applications for the system include cane-free navigation. The current version determines the distance from objects by comparing two images taken by the camera, and can even create a 3D map of the surrounding environment. Future implementations might include the ability to shoot live video and provide all sorts of useful information to the wearer.

The EyeRing project is still very much a work in progress, with Maes confirming that any moves toward commercialization are likely at least two years away. The group is already working on an iOS flavor, though, and size and shape refinements are certain to be made (think Ringbow).

Single-user tests with a volunteer have confirmed the existing system's assistance potential, but the addition of more sensors (such as an infrared light source or a laser module, a second camera, a depth sensor or inertial sensors) to the finger-worn device could open it up to whole new levels of usefulness.

Source: MIT

The following short video shows the EyeRing prototype in operation:

EyeRing-CHI2012.m4v

0 comments
There are no comments. Be the first!