EyeNav is an Eye-control interface for people with ALS.
EyeNav is using live video stream from a camera located on glasses frame, pointed directly to the user’s eye. The video is processed and analyzed using computer vision algorithm and translates the eye movement to keystrokes (up/down/left/right)
inspired by the eyeWriter project, and uses same low-cost hardware (webcam on plastic glasses).
What makes EyeNav different is that unlike direct, mouse-like cursor manipulation, which can be tiresome - it is based on discrete gestures (like keyboard), for navigation.
The project was conceived, born, and took its first steps in 12 hours, during a Hackathon for disabled people
back to Projects