180 likes | 304 Vues
This project focuses on developing a navigation system for visually impaired individuals through auditory feedback. It addresses the limitations of existing navigation solutions, such as white canes and guide dogs, which impose restrictions on freedom. Our innovative system combines visual information processing with 3D sound creation to provide a new way of navigating through environments. Users will rely primarily on their sense of hearing for obstacle detection and spatial awareness. The application is designed for Linux, utilizing advanced technologies like OpenCV and MySQL for effective performance.
E N D
Final Presentation May, 2011 AISN http://www.cs.bgu.ac.il/~royif/AISN Auditory Imaging for Sightless Navigation
Project Team Academic Advisor: Prof. Yuval Elovici Technical Advisor: Dr. Rami Puzis Team Members: YakirDahan Vitali Sepetnitsky RoyiFreifeld
General Overview • Introduction • Motivation & Problem Domain • Existing Solutions • Our Solution • Project Goals • System Architecture • General Diagram • Description of modules – A closer look • Technologies • The system in action …
IntroductionMotivation • Most of our navigation in the everyday life heavily depends on visual feedback that we get from our environment • When the ability to see the surroundings is missing due to visual impairments, the ability to navigate is also damaged
IntroductionExisting Solutions • Physical sense: • White Cane • Guide Dog Drawbacks: • Limit the freedom of their owner • Guide dogs are expensive to train and intended to only a small part of the blind community • Sensory substitution: • Warning of obstacles (e.g. Optical Radar) • Sonar-like images scanning (e.g. The vOICe) Drawbacks: • Requires an intensive training and is hard to get used to, due to a non-intuitive conversion between image and sound
IntroductionOur Solution A combination of visual information processing and 3D sound creation and positioning: • Taking a stream of frames • Processing the frames and retrieving visual information • Creating appropriate sounds • Performing an auditory spatialization of the sound
IntroductionProject Goals • Sightless navigation by sensory substitution A development of an application that allows a person to navigate, relying primarily on the sense of hearing • Integration with a spatial auditory environment • Providing a flexible environment for future research
System Architecture Sound_T pixel_t 3rd party
System ArchitectureA Closer Look • UI Unit: • Enables the user to configure the system through a user-friendly GUI • Core Unit: • The system’s engine • Responsible for: • Initializing the system • Running all the components • Communicating with hardware and database
System ArchitectureA Closer Look (cont.) • Image-Processing Unit: • Receives images obtained from a camera and retrieves visual information • Sound-Creation Unit: • Responsible for creating appropriate sounds according to the visual information • Sound-Positioning Unit: • Performs auditory spatialization of the soundin the 3D space
System ArchitectureA Closer Look (cont.) • General Database: • Contains the data relevant to an interaction of the system with users e.g. • Profiles • Data about installed technologies • Technologies Data Storage: • Contains the installed technologies used by the system
UsedTechnologies & PlatformsGeneral • The application was designed for Linux OS • Code written in C/C++ • Eclipse IDE & QtSDK for GUI • Handling HRTF data by calling MATLAB routines from the code • Database by MySQL RDBMS
UsedTechnologies & PlatformsWithin the system • Integration with Primesense 3D camera using OpenNI API • Image processing by OpenCVlibrary • Sound positioning by Integration with Sound Scape Renderer (SSR) and OpenAL API • Audio connection handling using JACK audio server daemon
aNY Questions ? Thank You!