{{youtube>nCqww9tXIJo?960x540|Beobot 2.0}} The VisualAid project is launched to develop a fully integrated visual aid device to allow the visually impaired users go to a location and find an object of interest. The system utilizes a number of state-of-the-art techniques such as SLAM (Simultaneous Localization and Mapping) and visual object detection. The device is designed to complement existing tools for navigation, such as seeing eye dogs and white canes. The system incorporates the use of a Smartphone, Head Mounted Camera, Computer, and Vibro-tactile and Bone conduction Headphones. Specifically, a user is able to choose from a number of tasks ("go somewhere", "find the Cheerios") through the Smartphone User interface. For our project, we start by focusing on the supermarket setting first. The head mounted camera then sees the outside world and pass it to the software. The vision algorithms then recognize the user whereabouts, and create a path to a goal location. Once the user arrives at the specified destination, the system then pinpoints the item(s) of importance to allow the user to grab it. The system is comprised of a few sub-systems: the [[visualaid/User_Interface |user interface]], [[visualaid/navigation |navigation]], and [[Beobot_2.0/object_recognition|object recognition and localization]] components. ====People==== ^ ^ Name ^ Task ^ |1. | [[http://ilab.usc.edu/siagian/ | Christian Siagian]] | Overall system architecture | |2. | [[http://ilab.usc.edu/| Nii Mante]] | Navigation | |3. | [[http://ilab.usc.edu/ | Kavi Thakkor]] | Object Recognition | |4. | [[http://ilab.usc.edu/ | Jack Eagan]] | User Interface | |4. | [[http://ilab.usc.edu/ | James Weiland]] | | |5. | [[http://ilab.usc.edu/ | Laurent Itti]] | | Internal notes can be found [[https://docs.google.com/document/d/19l5841BuzokBSV1E-LHNqk2u7V5MHGZ94Zi-DnLhGwo/edit | here]]. ====User's Manual==== The user's manual can be found [[visualaid/Users_Manual |here]]. ====System overview==== {{:tatrc_full_nav_rec_large.jpg?1200|}} The following describes the different subsystems in the visual aid device: ===User Interface=== The [[visualaid/User_Interface |user interface]] is used to obtain request from the user. ===Navigation=== The [[visualaid/navigation |navigation]] system is used to guide the user to their destination. ===Object Recognition And Localization=== The [[Beobot_2.0/object_recognition|object recognition and localization]] component is used to detect the requested object in the user field of view, and guide the user to reach and grab the object. ====Links (Related Projects)==== * [[http://grozi.calit2.net/ | Grozi]]: handheld scene text recognition system to recognize supermarket aisles. Also have object recognition to guide user to the object. * [[http://openglass.us | OpenGlass]]: Google Glass-based system, which utilizes crowdsourcing to recognize objects or gain other information from pictures. * [[http://fluid.media.mit.edu/projects/eyering | EyeRing]]: Finger-worn visual assistant that provides a voice feedback describing features such as color of objects, currency values, and price tags.