Aishwarya Jadhav, Software Engineer in the Autopilot AI Team at Tesla, presents the “Improved Navigation Assistance for the Blind via Real-time Edge AI,” tutorial at the May 2024 Embedded Vision Summit.
In this talk, Jadhav presents recent work on AI Guide Dog, a groundbreaking research project aimed at providing navigation assistance for the blind community. This multiyear project at Carnegie Mellon University leverages AI to predict sighted human reactions in real time and convey this information audibly to blind individuals, overcoming the limitations of existing GPS apps and mobility tools for the blind.
Jadhav discusses the various vision-only and multimodal models evaluated. She also discusses imitation learning approaches currently being explored. In addition, she highlights trade-offs among the strict requirements for models to ensure explainable predictions, high accuracy and real-time processing on mobile devices. And she shares insights gained through three iterations of this project, explaining data collection procedures, training pipelines and cutting-edge vision and multimodal modeling methodologies. She concludes with some exciting results.
See here for a PDF of the slides.