Seminar – Road Scene Understanding for the Visually Impaired



Time and place:

  • Fri 8:15-9:30, Room 09.150


Understanding road scenes and creating maps from urban environments is a challenging task in Computer Vision. It is used in self-driving cars, robotics and assistive navigation systems for blind or visually impaired pedestrians.
The topic of this seminar will be to build a software tools that will enable the creation of a navigation assistant for the blind and visually impaired. In particular the following issues will be addressed:
1. Define the hardware resources (e.g. computational resources, camera/cell phone, TOF/Lidar/focal length, sensors, ...)
2. Set up a web-based collaborative labeling environment using crowd-based labeling tools (e.g. EXACT (
3. define guidelines for crowdsourced labeling
4. Define use cases (face recognition, license plates) for managing and anonymizing the data.
The aim of this project seminar is to build tools that will enable preparation of data, e.g. image segmentation of roads, sidewalks and obstacles, etc. In a second step, the data should be ready for the creation of maps and respective annotations from the scenes so it can be used building for an assistive navigation system for blind or visually impaired pedestrians.

Recommended Literature

- J. C. Chang, S. Amershi, E. Kamar, Revolt: Collaborative Crowdsourcing for Labeling Machine Learning Datasets, CHI 2017, Denver, CO, USA, May 6-11, 2017, - Oana Inel, Khalid Khamkham, Tatiana Cristea et al., CrowdTruth: Machine-Human Computation Framework for Harnessing Disagreement in Gathering Annotated Data, International Semantic Web Conference, ISWC 2014: The Semantic Web - ISWC 2014, pp 486 504, - Brody Huval, Tao Wang, Sameep Tandon et al., An Empirical Evaluation of Deep Learning on Highway Driving, Cornell University, Apr. 7, 2015, - Joel Pazhayampallil, Free Space Detection with Deep Nets for Autonomous Driving,

Additional information

Expected participants: 30