@inproceedings{5649b15163f449f5ad47a79008265709,
title = "Outdoor walking guide for the visually-impaired people based on semantic segmentation and depth map",
abstract = "In this study, we proposed a wearable guiding system, which contains an embedded system-Jetson AGX Xavier launched by Nvidia and a RGB-D binocular depth camera-Stereolabs ZED2, for guiding visually-impaired people to walk outdoors. Using the deep learning image segmentation model and the depth map obtained by the ZED2, the front image of the blind is divided into seven divisions. Each division has its confidence of walkability which is computed by our specific methods. Based on the confidence of walkability, the most suitable direction for the visually-impaired people is selected and voice prompts are played to lead the visually-impaired people walking forward on the sidewalk or walking crosswalk to cross the road safely. An experiment is performed to verify the effectiveness of the proposed system. ",
keywords = "deep learning, depth map, obstacle avoidance, semantic segmentation, visually-impaired people, wearable device",
author = "Hsieh, {I. Hsuan} and Cheng, {Hsiao Chu} and Ke, {Hao Hsiang} and Chen, {Hsiang Chieh} and Wang, {Wen June}",
note = "Publisher Copyright: {\textcopyright} 2020 IEEE.; 1st International Conference on Pervasive Artificial Intelligence, ICPAI 2020 ; Conference date: 03-12-2020 Through 05-12-2020",
year = "2020",
month = dec,
doi = "10.1109/ICPAI51961.2020.00034",
language = "???core.languages.en_GB???",
series = "Proceedings - 2020 International Conference on Pervasive Artificial Intelligence, ICPAI 2020",
publisher = "Institute of Electrical and Electronics Engineers Inc.",
pages = "144--147",
booktitle = "Proceedings - 2020 International Conference on Pervasive Artificial Intelligence, ICPAI 2020",
}