Embedded solution to detect and classify head level objects using stereo vision for visually impaired people with audio feedback

Abstract This work presents an embedded solution for detecting and classifying head-level objects using stereo vision to assist blind individuals. A custom dataset was created, featuring five classes of head-level objects, selected based on a survey of visually impaired users. Object detection and c...

Full description

Saved in:
Bibliographic Details
Main Authors: Kevin Muñoz, Mario Chavarria, Luisa Ortiz, Silvan Suter, Klaus Schönenberger, Bladimir Bacca-Cortes
Format: Article
Language:English
Published: Nature Portfolio 2025-05-01
Series:Scientific Reports
Subjects:
Online Access:https://doi.org/10.1038/s41598-025-01529-7
Tags: Add Tag
No Tags, Be the first to tag this record!
Description
Summary:Abstract This work presents an embedded solution for detecting and classifying head-level objects using stereo vision to assist blind individuals. A custom dataset was created, featuring five classes of head-level objects, selected based on a survey of visually impaired users. Object detection and classification were achieved using deep-neural networks such as YoloV5. The system computes the relative range and orientation of detected head-level objects and provides audio feedback to alert the user about nearby objects. Four types of tests were conducted: a dataset-based test, achieving a mAP@0.95 of 0.89 for head-level objects classification; a quantitative assessment of range and orientation, with an average error of 0.028 m ± 0.004 and 2.05°±0.09, respectively; a field test conducted over a week at different times and lighting conditions, yielding a precision/recall of 98.21%/93.75% for head-level object classification; and user tests with Head-level identification accuracy of 91% and obstacle-avoidance/local-navigation where users reported an average of 88.75% for low or middle risk.
ISSN:2045-2322