Enhanced lightweight infrared object detection algorithm for assistive navigation in visually impaired individuals

Abstract This study introduces an advanced infrared scene detection algorithm, enhancing the YOLOv8 model for aiding visually impaired individuals in navigation. The focus is on the neck network, integrating attention scale sequences to boost multi‐level perception, particularly for small object det...

Full description

Saved in:
Bibliographic Details
Main Authors: Zhimin Bai, Yang Yang, Jian Wang, Zhengyang Li, Jiajun Wang, Chunxin Liu
Format: Article
Language:English
Published: Wiley 2024-12-01
Series:IET Image Processing
Subjects:
Online Access:https://doi.org/10.1049/ipr2.13233
Tags: Add Tag
No Tags, Be the first to tag this record!
Description
Summary:Abstract This study introduces an advanced infrared scene detection algorithm, enhancing the YOLOv8 model for aiding visually impaired individuals in navigation. The focus is on the neck network, integrating attention scale sequences to boost multi‐level perception, particularly for small object detection. This is achieved by adding upsampling and downsampling in the P2 module. Additionally, the CIoU loss function is refined with Inner‐SIoU, elevating bounding box detection precision. A distinctive feature of the approach is its monocular distance and velocity measurement integration, which operates independently of external devices, providing direct navigation support for visually impaired people. Further, the enhanced YOLOv8 is adapted for mobile use, employing pruning and lightweight methods, which substantially enhance its practicality. The experimental results on the FLIR and WOTR datasets demonstrate that, compared to the original YOLOv8n, the improved algorithm has achieved a 2.1% and 3.2% increase in mAP0.5, respectively. Furthermore, the mAP0.5--0.95 has seen a 2.2% and 3.8% improvement. Concurrently, the model size has been reduced by 55% and 60%, and the number of parameters has decreased by 60% and 67%. Compared to other assistive travel methods for visually impaired individuals, our work demonstrates superior practicality.
ISSN:1751-9659
1751-9667