Radar-Based Hand Gesture Recognition With Feature Fusion Using Robust CNN-LSTM and Attention Architecture

In Human-Computer Interaction (HCI), seamless hand gesture recognition is essential for intuitive and natural interactions. Gestures act as a universal language, bridging the gap between humans and machines. Radar-based recognition surpasses traditional optical methods, offering robust interaction c...

Full description

Saved in:
Bibliographic Details
Main Authors: Irshad Khan, Young-Woo Kwon
Format: Article
Language:English
Published: IEEE 2025-01-01
Series:IEEE Access
Subjects:
Online Access:https://ieeexplore.ieee.org/document/10950371/
Tags: Add Tag
No Tags, Be the first to tag this record!
_version_ 1849723299557277696
author Irshad Khan
Young-Woo Kwon
author_facet Irshad Khan
Young-Woo Kwon
author_sort Irshad Khan
collection DOAJ
description In Human-Computer Interaction (HCI), seamless hand gesture recognition is essential for intuitive and natural interactions. Gestures act as a universal language, bridging the gap between humans and machines. Radar-based recognition surpasses traditional optical methods, offering robust interaction capabilities in diverse environments. This article introduces a novel deep learning approach for hand gesture recognition, leveraging convolutional neural networks (CNNs), long short-term memory networks (LSTMs), and attention mechanisms. CNNs extract spatial features from radar signals, while LSTMs model the temporal dependencies crucial for dynamic gestures. Additionally, attention mechanisms enhance feature selection, ultimately improving recognition performance. We evaluate our method on the UWB-Gestures dataset, which has 12 gestures from eight people that were recorded using three X-Thru X4 UWB impulse radar sensors. Our processing pipeline integrates feature extraction, LSTM-attention blocks, and dense layers for final classification. Early fusion techniques, which combine spatial and temporal features in the initial stages, yield superior results, achieving an overall accuracy of 98.33% and outperforming intermediate fusion methods across gesture classes. To enhance the model’s robustness, we evaluated its performance under common contributors of radar-specific noise scenarios in practical applications, including Gaussian noise, signal inversion, and multipath interference. Our model demonstrates high resilience, maintaining performance despite adverse conditions. As compared to state-of-the-art approaches, our approach delivers competitive accuracy and enhanced robustness, offering a reliable solution for noise-resilient radar-based hand gesture recognition in real-world applications.
format Article
id doaj-art-845aa0b968f8477bbf3d2e7918e2027f
institution DOAJ
issn 2169-3536
language English
publishDate 2025-01-01
publisher IEEE
record_format Article
series IEEE Access
spelling doaj-art-845aa0b968f8477bbf3d2e7918e2027f2025-08-20T03:11:04ZengIEEEIEEE Access2169-35362025-01-0113692816929110.1109/ACCESS.2025.355829310950371Radar-Based Hand Gesture Recognition With Feature Fusion Using Robust CNN-LSTM and Attention ArchitectureIrshad Khan0https://orcid.org/0000-0001-6960-2083Young-Woo Kwon1https://orcid.org/0000-0003-0625-8232School of Computer Science and Engineering, Kyungpook National University, Daegu, South KoreaSchool of Computer Science and Engineering, Kyungpook National University, Daegu, South KoreaIn Human-Computer Interaction (HCI), seamless hand gesture recognition is essential for intuitive and natural interactions. Gestures act as a universal language, bridging the gap between humans and machines. Radar-based recognition surpasses traditional optical methods, offering robust interaction capabilities in diverse environments. This article introduces a novel deep learning approach for hand gesture recognition, leveraging convolutional neural networks (CNNs), long short-term memory networks (LSTMs), and attention mechanisms. CNNs extract spatial features from radar signals, while LSTMs model the temporal dependencies crucial for dynamic gestures. Additionally, attention mechanisms enhance feature selection, ultimately improving recognition performance. We evaluate our method on the UWB-Gestures dataset, which has 12 gestures from eight people that were recorded using three X-Thru X4 UWB impulse radar sensors. Our processing pipeline integrates feature extraction, LSTM-attention blocks, and dense layers for final classification. Early fusion techniques, which combine spatial and temporal features in the initial stages, yield superior results, achieving an overall accuracy of 98.33% and outperforming intermediate fusion methods across gesture classes. To enhance the model’s robustness, we evaluated its performance under common contributors of radar-specific noise scenarios in practical applications, including Gaussian noise, signal inversion, and multipath interference. Our model demonstrates high resilience, maintaining performance despite adverse conditions. As compared to state-of-the-art approaches, our approach delivers competitive accuracy and enhanced robustness, offering a reliable solution for noise-resilient radar-based hand gesture recognition in real-world applications.https://ieeexplore.ieee.org/document/10950371/Hand gesture recognitionUWB-sensorsdeep learningfusionnoise robust
spellingShingle Irshad Khan
Young-Woo Kwon
Radar-Based Hand Gesture Recognition With Feature Fusion Using Robust CNN-LSTM and Attention Architecture
IEEE Access
Hand gesture recognition
UWB-sensors
deep learning
fusion
noise robust
title Radar-Based Hand Gesture Recognition With Feature Fusion Using Robust CNN-LSTM and Attention Architecture
title_full Radar-Based Hand Gesture Recognition With Feature Fusion Using Robust CNN-LSTM and Attention Architecture
title_fullStr Radar-Based Hand Gesture Recognition With Feature Fusion Using Robust CNN-LSTM and Attention Architecture
title_full_unstemmed Radar-Based Hand Gesture Recognition With Feature Fusion Using Robust CNN-LSTM and Attention Architecture
title_short Radar-Based Hand Gesture Recognition With Feature Fusion Using Robust CNN-LSTM and Attention Architecture
title_sort radar based hand gesture recognition with feature fusion using robust cnn lstm and attention architecture
topic Hand gesture recognition
UWB-sensors
deep learning
fusion
noise robust
url https://ieeexplore.ieee.org/document/10950371/
work_keys_str_mv AT irshadkhan radarbasedhandgesturerecognitionwithfeaturefusionusingrobustcnnlstmandattentionarchitecture
AT youngwookwon radarbasedhandgesturerecognitionwithfeaturefusionusingrobustcnnlstmandattentionarchitecture