Radar-Based Hand Gesture Recognition With Feature Fusion Using Robust CNN-LSTM and Attention Architecture
In Human-Computer Interaction (HCI), seamless hand gesture recognition is essential for intuitive and natural interactions. Gestures act as a universal language, bridging the gap between humans and machines. Radar-based recognition surpasses traditional optical methods, offering robust interaction c...
Saved in:
| Main Authors: | , |
|---|---|
| Format: | Article |
| Language: | English |
| Published: |
IEEE
2025-01-01
|
| Series: | IEEE Access |
| Subjects: | |
| Online Access: | https://ieeexplore.ieee.org/document/10950371/ |
| Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
| _version_ | 1849723299557277696 |
|---|---|
| author | Irshad Khan Young-Woo Kwon |
| author_facet | Irshad Khan Young-Woo Kwon |
| author_sort | Irshad Khan |
| collection | DOAJ |
| description | In Human-Computer Interaction (HCI), seamless hand gesture recognition is essential for intuitive and natural interactions. Gestures act as a universal language, bridging the gap between humans and machines. Radar-based recognition surpasses traditional optical methods, offering robust interaction capabilities in diverse environments. This article introduces a novel deep learning approach for hand gesture recognition, leveraging convolutional neural networks (CNNs), long short-term memory networks (LSTMs), and attention mechanisms. CNNs extract spatial features from radar signals, while LSTMs model the temporal dependencies crucial for dynamic gestures. Additionally, attention mechanisms enhance feature selection, ultimately improving recognition performance. We evaluate our method on the UWB-Gestures dataset, which has 12 gestures from eight people that were recorded using three X-Thru X4 UWB impulse radar sensors. Our processing pipeline integrates feature extraction, LSTM-attention blocks, and dense layers for final classification. Early fusion techniques, which combine spatial and temporal features in the initial stages, yield superior results, achieving an overall accuracy of 98.33% and outperforming intermediate fusion methods across gesture classes. To enhance the model’s robustness, we evaluated its performance under common contributors of radar-specific noise scenarios in practical applications, including Gaussian noise, signal inversion, and multipath interference. Our model demonstrates high resilience, maintaining performance despite adverse conditions. As compared to state-of-the-art approaches, our approach delivers competitive accuracy and enhanced robustness, offering a reliable solution for noise-resilient radar-based hand gesture recognition in real-world applications. |
| format | Article |
| id | doaj-art-845aa0b968f8477bbf3d2e7918e2027f |
| institution | DOAJ |
| issn | 2169-3536 |
| language | English |
| publishDate | 2025-01-01 |
| publisher | IEEE |
| record_format | Article |
| series | IEEE Access |
| spelling | doaj-art-845aa0b968f8477bbf3d2e7918e2027f2025-08-20T03:11:04ZengIEEEIEEE Access2169-35362025-01-0113692816929110.1109/ACCESS.2025.355829310950371Radar-Based Hand Gesture Recognition With Feature Fusion Using Robust CNN-LSTM and Attention ArchitectureIrshad Khan0https://orcid.org/0000-0001-6960-2083Young-Woo Kwon1https://orcid.org/0000-0003-0625-8232School of Computer Science and Engineering, Kyungpook National University, Daegu, South KoreaSchool of Computer Science and Engineering, Kyungpook National University, Daegu, South KoreaIn Human-Computer Interaction (HCI), seamless hand gesture recognition is essential for intuitive and natural interactions. Gestures act as a universal language, bridging the gap between humans and machines. Radar-based recognition surpasses traditional optical methods, offering robust interaction capabilities in diverse environments. This article introduces a novel deep learning approach for hand gesture recognition, leveraging convolutional neural networks (CNNs), long short-term memory networks (LSTMs), and attention mechanisms. CNNs extract spatial features from radar signals, while LSTMs model the temporal dependencies crucial for dynamic gestures. Additionally, attention mechanisms enhance feature selection, ultimately improving recognition performance. We evaluate our method on the UWB-Gestures dataset, which has 12 gestures from eight people that were recorded using three X-Thru X4 UWB impulse radar sensors. Our processing pipeline integrates feature extraction, LSTM-attention blocks, and dense layers for final classification. Early fusion techniques, which combine spatial and temporal features in the initial stages, yield superior results, achieving an overall accuracy of 98.33% and outperforming intermediate fusion methods across gesture classes. To enhance the model’s robustness, we evaluated its performance under common contributors of radar-specific noise scenarios in practical applications, including Gaussian noise, signal inversion, and multipath interference. Our model demonstrates high resilience, maintaining performance despite adverse conditions. As compared to state-of-the-art approaches, our approach delivers competitive accuracy and enhanced robustness, offering a reliable solution for noise-resilient radar-based hand gesture recognition in real-world applications.https://ieeexplore.ieee.org/document/10950371/Hand gesture recognitionUWB-sensorsdeep learningfusionnoise robust |
| spellingShingle | Irshad Khan Young-Woo Kwon Radar-Based Hand Gesture Recognition With Feature Fusion Using Robust CNN-LSTM and Attention Architecture IEEE Access Hand gesture recognition UWB-sensors deep learning fusion noise robust |
| title | Radar-Based Hand Gesture Recognition With Feature Fusion Using Robust CNN-LSTM and Attention Architecture |
| title_full | Radar-Based Hand Gesture Recognition With Feature Fusion Using Robust CNN-LSTM and Attention Architecture |
| title_fullStr | Radar-Based Hand Gesture Recognition With Feature Fusion Using Robust CNN-LSTM and Attention Architecture |
| title_full_unstemmed | Radar-Based Hand Gesture Recognition With Feature Fusion Using Robust CNN-LSTM and Attention Architecture |
| title_short | Radar-Based Hand Gesture Recognition With Feature Fusion Using Robust CNN-LSTM and Attention Architecture |
| title_sort | radar based hand gesture recognition with feature fusion using robust cnn lstm and attention architecture |
| topic | Hand gesture recognition UWB-sensors deep learning fusion noise robust |
| url | https://ieeexplore.ieee.org/document/10950371/ |
| work_keys_str_mv | AT irshadkhan radarbasedhandgesturerecognitionwithfeaturefusionusingrobustcnnlstmandattentionarchitecture AT youngwookwon radarbasedhandgesturerecognitionwithfeaturefusionusingrobustcnnlstmandattentionarchitecture |