VBCNet: A Hybird Network for Human Activity Recognition

In recent years, the research on human activity recognition based on channel state information (CSI) of Wi-Fi has gradually attracted much attention in order to avoid the deployment of additional devices and reduce the risk of personal privacy leakage. In this paper, we propose a hybrid network arch...

Full description

Saved in:
Bibliographic Details
Main Authors: Fei Ge, Zhenyang Dai, Zhimin Yang, Fei Wu, Liansheng Tan
Format: Article
Language:English
Published: MDPI AG 2024-12-01
Series:Sensors
Subjects:
Online Access:https://www.mdpi.com/1424-8220/24/23/7793
Tags: Add Tag
No Tags, Be the first to tag this record!
Description
Summary:In recent years, the research on human activity recognition based on channel state information (CSI) of Wi-Fi has gradually attracted much attention in order to avoid the deployment of additional devices and reduce the risk of personal privacy leakage. In this paper, we propose a hybrid network architecture, named VBCNet, that can effectively identify human activity postures. Firstly, we extract CSI sequences from each antenna of Wi-Fi signals, and the data are preprocessed and tokenised. Then, in the encoder part of the model, we introduce a layer of long short-term memory network to further extract the temporal features in the sequences and enhance the ability of the model to capture the temporal information. Meanwhile, VBCNet employs a convolutional feed-forward network instead of the traditional feed-forward network to enhance the model’s ability to process local and multi-scale features. Finally, the model classifies the extracted features into human behaviours through a classification layer. To validate the effectiveness of VBCNet, we conducted experimental evaluations on the classical human activity recognition datasets UT-HAR and Widar3.0 and achieved an accuracy of 98.65% and 77.92%. These results show that VBCNet exhibits extremely high effectiveness and robustness in human activity recognition tasks in complex scenarios.
ISSN:1424-8220