1. Multiscale Residual Weighted Classification Network for Human Activity Recognition in Microwave Radar
- Author
-
Yukun Gao, Lin Cao, Zongmin Zhao, Dongfeng Wang, Chong Fu, and Yanan Guo
- Subjects
deep learning (DL) ,human activity recognition (HAR) ,contrastive learning ,radar micro-Doppler signatures ,time-Doppler images ,Chemical technology ,TP1-1185 - Abstract
Human activity recognition by radar sensors plays an important role in healthcare and smart homes. However, labeling a large number of radar datasets is difficult and time-consuming, and it is difficult for models trained on insufficient labeled data to obtain exact classification results. In this paper, we propose a multiscale residual weighted classification network with large-scale, medium-scale, and small-scale residual networks. Firstly, an MRW image encoder is used to extract salient feature representations from all time-Doppler images through contrastive learning. This can extract the representative vector of each image and also obtain the pre-training parameters of the MRW image encoder. During the pre-training process, large-scale residual networks, medium-scale residual networks, and small-scale residual networks are used to extract global information, texture information, and semantic information, respectively. Moreover, the time–channel weighting mechanism can allocate weights to important time and channel dimensions to achieve more effective extraction of feature information. The model parameters obtained from pre-training are frozen, and the classifier is added to the backend. Finally, the classifier is fine-tuned using a small amount of labeled data. In addition, we constructed a new dataset with eight dangerous activities. The proposed MRW-CN model was trained on this dataset and achieved a classification accuracy of 96.9%. We demonstrated that our method achieves state-of-the-art performance. The ablation analysis also demonstrated the role of multi-scale convolutional kernels and time–channel weighting mechanisms in classification.
- Published
- 2025
- Full Text
- View/download PDF