Affiliation:
1. Changzhou University, China
Abstract
The latest advancement in human activity recognition (HAR) involves the use of deep neural networks to achieve greater accuracy in the classification of various activities. A popular approach in the field is to encode time series data from inertial sensors into images and then apply techniques from computer vision to analyze the data. However, encoding into images often leads to a significant surge in the amount of data and a subsequent rise in computational cost, making this method less efficient for real-world applications. In this paper, we propose a novel image-coding approach, alternating sampling amplitude-phase field (AS-APF), and a multi-sensor fusion framework based on selective kernel (SK). AS-APF can reduce the amount of image data while ensuring the integrity and representativeness of the data. Because it splits the time series and preserves the main feature information. We introduce SK to learn multi-scale features in HAR instead of a fixed receptive fields (RFs) size. Our experimental results demonstrate that our approach outperforms previous encoding methods in both accuracy and time efficiency.
Funder
Key Laboratory of Brain Machine Collaborative Intelligence Foundation of Zhejiang Province
Changzhou Sci & Tech Program
The Project of Jiangsu Key Research and Development Plan
Jiangsu Agriculture Science and Technology Innovation Fund
Reference45 articles.
1. Abdel-Salam R, Mostafa R, Hadhood M (2021) Human activity recognition using wearable sensors: Review, challenges, evaluation benchmark. ArXiv [online] abs/2101.01665. Available at: https://arxiv.org/abs/2101.01665 (accessed 2 August 2024).
2. Inertial Sensor Data to Image Encoding for Human Action Recognition
3. Deep Learning Based Real-time Daily Human Activity Recognition and Its Implementation in a Smartphone
4. Anguita D, Ghio A, Oneto L, et al. (2013) A public domain dataset for human activity recognition using smartphones. In: Proceedings of the Esann, Bruges, 24–26 April 2013.
5. UTD-MHAD: A multimodal dataset for human action recognition utilizing a depth camera and a wearable inertial sensor