ConvBoost
-
Published:2023-06-12
Issue:2
Volume:7
Page:1-21
-
ISSN:2474-9567
-
Container-title:Proceedings of the ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies
-
language:en
-
Short-container-title:Proc. ACM Interact. Mob. Wearable Ubiquitous Technol.
Author:
Shao Shuai1ORCID, Guan Yu1ORCID, Zhai Bing2ORCID, Missier Paolo3ORCID, Plötz Thomas4ORCID
Affiliation:
1. Department of Computer Science, University of Warwick, Coventry, UK 2. Computer and Information Sciences, Northumbria University, Newcastle upon Tyne, UK 3. School of Computing, Newcastle University, Newcastle upon Tyne, UK 4. School of Interactive Computing, Georgia Institute of Technology, Atlanta, USA
Abstract
Human activity recognition (HAR) is one of the core research themes in ubiquitous and wearable computing. With the shift to deep learning (DL) based analysis approaches, it has become possible to extract high-level features and perform classification in an end-to-end manner. Despite their promising overall capabilities, DL-based HAR may suffer from overfitting due to the notoriously small, often inadequate, amounts of labeled sample data that are available for typical HAR applications. In response to such challenges, we propose ConvBoost -- a novel, three-layer, structured model architecture and boosting framework for convolutional network based HAR. Our framework generates additional training data from three different perspectives for improved HAR, aiming to alleviate the shortness of labeled training data in the field. Specifically, with the introduction of three conceptual layers--Sampling Layer, Data Augmentation Layer, and Resilient Layer--we develop three "boosters"--R-Frame, Mix-up, and C-Drop--to enrich the per-epoch training data by dense-sampling, synthesizing, and simulating, respectively. These new conceptual layers and boosters, that are universally applicable for any kind of convolutional network, have been designed based on the characteristics of the sensor data and the concept of frame-wise HAR. In our experimental evaluation on three standard benchmarks (Opportunity, PAMAP2, GOTOV) we demonstrate the effectiveness of our ConvBoost framework for HAR applications based on variants of convolutional networks: vanilla CNN, ConvLSTM, and Attention Models. We achieved substantial performance gains for all of them, which suggests that the proposed approach is generic and can serve as a practical solution for boosting the performance of existing ConvNet-based HAR models. This is an open-source project, and the code can be found at https://github.com/sshao2013/ConvBoost
Publisher
Association for Computing Machinery (ACM)
Subject
Computer Networks and Communications,Hardware and Architecture,Human-Computer Interaction
Reference48 articles.
1. Akin Avci , Stephan Bosch , Mihai Marin-Perianu , Raluca Marin-Perianu , and Paul Havinga . 2010 . Activity recognition using inertial sensing for healthcare, wellbeing and sports applications: A survey . In 23th International conference on architecture of computing systems 2010. VDE, VDE, The Netherlands, 1--10. Akin Avci, Stephan Bosch, Mihai Marin-Perianu, Raluca Marin-Perianu, and Paul Havinga. 2010. Activity recognition using inertial sensing for healthcare, wellbeing and sports applications: A survey. In 23th International conference on architecture of computing systems 2010. VDE, VDE, The Netherlands, 1--10. 2. Adversarial multi-view networks for activity recognition;Bai Lei;Proceedings of the ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies,2020 3. The Opportunity challenge: A benchmark database for on-body sensor-based activity recognition 4. METIER: A deep multi-task learning based activity and user recognition model using wearable sensors;Chen Ling;Proceedings of the ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies,2020 5. Michael R Chernick and Robert A LaBudde . 2014. An introduction to bootstrap methods with applications to R . John Wiley & Sons . Michael R Chernick and Robert A LaBudde. 2014. An introduction to bootstrap methods with applications to R. John Wiley & Sons.
Cited by
4 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献
|
|