Affiliation:
1. Carnegie Mellon University, Pittsburgh, PA, USA
2. University of Toronto, Toronto, ON, Canada
Abstract
Despite advances in audio- and motion-based human activity recognition (HAR) systems, a practical, power-efficient, and privacy-sensitive activity recognition system has remained elusive. State-of-the-art activity recognition systems often require power-hungry and privacy-invasive audio data. This is especially challenging for resource-constrained wearables, such as smartwatches. To counter the need for an always-on audio-based activity classification system, we first make use of power and compute-optimized IMUs sampled at 50 Hz to act as a trigger for detecting activity events. Once detected, we use a multimodal deep learning model that augments the motion data with audio data captured on a smartwatch. We subsample this audio to rates ≤ 1 kHz, rendering spoken content unintelligible, while also reducing power consumption on mobile devices. Our multimodal deep learning model achieves a recognition accuracy of 92.2% across 26 daily activities in four indoor environments. Our findings show that subsampling audio from 16 kHz down to 1 kHz, in concert with motion data, does not result in a significant drop in inference accuracy. We also analyze the speech content intelligibility and power requirements of audio sampled at less than 1 kHz and demonstrate that our proposed approach can improve the practicality of human activity recognition systems.
Publisher
Association for Computing Machinery (ACM)
Subject
Computer Networks and Communications,Hardware and Architecture,Human-Computer Interaction
Reference69 articles.
1. Alireza Abedin , Mahsa Ehsanpour , Qinfeng Shi , Hamid Rezatofighi , and Damith C. Ranasinghe . 2021. Attend and Discriminate: Beyond the State-of-the-Art for Human Activity Recognition Using Wearable Sensors . Proc. ACM Interact. Mob. Wearable Ubiquitous Technol. 5, 1, Article 1 (mar 2021 ), 22 pages. https://doi.org/10.1145/3448083 10.1145/3448083 Alireza Abedin, Mahsa Ehsanpour, Qinfeng Shi, Hamid Rezatofighi, and Damith C. Ranasinghe. 2021. Attend and Discriminate: Beyond the State-of-the-Art for Human Activity Recognition Using Wearable Sensors. Proc. ACM Interact. Mob. Wearable Ubiquitous Technol. 5, 1, Article 1 (mar 2021), 22 pages. https://doi.org/10.1145/3448083
2. Pose-on-the-Go: Approximating User Pose with Smartphone Sensor Fusion and Inverse Kinematics
3. VQA: Visual Question Answering
4. CHARM-Deep: Continuous Human Activity Recognition Model Based on Deep Neural Network Using IMU Sensors of Smartwatch
5. Effects of low pass filtering on the intelligibility of speech in noise for people with and without dead regions at high frequencies
Cited by
22 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献
1. Collecting Self-reported Physical Activity and Posture Data Using Audio-based Ecological Momentary Assessment;Proceedings of the ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies;2024-08-22
2. UMAHand: A dataset of inertial signals of typical hand activities;Data in Brief;2024-08
3. TouchTone: Smartwatch Privacy Protection via Unobtrusive Finger Touch Gestures;Proceedings of the 22nd Annual International Conference on Mobile Systems, Applications and Services;2024-06-03
4. The EarSAVAS Dataset;Proceedings of the ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies;2024-05-13
5. EchoWrist: Continuous Hand Pose Tracking and Hand-Object Interaction Recognition Using Low-Power Active Acoustic Sensing On a Wristband;Proceedings of the CHI Conference on Human Factors in Computing Systems;2024-05-11