Photoplethysmography Signal Quality Assessment using Attentive-CNN Models
Author:
Silva Leonardo,Lima Rafael,Lucafo Giovani,Sandoval Italo,Freitas Pedro Garcia,Penatti Otávio A. B.
Abstract
Due to the rapid popularization of wearable computers such as smartwatches, Health Monitoring Applications (HMA) are becoming increasingly popular because of their capability to track different health indicators, including sleep patterns, heart rate, and activity tracking movements. These applications usually employ Photoplethysmography (PPG) sensors to monitor various aspects of an individual’s health and well-being. PPG is a non-invasive and cost-effective optical technique based on the detection of blood volume changes in the microvascular bed of tissue, capturing the dynamic physiological changes in the body with continuous measurements taken over time. Analyzing PPG as a time series enables the extraction of meaningful information about cardiovascular health and other physiological parameters, such as Heart Rate Variability (HRV), Peripheral Oxygen Saturation (SpO2), and sleep status. To enable reliable health indicators, it is important to have robustly sampled PPG signals. However, in practice, the PPG signal is often corrupted with different types of noise and artifacts due to motion, especially in scenarios where wearables are used. Therefore, Signal Quality Assessment (SQA) plays a fundamental role in determining the reliability of a given PPG for use in HMA. Considering this, in this work, we propose a novel PPG SQA method focused on the balance between storage size and classifier quality, aiming to achieve a lightweight and robust model. This model is developed using recent advances in attention-based strategies to significantly improve the performance of purely Convolutional Neural Network (CNN)-based SQA classifiers.
Publisher
Sociedade Brasileira de Computação - SBC
Reference28 articles.
1. Azar, J., Makhoul, A., Couturier, R., and Demerjian, J. (2021). Deep recurrent neural network-based autoencoder for photoplethysmogram artifacts filtering. Computers & Electrical Engineering, 92:107065. 2. Bahdanau, D., Cho, K., and Bengio, Y. (2015). Neural machine translation by jointly learning to align and translate. 3rd International Conference on Learning Representations, ICLR 2015. 3. Brown, T., Mann, B., Ryder, N., Subbiah, M., Kaplan, J. D., Dhariwal, P., Neelakantan, A., Shyam, P., Sastry, G., Askell, A., Agarwal, S., Herbert-Voss, A., Krueger, G., Henighan, T., Child, R., Ramesh, A., Ziegler, D., Wu, J., Winter, C., Hesse, C., Chen, M., Sigler, E., Litwin, M., Gray, S., Chess, B., Clark, J., Berner, C., McCandlish, S., Radford, A., Sutskever, I., and Amodei, D. (2020). Language models are few-shot learners. In Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M., and Lin, H., editors, Advances in Neural Information Processing Systems, volume 33, pages 1877–1901. Curran Associates, Inc. 4. Chatterjee, T., Ghosh, A., and Sarkar, S. (2022). Signal quality assessment of photoplethysmogram signals using quantum pattern recognition technique and lightweight cnn module. In International Conference of the IEEE Engineering in Medicine & Biology Society (EMBC), pages 3382–3386. 5. Chong, J. W., Dao, D. K., Salehizadeh, S., McManus, D. D., Darling, C. E., Chon, K. H., and Mendelson, Y. (2014). Photoplethysmograph signal reconstruction based on a novel hybrid motion artifact detection–reduction approach. part i: Motion and noise artifact detection. Annals of biomedical engineering, 42:2238–2250.
|
|