Author:
Shickel Benjamin,Silva Brandon,Ozrazgat-Baslanti Tezcan,Ren Yuanfang,Khezeli Kia,Guan Ziyuan,Tighe Patrick J.,Bihorac Azra,Rashidi Parisa
Abstract
Transformer model architectures have revolutionized the natural language processing (NLP) domain and continue to produce state-of-the-art results in text-based applications. Prior to the emergence of transformers, traditional NLP models such as recurrent and convolutional neural networks demonstrated promising utility for patient-level predictions and health forecasting from longitudinal datasets. However, to our knowledge only few studies have explored transformers for predicting clinical outcomes from electronic health record (EHR) data, and in our estimation, none have adequately derived a health-specific tokenization scheme to fully capture the heterogeneity of EHR systems. In this study, we propose a dynamic method for tokenizing both discrete and continuous patient data, and present a transformer-based classifier utilizing a joint embedding space for integrating disparate temporal patient measurements. We demonstrate the feasibility of our clinical AI framework through multi-task ICU patient acuity estimation, where we simultaneously predict six mortality and readmission outcomes. Our longitudinal EHR tokenization and transformer modeling approaches resulted in more accurate predictions compared with baseline machine learning models, which suggest opportunities for future multimodal data integrations and algorithmic support tools using clinical transformer networks.
Funder
National Institute of General Medical Sciences
NIH Office of the Director
National Institute of Diabetes and Digestive and Kidney Diseases
National Institute of Neurological Disorders and Stroke
National Institute of Biomedical Imaging and Bioengineering
National Science Foundation
National Institute of General Medical Science
Cited by
7 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献