Affiliation:
1. University of Massachusetts Amherst
2. University of Massachusetts Lowell
Abstract
Abstract
Deep learning transformer-based models using longitudinal electronic health records (EHRs) have shown a great success in prediction of clinical diseases or outcomes. Pretraining on a large dataset can help such models map the input space better and boost their performance on relevant tasks through finetuning with limited data. In this study, we present Enformer, a generative encoder-decoder model with transformer that was pretrained using a new pretraining objective - predicting all diseases and outcomes of a patient at a future visit from previous visits. Enformer’s encoder-decoder framework, paired with the novel pretraining objective, helped it achieve the new state-of-the-art (SOTA) performance on multiple clinical prediction tasks. Comparing with the previous SOTA model, Enformer improved area under the precision–recall curve (AUPRC) by 2% (p<0.001) for pancreatic cancer onset and by 24% (p<0.001) for intentional self-harm in patients with PTSD. The high performance in predicting intentional self-harm shows the potential of Enformer in building effective clinical intervention systems. Enformer is also generalizable and can be easily finetuned for clinical prediction tasks with limited data.
Publisher
Research Square Platform LLC
Reference54 articles.
1. 1. Kessler RC, Bauer MS, Bishop TM, et al. Using Administrative Data to Predict Suicide After Psychiatric Hospitalization in the Veterans Health Administration System. Front Psychiatry. 2020;11:390. doi:10.3389/fpsyt.2020.00390
2. 2. Zhao W, Jiang W, Qiu X. Deep learning for COVID-19 detection based on CT images. Sci Rep. 2021;11.
3. 3. Goh KH, Wang L, Yeow AYK, et al. Artificial intelligence in sepsis early prediction and diagnosis using unstructured data in healthcare. Nat Commun. 2021;12.
4. 4. Choi E, Bahadori MT, Sun J, Kulas J, Schuetz A, Stewart W. RETAIN: An Interpretable Predictive Model for Healthcare using Reverse Time Attention Mechanism. In: Advances in Neural Information Processing Systems. ; 2016:3504–3512. Accessed January 12, 2017. http://papers.nips.cc/paper/6321-retain-an-interpretable-predictive-model-for-healthcare-using-reverse-time-attention-mechanism
5. 5. Wu M, Hughes MC, Parbhoo S, Zazzi M, Roth V, Doshi-Velez F. Beyond Sparsity: Tree Regularization of Deep Models for Interpretability. In: AAAI Conference on Artificial Intelligence. ; 2017.