Author:
Yagimoto Kanami,Hosoda Shion,Sato Miwa,Hamada Michiaki
Abstract
AbstractMotivationAntibiotic resistance has emerged as a major global health threat, with an increasing number of bacterial infections becoming difficult to treat. Predicting the underlying resistance mechanisms of antibiotic resistance genes (ARGs) is crucial for understanding and combating this problem. However, existing methods struggle to accurately predict resistance mechanisms for ARGs with low similarity to known sequences and lack sufficient interpretability of the prediction models.ResultsIn this study, we present a novel approach for predicting ARG resistance mechanisms using Protein-BERT, a protein language model based on deep learning. Our method outperforms state-of-the-art techniques on diverse ARG datasets, including those with low homology to the training data, highlighting its potential for predicting the resistance mechanisms of unknown ARGs. Attention analysis of the model reveals that it considers biologically relevant features, such as conserved amino acid residues and antibiotic target binding sites, when making predictions. These findings provide valuable insights into the molecular basis of antibiotic resistance and demonstrate the interpretability of protein language models, offering a new perspective on their application in bioinformatics.AvailabilityThe source code is available for free athttps://github.com/hmdlab/ARG-BERT. The output results of the model are published athttps://waseda.box.com/v/ARG-BERT-suppl.Contactmhamada@waseda.jp
Publisher
Cold Spring Harbor Laboratory