Abstract
Abstract
Background
Biomedical natural language processing tasks are best performed with English models, and translation tools have undergone major improvements. On the other hand, building annotated biomedical data sets remains a challenge.
Objective
The aim of our study is to determine whether the use of English tools to extract and normalize French medical concepts based on translations provides comparable performance to that of French models trained on a set of annotated French clinical notes.
Methods
We compared 2 methods: 1 involving French-language models and 1 involving English-language models. For the native French method, the named entity recognition and normalization steps were performed separately. For the translated English method, after the first translation step, we compared a 2-step method and a terminology-oriented method that performs extraction and normalization at the same time. We used French, English, and bilingual annotated data sets to evaluate all stages (named entity recognition, normalization, and translation) of our algorithms.
Results
The native French method outperformed the translated English method, with an overall F1-score of 0.51 (95% CI 0.47-0.55), compared with 0.39 (95% CI 0.34-0.44) and 0.38 (95% CI 0.36-0.40) for the 2 English methods tested.
Conclusions
Despite recent improvements in translation models, there is a significant difference in performance between the 2 approaches in favor of the native French method, which is more effective on French medical texts, even with few annotated documents.
Reference46 articles.
1. The Unified Medical Language System (UMLS): integrating biomedical terminology;Bodenreider;Nucleic Acids Res
2. Vaswani A Shazeer N Parmar N et al. Attention is all you need. In: Guyon I von Luxburg U Bengio S et al, editors. Advances in Neural Information Processing Systems 30 (NIPS 2017). 2017. URL: https://papers.nips.cc/paper_files/paper/2017/hash/3f5ee243547dee91fbd053c1c4a845aa-Abstract.html [Accessed 15-03-2024]
3. Devlin J Chang MW Lee K Toutanova K . BERT: pre-training of deep bidirectional transformers for language understanding. In: Burstein J Doran C Solorio T , editors. Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers). Association for Computational Linguistics; 2019;4171-4186. [doi: 10.18653/v1/N19-1423]
4. Clinical natural language processing in languages other than English: opportunities and challenges;Névéol;J Biomed Semantics
5. van Mulligen EM Afzal Z Akhondi SA Vo D Kors JA . Erasmus MC at CLEF Ehealth 2016: concept recognition and coding in French texts. In: Balog K Cappellato L Ferro N Macdonald C , editors. Working Notes of CLEF 2016 - Conference and Labs of the Evaluation Forum CEUR Workshop Proceedings, Vol 1609. CEUR-WS.org; 2016;171-178. URL: https://ceur-ws.org/Vol-1609/16090171.pdf [Accessed 15-03-2024]