1. Developing a Clinical Language Model for Swedish : Continued Pretraining of Generic BERT with In-Domain Data
- Author
-
Anastasios Lamproudis, Aron Henriksson, and Hercules Dalianis
- Subjects
Vocabulary ,Computer and Information Sciences ,Downstream (software development) ,Computer science ,business.industry ,media_common.quotation_subject ,Data- och informationsvetenskap ,computer.software_genre ,Task (project management) ,Domain (software engineering) ,clinical text ,language models ,Added value ,Language model ,Diagnosis code ,Artificial intelligence ,natural language processing ,business ,computer ,Natural language processing ,Protected health information ,media_common - Abstract
The use of pretrained language models, fine-tuned to perform a specific downstream task, has become widespread in NLP. Using a generic language model in specialized domains may, however, be sub-optimal due to differences in language use and vocabulary. In this paper, it is investigated whether an existing, generic language model for Swedish can be improved for the clinical domain through continued pretraining with clinical text. The generic and domain-specific language models are fine-tuned and evaluated on three representative clinical NLP tasks: (i) identifying protected health information, (ii) assigning ICD-10 diagnosis codes to discharge summaries, and (iii) sentence-level uncertainty prediction. The results show that continued pretraining on in-domain data leads to improved performance on all three downstream tasks, indicating that there is a potential added value of domain-specific language models for clinical NLP.
- Published
- 2021