1. A comparative study of large language model-based zero-shot inference and task-specific supervised classification of breast cancer pathology reports
- Author
-
Sushil, Madhumita, Zack, Travis, Mandair, Divneet, Zheng, Zhiwei, Wali, Ahmed, Yu, Yan-Ning, Quan, Yuwei, Lituiev, Dmytro, and Butte, Atul J
- Subjects
Information and Computing Sciences ,Machine Learning ,Bioengineering ,Cancer ,Women's Health ,Networking and Information Technology R&D (NITRD) ,Breast Cancer ,2.5 Research design and methodologies (aetiology) ,Humans ,Breast Neoplasms ,Female ,Supervised Machine Learning ,Natural Language Processing ,Datasets as Topic ,Electronic Health Records ,Data Mining ,electronic health records ,large language models ,breast cancer ,pathology ,natural language processing ,Engineering ,Medical and Health Sciences ,Medical Informatics ,Biomedical and clinical sciences ,Health sciences ,Information and computing sciences - Abstract
ObjectiveAlthough supervised machine learning is popular for information extraction from clinical notes, creating large annotated datasets requires extensive domain expertise and is time-consuming. Meanwhile, large language models (LLMs) have demonstrated promising transfer learning capability. In this study, we explored whether recent LLMs could reduce the need for large-scale data annotations.Materials and methodsWe curated a dataset of 769 breast cancer pathology reports, manually labeled with 12 categories, to compare zero-shot classification capability of the following LLMs: GPT-4, GPT-3.5, Starling, and ClinicalCamel, with task-specific supervised classification performance of 3 models: random forests, long short-term memory networks with attention (LSTM-Att), and the UCSF-BERT model.ResultsAcross all 12 tasks, the GPT-4 model performed either significantly better than or as well as the best supervised model, LSTM-Att (average macro F1-score of 0.86 vs 0.75), with advantage on tasks with high label imbalance. Other LLMs demonstrated poor performance. Frequent GPT-4 error categories included incorrect inferences from multiple samples and from history, and complex task design, and several LSTM-Att errors were related to poor generalization to the test set.DiscussionOn tasks where large annotated datasets cannot be easily collected, LLMs can reduce the burden of data labeling. However, if the use of LLMs is prohibitive, the use of simpler models with large annotated datasets can provide comparable results.ConclusionsGPT-4 demonstrated the potential to speed up the execution of clinical NLP studies by reducing the need for large annotated datasets. This may increase the utilization of NLP-based variables and outcomes in clinical studies.
- Published
- 2024