254 results on '"Verspoor, K"'
Search Results
2. Electronic Health Records for Predicting Outcomes to Work-Related Musculoskeletal Disorders: A Scoping Review
- Author
-
Wassell, M., Vitiello, A., Butler-Henderson, K., Verspoor, K., McCann, P., and Pollard, H.
- Published
- 2024
- Full Text
- View/download PDF
3. Automating Quality Assessment of Medical Evidence in Systematic Reviews: Model Development and Validation Study
- Author
-
Suster, S, Baldwin, T, Lau, JH, Yepes, AJ, Iraola, DM, Otmakhova, Y, Verspoor, K, Suster, S, Baldwin, T, Lau, JH, Yepes, AJ, Iraola, DM, Otmakhova, Y, and Verspoor, K
- Abstract
BACKGROUND: Assessment of the quality of medical evidence available on the web is a critical step in the preparation of systematic reviews. Existing tools that automate parts of this task validate the quality of individual studies but not of entire bodies of evidence and focus on a restricted set of quality criteria. OBJECTIVE: We proposed a quality assessment task that provides an overall quality rating for each body of evidence (BoE), as well as finer-grained justification for different quality criteria according to the Grading of Recommendation, Assessment, Development, and Evaluation formalization framework. For this purpose, we constructed a new data set and developed a machine learning baseline system (EvidenceGRADEr). METHODS: We algorithmically extracted quality-related data from all summaries of findings found in the Cochrane Database of Systematic Reviews. Each BoE was defined by a set of population, intervention, comparison, and outcome criteria and assigned a quality grade (high, moderate, low, or very low) together with quality criteria (justification) that influenced that decision. Different statistical data, metadata about the review, and parts of the review text were extracted as support for grading each BoE. After pruning the resulting data set with various quality checks, we used it to train several neural-model variants. The predictions were compared against the labels originally assigned by the authors of the systematic reviews. RESULTS: Our quality assessment data set, Cochrane Database of Systematic Reviews Quality of Evidence, contains 13,440 instances, or BoEs labeled for quality, originating from 2252 systematic reviews published on the internet from 2002 to 2020. On the basis of a 10-fold cross-validation, the best neural binary classifiers for quality criteria detected risk of bias at 0.78 F1 (P=.68; R=0.92) and imprecision at 0.75 F1 (P=.66; R=0.86), while the performance on inconsistency, indirectness, and publication bias criteria was l
- Published
- 2023
4. Detecting evidence of invasive fungal infections in cytology and histopathology reports enriched with concept-level annotations
- Author
-
Rozova, V, Khanina, A, Teng, JC, Teh, JSK, Worth, LJ, Slavin, MA, Thursky, KA, Verspoor, K, Rozova, V, Khanina, A, Teng, JC, Teh, JSK, Worth, LJ, Slavin, MA, Thursky, KA, and Verspoor, K
- Abstract
Invasive fungal infections (IFIs) are particularly dangerous to high-risk patients with haematological malignancies and are responsible for excessive mortality and delays in cancer therapy. Surveillance of IFI in clinical settings offers an opportunity to identify potential risk factors and evaluate new therapeutic strategies. However, manual surveillance is both time- and resource-intensive. As part of a broader project aimed to develop a system for automated IFI surveillance by leveraging electronic medical records, we present our approach to detecting evidence of IFI in the key diagnostic domain of histopathology. Using natural language processing (NLP), we analysed cytology and histopathology reports to identify IFI-positive reports. We compared a conventional bag-of-words classification model to a method that relies on concept-level annotations. Although the investment to prepare data supporting concept annotations is substantial, extracting targeted information specific to IFI as a pre-processing step increased the performance of the classifier from the PR AUC of 0.84 to 0.92 and enabled model interpretability. We have made publicly available the annotated dataset of 283 reports, the Cytology and Histopathology IFI Reports corpus (CHIFIR), to allow the clinical NLP research community to further build on our results.
- Published
- 2023
5. Analysis of predictive performance and reliability of classifiers for quality assessment of medical evidence revealed important variation by medical area
- Author
-
Suster, S, Baldwin, T, Verspoor, K, Suster, S, Baldwin, T, and Verspoor, K
- Abstract
OBJECTIVES: A major obstacle in deployment of models for automated quality assessment is their reliability. To analyze their calibration and selective classification performance. STUDY DESIGN AND SETTING: We examine two systems for assessing the quality of medical evidence, EvidenceGRADEr and RobotReviewer, both developed from Cochrane Database of Systematic Reviews (CDSR) to measure strength of bodies of evidence and risk of bias (RoB) of individual studies, respectively. We report their calibration error and Brier scores, present their reliability diagrams, and analyze the risk-coverage trade-off in selective classification. RESULTS: The models are reasonably well calibrated on most quality criteria (expected calibration error [ECE] 0.04-0.09 for EvidenceGRADEr, 0.03-0.10 for RobotReviewer). However, we discover that both calibration and predictive performance vary significantly by medical area. This has ramifications for the application of such models in practice, as average performance is a poor indicator of group-level performance (e.g., health and safety at work, allergy and intolerance, and public health see much worse performance than cancer, pain, and anesthesia, and Neurology). We explore the reasons behind this disparity. CONCLUSION: Practitioners adopting automated quality assessment should expect large fluctuations in system reliability and predictive performance depending on the medical area. Prospective indicators of such behavior should be further researched.
- Published
- 2023
6. Stratification of keratoconus progression using unsupervised machine learning analysis of tomographical parameters
- Author
-
Cao, K, Verspoor, K, Chan, E, Daniell, M, Sahebjada, S, Baird, PN, Cao, K, Verspoor, K, Chan, E, Daniell, M, Sahebjada, S, and Baird, PN
- Abstract
Purpose: This study aimed to stratify eyes with keratoconus (KC) based on longitudinal changes in all Pentacam parameters into clusters using unsupervised machine learning, with the broader objective of more clearly defining the characteristics of KC progression. Methods: A data-driven cluster analysis (hierarchical clustering) was undertaken on a retrospective cohort of 1017 kC eyes and 128 control eyes. Clusters were derived using 6-month tomographical change in individual eyes from analysis of the reduced dimensionality parameter space using all available Pentacam parameters (406 principal components). The optimal number of clusters was determined by the clustering's capacity to discriminate progression between KC and control eyes based on change across parameters. One-way ANOVA was used to compare parameters between inferred clusters. Complete Pentacam data changes at 6, 12 and 18-month time points provided validation datasets to determine the generalizability of the clustering model. Results: We identified three clusters in KC progression patterns. Eyes designated within cluster 3 had the most rapidly changing tomographical parameters compared to eyes in either cluster 1 or 2. Eyes designated within cluster 1 reflected minimal changes in tomographical parameters, closest to the tomographical changes of control (non-KC) eyes. Thirty-nine corneal curvature parameters were identified and associated with these stratified clusters, with each of these parameters changing significantly different between three clusters. Similar clusters were identified at the 6, 12 and 18-month follow-up. Conclusions: The clustering model developed was able to automatically detect and categorize KC tomographical features into fast, slow, or limited change at different time points. This new KC stratification tool may provide an opportunity to provide a precision medicine approach to KC.
- Published
- 2023
7. Early prediction of incident liver disease using conventional risk factors and gut-microbiome-augmented gradient boosting
- Author
-
Liu, Y, Meric, G, Havulinna, AS, Teo, SM, Aberg, F, Ruuskanen, M, Sanders, J, Zhu, Q, Tripathi, A, Verspoor, K, Cheng, S, Jain, M, Jousilahti, P, Vazquez-Baeza, Y, Loomba, R, Lahti, L, Niiranen, T, Salomaa, V, Knight, R, Inouye, M, Liu, Y, Meric, G, Havulinna, AS, Teo, SM, Aberg, F, Ruuskanen, M, Sanders, J, Zhu, Q, Tripathi, A, Verspoor, K, Cheng, S, Jain, M, Jousilahti, P, Vazquez-Baeza, Y, Loomba, R, Lahti, L, Niiranen, T, Salomaa, V, Knight, R, and Inouye, M
- Abstract
The gut microbiome has shown promise as a predictive biomarker for various diseases. However, the potential of gut microbiota for prospective risk prediction of liver disease has not been assessed. Here, we utilized shallow shotgun metagenomic sequencing of a large population-based cohort (N > 7,000) with ∼15 years of follow-up in combination with machine learning to investigate the predictive capacity of gut microbial predictors individually and in conjunction with conventional risk factors for incident liver disease. Separately, conventional and microbial factors showed comparable predictive capacity. However, microbiome augmentation of conventional risk factors using machine learning significantly improved the performance. Similarly, disease-free survival analysis showed significantly improved stratification using microbiome-augmented models. Investigation of predictive microbial signatures revealed previously unknown taxa for liver disease, as well as those previously associated with hepatic function and disease. This study supports the potential clinical validity of gut metagenomic sequencing to complement conventional risk factors for prediction of liver diseases.
- Published
- 2022
8. COVID-19 Drug Repurposing: A Network-Based Framework for Exploring Biomedical Literature and Clinical Trials for Possible Treatments
- Author
-
Hamed, AA, Fandy, TE, Tkaczuk, KL, Verspoor, K, Lee, BS, Hamed, AA, Fandy, TE, Tkaczuk, KL, Verspoor, K, and Lee, BS
- Abstract
BACKGROUND: With the Coronavirus becoming a new reality of our world, global efforts continue to seek answers to many questions regarding the spread, variants, vaccinations, and medications. Particularly, with the emergence of several strains (e.g., Delta, Omicron), vaccines will need further development to offer complete protection against the new variants. It is critical to identify antiviral treatments while the development of vaccines continues. In this regard, the repurposing of already FDA-approved drugs remains a major effort. In this paper, we investigate the hypothesis that a combination of FDA-approved drugs may be considered as a candidate for COVID-19 treatment if (1) there exists an evidence in the COVID-19 biomedical literature that suggests such a combination, and (2) there is match in the clinical trials space that validates this drug combination. METHODS: We present a computational framework that is designed for detecting drug combinations, using the following components (a) a Text-mining module: to extract drug names from the abstract section of the biomedical publications and the intervention/treatment sections of clinical trial records. (b) a network model constructed from the drug names and their associations, (c) a clique similarity algorithm to identify candidate drug treatments. RESULT AND CONCLUSIONS: Our framework has identified treatments in the form of two, three, or four drug combinations (e.g., hydroxychloroquine, doxycycline, and azithromycin). The identifications of the various treatment candidates provided sufficient evidence that supports the trustworthiness of our hypothesis.
- Published
- 2022
9. Accuracy of Machine Learning Assisted Detection of Keratoconus: A Systematic Review and Meta-Analysis
- Author
-
Cao, K, Verspoor, K, Sahebjada, S, Baird, PN, Cao, K, Verspoor, K, Sahebjada, S, and Baird, PN
- Abstract
(1) Background: The objective of this review was to synthesize available data on the use of machine learning to evaluate its accuracy (as determined by pooled sensitivity and specificity) in detecting keratoconus (KC), and measure reporting completeness of machine learning models in KC based on TRIPOD (the transparent reporting of multivariable prediction models for individual prognosis or diagnosis) statement. (2) Methods: Two independent reviewers searched the electronic databases for all potential articles on machine learning and KC published prior to 2021. The TRIPOD 29-item checklist was used to evaluate the adherence to reporting guidelines of the studies, and the adherence rate to each item was computed. We conducted a meta-analysis to determine the pooled sensitivity and specificity of machine learning models for detecting KC. (3) Results: Thirty-five studies were included in this review. Thirty studies evaluated machine learning models for detecting KC eyes from controls and 14 studies evaluated machine learning models for detecting early KC eyes from controls. The pooled sensitivity for detecting KC was 0.970 (95% CI 0.949-0.982), with a pooled specificity of 0.985 (95% CI 0.971-0.993), whereas the pooled sensitivity of detecting early KC was 0.882 (95% CI 0.822-0.923), with a pooled specificity of 0.947 (95% CI 0.914-0.967). Between 3% and 48% of TRIPOD items were adhered to in studies, and the average (median) adherence rate for a single TRIPOD item was 23% across all studies. (4) Conclusions: Application of machine learning model has the potential to make the diagnosis and monitoring of KC more efficient, resulting in reduced vision loss to the patients. This review provides current information on the machine learning models that have been developed for detecting KC and early KC. Presently, the machine learning models performed poorly in identifying early KC from control eyes and many of these research studies did not follow established reporting standa
- Published
- 2022
10. Large-scale protein-protein post-translational modification extraction with distant supervision and confidence calibrated BioBERT
- Author
-
Elangovan, A, Li, Y, Pires, DE, Davis, MJ, Verspoor, K, Elangovan, A, Li, Y, Pires, DE, Davis, MJ, and Verspoor, K
- Abstract
MOTIVATION: Protein-protein interactions (PPIs) are critical to normal cellular function and are related to many disease pathways. A range of protein functions are mediated and regulated by protein interactions through post-translational modifications (PTM). However, only 4% of PPIs are annotated with PTMs in biological knowledge databases such as IntAct, mainly performed through manual curation, which is neither time- nor cost-effective. Here we aim to facilitate annotation by extracting PPIs along with their pairwise PTM from the literature by using distantly supervised training data using deep learning to aid human curation. METHOD: We use the IntAct PPI database to create a distant supervised dataset annotated with interacting protein pairs, their corresponding PTM type, and associated abstracts from the PubMed database. We train an ensemble of BioBERT models-dubbed PPI-BioBERT-x10-to improve confidence calibration. We extend the use of ensemble average confidence approach with confidence variation to counteract the effects of class imbalance to extract high confidence predictions. RESULTS AND CONCLUSION: The PPI-BioBERT-x10 model evaluated on the test set resulted in a modest F1-micro 41.3 (P =5 8.1, R = 32.1). However, by combining high confidence and low variation to identify high quality predictions, tuning the predictions for precision, we retained 19% of the test predictions with 100% precision. We evaluated PPI-BioBERT-x10 on 18 million PubMed abstracts and extracted 1.6 million (546507 unique PTM-PPI triplets) PTM-PPI predictions, and filter [Formula: see text] (4584 unique) high confidence predictions. Of the 5700, human evaluation on a small randomly sampled subset shows that the precision drops to 33.7% despite confidence calibration and highlights the challenges of generalisability beyond the test set even with confidence calibration. We circumvent the problem by only including predictions associated with multiple papers, improving the precision to 5
- Published
- 2022
11. What does it take to bake a cake? The RecipeRef corpus and anaphora resolution in procedural text
- Author
-
Fang, B, Baldwin, T, Verspoor, K, Fang, B, Baldwin, T, and Verspoor, K
- Published
- 2022
12. Cross-linguistic comparison of linguistic feature encoding in BERT models for typologically different languages
- Author
-
Otmakhova, Y, Verspoor, K, Lau, JH, Otmakhova, Y, Verspoor, K, and Lau, JH
- Abstract
Though recently there have been an increased interest in how pre-trained language models encode different linguistic features, there is still a lack of systematic comparison between languages with different morphology and syntax. In this paper, using BERT as an example of a pre-trained model, we compare how three typologically different languages (English, Korean, and Russian) encode morphology and syntax features across different layers. In particular, we contrast languages which differ in a particular aspect, such as flexibility of word order, head directionality, morphological type, presence of grammatical gender, and morphological richness, across four different tasks.
- Published
- 2022
13. Tasks as needs: reframing the paradigm of clinical natural language processing research for real-world decision support
- Author
-
Lederman, A, Lederman, R, Verspoor, K, Lederman, A, Lederman, R, and Verspoor, K
- Abstract
Electronic medical records are increasingly used to store patient information in hospitals and other clinical settings. There has been a corresponding proliferation of clinical natural language processing (cNLP) systems aimed at using text data in these records to improve clinical decision-making, in comparison to manual clinician search and clinical judgment alone. However, these systems have delivered marginal practical utility and are rarely deployed into healthcare settings, leading to proposals for technical and structural improvements. In this paper, we argue that this reflects a violation of Friedman's "Fundamental Theorem of Biomedical Informatics," and that a deeper epistemological change must occur in the cNLP field, as a parallel step alongside any technical or structural improvements. We propose that researchers shift away from designing cNLP systems independent of clinical needs, in which cNLP tasks are ends in themselves-"tasks as decisions"-and toward systems that are directly guided by the needs of clinicians in realistic decision-making contexts-"tasks as needs." A case study example illustrates the potential benefits of developing cNLP systems that are designed to more directly support clinical needs.
- Published
- 2022
14. Uncertainty Estimation and Reduction of Pre-trained Models for Text Regression
- Author
-
Wang, Y, Beck, D, Baldwin, T, Verspoor, K, Wang, Y, Beck, D, Baldwin, T, and Verspoor, K
- Abstract
State-of-the-art classification and regression models are often not well calibrated, and cannot reliably provide uncertainty estimates, limiting their utility in safety-critical applications such as clinical decision-making. While recent work has focused on calibration of classifiers, there is almost no work in NLP on calibration in a regression setting. In this paper, we quantify the calibration of pretrained language models for text regression, both intrinsically and extrinsically. We further apply uncertainty estimates to augment training data in low-resource domains. Our experiments on three regression tasks in both self-training and active-learning settings show that uncertainty estimation can be used to increase overall performance and enhance model generalization.
- Published
- 2022
15. Exploring automatic inconsistency detection for literature-based gene ontology annotation
- Author
-
Chen, J, Goudey, B, Zobel, J, Geard, N, Verspoor, K, Chen, J, Goudey, B, Zobel, J, Geard, N, and Verspoor, K
- Abstract
MOTIVATION: Literature-based gene ontology annotations (GOA) are biological database records that use controlled vocabulary to uniformly represent gene function information that is described in the primary literature. Assurance of the quality of GOA is crucial for supporting biological research. However, a range of different kinds of inconsistencies in between literature as evidence and annotated GO terms can be identified; these have not been systematically studied at record level. The existing manual-curation approach to GOA consistency assurance is inefficient and is unable to keep pace with the rate of updates to gene function knowledge. Automatic tools are therefore needed to assist with GOA consistency assurance. This article presents an exploration of different GOA inconsistencies and an early feasibility study of automatic inconsistency detection. RESULTS: We have created a reliable synthetic dataset to simulate four realistic types of GOA inconsistency in biological databases. Three automatic approaches are proposed. They provide reasonable performance on the task of distinguishing the four types of inconsistency and are directly applicable to detect inconsistencies in real-world GOA database records. Major challenges resulting from such inconsistencies in the context of several specific application settings are reported. This is the first study to introduce automatic approaches that are designed to address the challenges in current GOA quality assurance workflows. The data underlying this article are available in Github at https://github.com/jiyuc/AutoGOAConsistency.
- Published
- 2022
16. The patient is more dead than alive: exploring the current state of the multi-document summarization of the biomedical literature
- Author
-
Otmakhova, Y, Verspoor, K, Baldwin, T, Lau, JH, Otmakhova, Y, Verspoor, K, Baldwin, T, and Lau, JH
- Published
- 2022
17. Overcoming challenges in extracting prescribing habits from veterinary clinics using big data and deep learning
- Author
-
Hur, B, Hardefeldt, LY, Verspoor, K, Baldwin, T, Gilkerson, JR, Hur, B, Hardefeldt, LY, Verspoor, K, Baldwin, T, and Gilkerson, JR
- Abstract
Understanding antimicrobial usage patterns and encouraging appropriate antimicrobial usage is a critical component of antimicrobial stewardship. Studies using VetCompass Australia and Natural Language Processing (NLP) have demonstrated antimicrobial usage patterns in companion animal practices across Australia. Doing so has highlighted the many obstacles and barriers to the task of converting raw clinical notes into a format that can be readily queried and analysed. We developed NLP systems using rules-based algorithms and machine learning to automate the extraction of data describing the key elements to assess appropriate antimicrobial use. These included the clinical indication, antimicrobial agent selection, dose and duration of therapy. Our methods were applied to over 4.4 million companion animal clinical records across Australia on all consultations with antimicrobial use to help us understand what antibiotics are being given and why on a population level. Of these, approximately only 40% recorded the reason why antimicrobials were prescribed, along with the dose and duration of treatment. NLP and deep learning might be able to overcome the difficulties of harvesting free text data from clinical records, but when the essential data are not recorded in the clinical records, then, this becomes an insurmountable obstacle.
- Published
- 2022
18. Propagation, detection and correction of errors using the sequence database network
- Author
-
Goudey, B, Geard, N, Verspoor, K, Zobel, J, Goudey, B, Geard, N, Verspoor, K, and Zobel, J
- Abstract
Nucleotide and protein sequences stored in public databases are the cornerstone of many bioinformatics analyses. The records containing these sequences are prone to a wide range of errors, including incorrect functional annotation, sequence contamination and taxonomic misclassification. One source of information that can help to detect errors are the strong interdependency between records. Novel sequences in one database draw their annotations from existing records, may generate new records in multiple other locations and will have varying degrees of similarity with existing records across a range of attributes. A network perspective of these relationships between sequence records, within and across databases, offers new opportunities to detect-or even correct-erroneous entries and more broadly to make inferences about record quality. Here, we describe this novel perspective of sequence database records as a rich network, which we call the sequence database network, and illustrate the opportunities this perspective offers for quantification of database quality and detection of spurious entries. We provide an overview of the relevant databases and describe how the interdependencies between sequence records across these databases can be exploited by network analyses. We review the process of sequence annotation and provide a classification of sources of error, highlighting propagation as a major source. We illustrate the value of a network perspective through three case studies that use network analysis to detect errors, and explore the quality and quantity of critical relationships that would inform such network analyses. This systematic description of a network perspective of sequence database records provides a novel direction to combat the proliferation of errors within these critical bioinformatics resources.
- Published
- 2022
19. Predicting Publication of Clinical Trials Using Structured and Unstructured Data: Model Development and Validation Study.
- Author
-
Wang, S, Šuster, S, Baldwin, T, Verspoor, K, Wang, S, Šuster, S, Baldwin, T, and Verspoor, K
- Abstract
BACKGROUND: Publication of registered clinical trials is a critical step in the timely dissemination of trial findings. However, a significant proportion of completed clinical trials are never published, motivating the need to analyze the factors behind success or failure to publish. This could inform study design, help regulatory decision-making, and improve resource allocation. It could also enhance our understanding of bias in the publication of trials and publication trends based on the research direction or strength of the findings. Although the publication of clinical trials has been addressed in several descriptive studies at an aggregate level, there is a lack of research on the predictive analysis of a trial's publishability given an individual (planned) clinical trial description. OBJECTIVE: We aimed to conduct a study that combined structured and unstructured features relevant to publication status in a single predictive approach. Established natural language processing techniques as well as recent pretrained language models enabled us to incorporate information from the textual descriptions of clinical trials into a machine learning approach. We were particularly interested in whether and which textual features could improve the classification accuracy for publication outcomes. METHODS: In this study, we used metadata from ClinicalTrials.gov (a registry of clinical trials) and MEDLINE (a database of academic journal articles) to build a data set of clinical trials (N=76,950) that contained the description of a registered trial and its publication outcome (27,702/76,950, 36% published and 49,248/76,950, 64% unpublished). This is the largest data set of its kind, which we released as part of this work. The publication outcome in the data set was identified from MEDLINE based on clinical trial identifiers. We carried out a descriptive analysis and predicted the publication outcome using 2 approaches: a neural network with a large domain-specific language mo
- Published
- 2022
20. Overview of ChEMU 2022 Evaluation Campaign: Information Extraction in Chemical Patents
- Author
-
Barron-Cedeno, A, DaSanMartino, G, Esposti, MD, Sebastiani, F, Macdonald, C, Pasi, G, Hanbury, A, Potthast, M, Faggioli, G, Ferro, N, Li, Y, Fang, B, He, J, Yoshikawa, H, Akhondi, SA, Druckenbrodt, C, Thorne, C, Afzal, Z, Zhai, Z, Baldwin, T, Verspoor, K, Barron-Cedeno, A, DaSanMartino, G, Esposti, MD, Sebastiani, F, Macdonald, C, Pasi, G, Hanbury, A, Potthast, M, Faggioli, G, Ferro, N, Li, Y, Fang, B, He, J, Yoshikawa, H, Akhondi, SA, Druckenbrodt, C, Thorne, C, Afzal, Z, Zhai, Z, Baldwin, T, and Verspoor, K
- Published
- 2022
21. Cross-modal Clinical Graph Transformer for Ophthalmic Report Generation
- Author
-
Li, M, Cai, W, Verspoor, K, Pan, S, Liang, X, Chang, X, Li, M, Cai, W, Verspoor, K, Pan, S, Liang, X, and Chang, X
- Abstract
Automatic generation of ophthalmic reports using datadriven neural networks has great potential in clinical practice. When writing a report, ophthalmologists make inferences with prior clinical knowledge. This knowledge has been neglected in prior medical report generation methods. To endow models with the capability of incorporating expert knowledge, we propose a Cross-modal clinical Graph Transformer (CGT) for ophthalmic report generation (ORG), in which clinical relation triples are injected into the visual features as prior knowledge to drive the decoding procedure. However, two major common Knowledge Noise (KN) issues may affect models' effectiveness. 1) Existing general biomedical knowledge bases such as the UMLS may not align meaningfully to the specific context and language of the report, limiting their utility for knowledge injection. 2) Incorporating too much knowledge may divert the visual features from their correct meaning. To overcome these limitations, we design an automatic information extraction scheme based on natural language processing to obtain clinical entities and relations directly from in-domain training reports. Given a set of ophthalmic images, our CGT first restores a sub-graph from the clinical graph and injects the restored triples into visual features. Then visible matrix is employed during the encoding procedure to limit the impact of knowledge. Finally, reports are predicted by the encoded cross-modal features via a Transformer decoder. Extensive experiments on the large-scale FFA-IR benchmark demonstrate that the proposed CGT is able to outperform previous benchmark methods and achieve state-of-the-art performances.
- Published
- 2022
22. Overcoming challenges in extracting prescribing habits from veterinary clinics using big data and deep learning
- Author
-
Hur, B, primary, Hardefeldt, LY, additional, Verspoor, K, additional, Baldwin, T, additional, and Gilkerson, JR, additional
- Published
- 2022
- Full Text
- View/download PDF
23. Use and validation of text mining and cluster algorithms to derive insights from Corona Virus Disease-2019 (COVID-19) medical literature
- Author
-
Reddy, Sandeep, Bhaskar, R, Padmanabhan, S, Verspoor, K, Mamillapalli, C, Lahoti, R, Makinen, V-P, Pradhan, S, Kushwah, P, Sinha, S, Reddy, Sandeep, Bhaskar, R, Padmanabhan, S, Verspoor, K, Mamillapalli, C, Lahoti, R, Makinen, V-P, Pradhan, S, Kushwah, P, and Sinha, S
- Published
- 2021
24. ChemTables: a dataset for semantic classification on tables in chemical patents
- Author
-
Zhai, Z, Druckenbrodt, C, Thorne, C, Akhondi, SA, Dat, QN, Cohn, T, Verspoor, K, Zhai, Z, Druckenbrodt, C, Thorne, C, Akhondi, SA, Dat, QN, Cohn, T, and Verspoor, K
- Abstract
Chemical patents are a commonly used channel for disclosing novel compounds and reactions, and hence represent important resources for chemical and pharmaceutical research. Key chemical data in patents is often presented in tables. Both the number and the size of tables can be very large in patent documents. In addition, various types of information can be presented in tables in patents, including spectroscopic and physical data, or pharmacological use and effects of chemicals. Since images of Markush structures and merged cells are commonly used in these tables, their structure also shows substantial variation. This heterogeneity in content and structure of tables in chemical patents makes relevant information difficult to find. We therefore propose a new text mining task of automatically categorising tables in chemical patents based on their contents. Categorisation of tables based on the nature of their content can help to identify tables containing key information, improving the accessibility of information in patents that is highly relevant for new inventions. For developing and evaluating methods for the table classification task, we developed a new dataset, called CHEMTABLES, which consists of 788 chemical patent tables with labels of their content type. We introduce this data set in detail. We further establish strong baselines for the table classification task in chemical patents by applying state-of-the-art neural network models developed for natural language processing, including TabNet, ResNet and Table-BERT on CHEMTABLES. The best performing model, Table-BERT, achieves a performance of 88.66 micro-averaged [Formula: see text] score on the table classification task. The CHEMTABLES dataset is publicly available at https://doi.org/10.17632/g7tjh7tbrj.3 , subject to the CC BY NC 3.0 license. Code/models evaluated in this work are in a Github repository https://github.com/zenanz/ChemTables .
- Published
- 2021
25. Use and validation of text mining and cluster algorithms to derive insights from Corona Virus Disease-2019 (COVID-19) medical literature.
- Author
-
Reddy, S, Bhaskar, R, Padmanabhan, S, Verspoor, K, Mamillapalli, C, Lahoti, R, Makinen, V-P, Pradhan, S, Kushwah, P, Sinha, S, Reddy, S, Bhaskar, R, Padmanabhan, S, Verspoor, K, Mamillapalli, C, Lahoti, R, Makinen, V-P, Pradhan, S, Kushwah, P, and Sinha, S
- Abstract
The emergence of the severe acute respiratory syndrome coronavirus 2 (SARS-CoV-2) late last year has not only led to the world-wide coronavirus disease 2019 (COVID-19) pandemic but also a deluge of biomedical literature. Following the release of the COVID-19 open research dataset (CORD-19) comprising over 200,000 scholarly articles, we a multi-disciplinary team of data scientists, clinicians, medical researchers and software engineers developed an innovative natural language processing (NLP) platform that combines an advanced search engine with a biomedical named entity recognition extraction package. In particular, the platform was developed to extract information relating to clinical risk factors for COVID-19 by presenting the results in a cluster format to support knowledge discovery. Here we describe the principles behind the development, the model and the results we obtained.
- Published
- 2021
26. Machine learning with a reduced dimensionality representation of comprehensive Pentacam tomography parameters to identify subclinical keratoconus
- Author
-
Cao, K, Verspoor, K, Chan, E, Daniell, M, Sahebjada, S, Baird, PN, Cao, K, Verspoor, K, Chan, E, Daniell, M, Sahebjada, S, and Baird, PN
- Abstract
PURPOSE: To investigate the performance of a machine learning model based on a reduced dimensionality parameter space derived from complete Pentacam parameters to identify subclinical keratoconus (KC). METHODS: All 1692 available parameters were obtained from the Pentacam imaging machine on 145 subclinical KC and 122 control eyes. We applied a principal component analysis (PCA) to the complete Pentacam dataset to reduce its parameter dimensionality. Subsequently, we investigated machine learning performance of the random forest algorithm with increasing numbers of components to identify their optimal number for detecting subclinical KC from control eyes. RESULTS: The dimensionality of the complete set of 1692 Pentacam parameters was reduced to 267 principal components using PCA. Subsequent selection of 15 of these principal components explained over 85% of the variance of the original Pentacam-derived parameters and input to train a random forest machine learning model to achieve the best accuracy of 98% in detecting subclinical KC eyes. The model established also reached a high sensitivity of 97% in identification of subclinical KC and a specificity of 98% in recognizing control eyes. CONCLUSIONS: A random forest-based model trained using a modest number of components derived from a reduced dimensionality representation of complete Pentacam system parameters allowed for high accuracy of subclinical KC identification.
- Published
- 2021
27. ChEMU 2020: Natural Language Processing Methods Are Effective for Information Extraction From Chemical Patents.
- Author
-
He, J, Nguyen, DQ, Akhondi, SA, Druckenbrodt, C, Thorne, C, Hoessel, R, Afzal, Z, Zhai, Z, Fang, B, Yoshikawa, H, Albahem, A, Cavedon, L, Cohn, T, Baldwin, T, Verspoor, K, He, J, Nguyen, DQ, Akhondi, SA, Druckenbrodt, C, Thorne, C, Hoessel, R, Afzal, Z, Zhai, Z, Fang, B, Yoshikawa, H, Albahem, A, Cavedon, L, Cohn, T, Baldwin, T, and Verspoor, K
- Abstract
Chemical patents represent a valuable source of information about new chemical compounds, which is critical to the drug discovery process. Automated information extraction over chemical patents is, however, a challenging task due to the large volume of existing patents and the complex linguistic properties of chemical patents. The Cheminformatics Elsevier Melbourne University (ChEMU) evaluation lab 2020, part of the Conference and Labs of the Evaluation Forum 2020 (CLEF2020), was introduced to support the development of advanced text mining techniques for chemical patents. The ChEMU 2020 lab proposed two fundamental information extraction tasks focusing on chemical reaction processes described in chemical patents: (1) chemical named entity recognition, requiring identification of essential chemical entities and their roles in chemical reactions, as well as reaction conditions; and (2) event extraction, which aims at identification of event steps relating the entities involved in chemical reactions. The ChEMU 2020 lab received 37 team registrations and 46 runs. Overall, the performance of submissions for these tasks exceeded our expectations, with the top systems outperforming strong baselines. We further show the methods to be robust to variations in sampling of the test data. We provide a detailed overview of the ChEMU 2020 corpus and its annotation, showing that inter-annotator agreement is very strong. We also present the methods adopted by participants, provide a detailed analysis of their performance, and carefully consider the potential impact of data leakage on interpretation of the results. The ChEMU 2020 Lab has shown the viability of automated methods to support information extraction of key information in chemical patents.
- Published
- 2021
28. Memorization vs. Generalization: Quantifying data leakage in NLP performance evaluation
- Author
-
Elangovan, A, He, J, Verspoor, K, Elangovan, A, He, J, and Verspoor, K
- Published
- 2021
29. Automatic consistency assurance for literature-based gene ontology annotation
- Author
-
Chen, J, Geard, N, Zobel, J, Verspoor, K, Chen, J, Geard, N, Zobel, J, and Verspoor, K
- Abstract
BACKGROUND: Literature-based gene ontology (GO) annotation is a process where expert curators use uniform expressions to describe gene functions reported in research papers, creating computable representations of information about biological systems. Manual assurance of consistency between GO annotations and the associated evidence texts identified by expert curators is reliable but time-consuming, and is infeasible in the context of rapidly growing biological literature. A key challenge is maintaining consistency of existing GO annotations as new studies are published and the GO vocabulary is updated. RESULTS: In this work, we introduce a formalisation of biological database annotation inconsistencies, identifying four distinct types of inconsistency. We propose a novel and efficient method using state-of-the-art text mining models to automatically distinguish between consistent GO annotation and the different types of inconsistent GO annotation. We evaluate this method using a synthetic dataset generated by directed manipulation of instances in an existing corpus, BC4GO. We provide detailed error analysis for demonstrating that the method achieves high precision on more confident predictions. CONCLUSIONS: Two models built using our method for distinct annotation consistency identification tasks achieved high precision and were robust to updates in the GO vocabulary. Our approach demonstrates clear value for human-in-the-loop curation scenarios.
- Published
- 2021
30. Early prediction of diagnostic-related groups and estimation of hospital cost by processing clinical notes
- Author
-
Liu, J, Capurro, D, Nguyen, A, Verspoor, K, Liu, J, Capurro, D, Nguyen, A, and Verspoor, K
- Abstract
As healthcare providers receive fixed amounts of reimbursement for given services under DRG (Diagnosis-Related Groups) payment, DRG codes are valuable for cost monitoring and resource allocation. However, coding is typically performed retrospectively post-discharge. We seek to predict DRGs and DRG-based case mix index (CMI) at early inpatient admission using routine clinical text to estimate hospital cost in an acute setting. We examined a deep learning-based natural language processing (NLP) model to automatically predict per-episode DRGs and corresponding cost-reflecting weights on two cohorts (paid under Medicare Severity (MS) DRG or All Patient Refined (APR) DRG), without human coding efforts. It achieved macro-averaged area under the receiver operating characteristic curve (AUC) scores of 0·871 (SD 0·011) on MS-DRG and 0·884 (0·003) on APR-DRG in fivefold cross-validation experiments on the first day of ICU admission. When extended to simulated patient populations to estimate average cost-reflecting weights, the model increased its accuracy over time and obtained absolute CMI error of 2·40 (1·07%) and 12·79% (2·31%), respectively on the first day. As the model could adapt to variations in admission time, cohort size, and requires no extra manual coding efforts, it shows potential to help estimating costs for active patients to support better operational decision-making in hospitals.
- Published
- 2021
31. ChEMU-Ref: A corpus for modeling anaphora resolution in the chemical domain
- Author
-
Fang, B, Druckenbrodt, C, Akhondi, SA, He, J, Baldwin, T, Verspoor, K, Fang, B, Druckenbrodt, C, Akhondi, SA, He, J, Baldwin, T, and Verspoor, K
- Abstract
Chemical patents contain rich coreference and bridging links, which are the target of this research. Specially, we introduce a novel annotation scheme, based on which we create the ChEMU-Ref dataset from reaction description snippets in English-language chemical patents. We propose a neural approach to anaphora resolution, which we show to achieve strong results, especially when jointly trained over coreference and bridging links.
- Published
- 2021
32. Bow-tie architecture of gene regulatory networks in species of varying complexity
- Author
-
Roy, GG, He, S, Geard, N, Verspoor, K, Roy, GG, He, S, Geard, N, and Verspoor, K
- Abstract
The gene regulatory network (GRN) architecture plays a key role in explaining the biological differences between species. We aim to understand species differences in terms of some universally present dynamical properties of their gene regulatory systems. A network architectural feature associated with controlling system-level dynamical properties is the bow-tie, identified by a strongly connected subnetwork, the core layer, between two sets of nodes, the in and the out layers. Though a bow-tie architecture has been observed in many networks, its existence has not been extensively investigated in GRNs of species of widely varying biological complexity. We analyse publicly available GRNs of several well-studied species from prokaryotes to unicellular eukaryotes to multicellular organisms. In their GRNs, we find the existence of a bow-tie architecture with a distinct largest strongly connected core layer. We show that the bow-tie architecture is a characteristic feature of GRNs. We observe an increasing trend in the relative core size with species complexity. Using studied relationships of the core size with dynamical properties like robustness and fragility, flexibility, criticality, controllability and evolvability, we hypothesize how these regulatory system properties have emerged differently with biological complexity, based on the observed differences of the GRN bow-tie architectures.
- Published
- 2021
33. The Evolution of Clinical Knowledge During COVID-19: Towards a Global Learning Health System.
- Author
-
Verspoor, K and Verspoor, K
- Abstract
OBJECTIVES: We examine the knowledge ecosystem of COVID-19, focusing on clinical knowledge and the role of health informatics as enabling technology. We argue for commitment to the model of a global learning health system to facilitate rapid knowledge translation supporting health care decision making in the face of emerging diseases. METHODS AND RESULTS: We frame the evolution of knowledge in the COVID-19 crisis in terms of learning theory, and present a view of what has occurred during the pandemic to rapidly derive and share knowledge as an (underdeveloped) instance of a global learning health system. We identify the key role of information technologies for electronic data capture and data sharing, computational modelling, evidence synthesis, and knowledge dissemination. We further highlight gaps in the system and barriers to full realisation of an efficient and effective global learning health system. CONCLUSIONS: The need for a global knowledge ecosystem supporting rapid learning from clinical practice has become more apparent than ever during the COVID-19 pandemic. Continued effort to realise the vision of a global learning health system, including establishing effective approaches to data governance and ethics to support the system, is imperative to enable continuous improvement in our clinical care.
- Published
- 2021
34. Measuring emotions in the COVID-19 real world worry dataset
- Author
-
Kleinberg, Bennett, van der Vegt, Isabelle, Mozes, Maximilian, Verspoor, K., Bretonnel Cohen, K., Dredze, M., Ferrara, E., May, J., Munro, R., Paris, C., Wallace, B., and Department of Methodology and Statistics
- Abstract
The COVID-19 pandemic is having a dramatic impact on societies and economies around the world. With various measures of lockdowns and social distancing in place, it becomes important to understand emotional responses on a large scale. In this paper, we present the first ground truth dataset of emotional responses to COVID-19. We asked participants to indicate their emotions and express these in text. This resulted in the Real World Worry Dataset of 5,000 texts (2,500 short + 2,500 long texts). Our analyses suggest that emotional responses correlated with linguistic measures. Topic modeling further revealed that people in the UK worry about their family and the economic situation. Tweet-sized texts functioned as a call for solidarity, while longer texts shed light on worries and concerns. Using predictive modeling approaches, we were able to approximate the emotional responses of participants from text within 14% of their actual value. We encourage others to use the dataset and improve how we can use automated methods to learn about emotional responses and worries about an urgent problem.
- Published
- 2020
35. PoLoBag: Polynomial Lasso Bagging for signed gene regulatory network inference from expression data.
- Author
-
Luigi Martelli, P, Roy, GG, Geard, N, Verspoor, K, He, S, Luigi Martelli, P, Roy, GG, Geard, N, Verspoor, K, and He, S
- Abstract
MOTIVATION: Inferring gene regulatory networks (GRNs) from expression data is a significant systems biology problem. A useful inference algorithm should not only unveil the global structure of the regulatory mechanisms but also the details of regulatory interactions such as edge direction (from regulator to target) and sign (activation/inhibition). Many popular GRN inference algorithms cannot infer edge signs, and those that can infer signed GRNs cannot simultaneously infer edge directions or network cycles. RESULTS: To address these limitations of existing algorithms we propose Polynomial Lasso Bagging (PoLoBag) for signed GRN inference with both edge directions and network cycles. PoLoBag is an ensemble regression algorithm in a bagging framework where Lasso weights estimated on bootstrap samples are averaged. These bootstrap samples incorporate polynomial features to capture higher order interactions. Results demonstrate that PoLoBag is consistently more accurate for signed inference than state-of-the-art algorithms on simulated and real-world expression datasets. AVAILABILITY: Algorithm and data are freely available at https://github.com/gourabghoshroy/PoLoBag. SUPPLEMENTARY INFORMATION: Supplementary data are available at Bioinformatics online.
- Published
- 2020
36. Domain Adaptation and Instance Selection for Disease Syndrome Classification over Veterinary Clinical Notes
- Author
-
Hur, B, Baldwin, T, Verspoor, K, Hardefeldt, L, Gilkerson, J, Hur, B, Baldwin, T, Verspoor, K, Hardefeldt, L, and Gilkerson, J
- Published
- 2020
37. Development of a Self-Harm Monitoring System for Victoria
- Author
-
Robinson, J, Witt, K, Lamblin, M, Spittal, MJ, Carter, G, Verspoor, K, Page, A, Rajaram, G, Rozova, V, Hill, NTM, Pirkis, J, Bleeker, C, Pleban, A, Knott, JC, Robinson, J, Witt, K, Lamblin, M, Spittal, MJ, Carter, G, Verspoor, K, Page, A, Rajaram, G, Rozova, V, Hill, NTM, Pirkis, J, Bleeker, C, Pleban, A, and Knott, JC
- Abstract
The prevention of suicide and suicide-related behaviour are key policy priorities in Australia and internationally. The World Health Organization has recommended that member states develop self-harm surveillance systems as part of their suicide prevention efforts. This is also a priority under Australia's Fifth National Mental Health and Suicide Prevention Plan. The aim of this paper is to describe the development of a state-based self-harm monitoring system in Victoria, Australia. In this system, data on all self-harm presentations are collected from eight hospital emergency departments in Victoria. A natural language processing classifier that uses machine learning to identify episodes of self-harm is currently being developed. This uses the free-text triage case notes, together with certain structured data fields, contained within the metadata of the incoming records. Post-processing is undertaken to identify primary mechanism of injury, substances consumed (including alcohol, illicit drugs and pharmaceutical preparations) and presence of psychiatric disorders. This system will ultimately leverage routinely collected data in combination with advanced artificial intelligence methods to support robust community-wide monitoring of self-harm. Once fully operational, this system will provide accurate and timely information on all presentations to participating emergency departments for self-harm, thereby providing a useful indicator for Australia's suicide prevention efforts.
- Published
- 2020
38. Artificial intelligence for clinical decision support in neurology
- Author
-
Pedersen, M, Verspoor, K, Jenkinson, M, Law, M, Abbott, DF, Jackson, GD, Pedersen, M, Verspoor, K, Jenkinson, M, Law, M, Abbott, DF, and Jackson, GD
- Abstract
Artificial intelligence is one of the most exciting methodological shifts in our era. It holds the potential to transform healthcare as we know it, to a system where humans and machines work together to provide better treatment for our patients. It is now clear that cutting edge artificial intelligence models in conjunction with high-quality clinical data will lead to improved prognostic and diagnostic models in neurological disease, facilitating expert-level clinical decision tools across healthcare settings. Despite the clinical promise of artificial intelligence, machine and deep-learning algorithms are not a one-size-fits-all solution for all types of clinical data and questions. In this article, we provide an overview of the core concepts of artificial intelligence, particularly contemporary deep-learning methods, to give clinician and neuroscience researchers an appreciation of how artificial intelligence can be harnessed to support clinical decisions. We clarify and emphasize the data quality and the human expertise needed to build robust clinical artificial intelligence models in neurology. As artificial intelligence is a rapidly evolving field, we take the opportunity to iterate important ethical principles to guide the field of medicine is it moves into an artificial intelligence enhanced future.
- Published
- 2020
39. Evaluating the Performance of Various Machine Learning Algorithms to Detect Subclinical Keratoconus
- Author
-
Cao, K, Verspoor, K, Sahebjada, S, Baird, PN, Cao, K, Verspoor, K, Sahebjada, S, and Baird, PN
- Abstract
Purpose: Keratoconus (KC) represents one of the leading causes of corneal transplantation worldwide. Detecting subclinical KC would lead to better management to avoid the need for corneal grafts, but the condition is clinically challenging to diagnose. We wished to compare eight commonly used machine learning algorithms using a range of parameter combinations by applying them to our KC dataset and build models to better differentiate subclinical KC from non-KC eyes. Methods: Oculus Pentacam was used to obtain corneal parameters on 49 subclinical KC and 39 control eyes, along with clinical and demographic parameters. Eight machine learning methods were applied to build models to differentiate subclinical KC from control eyes. Dominant algorithms were trained with all combinations of the considered parameters to select important parameter combinations. The performance of each model was evaluated and compared. Results: Using a total of eleven parameters, random forest, support vector machine and k-nearest neighbors had better performance in detecting subclinical KC. The highest area under the curve of 0.97 for detecting subclinical KC was achieved using five parameters by the random forest method. The highest sensitivity (0.94) and specificity (0.90) were obtained by the support vector machine and the k-nearest neighbor model, respectively. Conclusions: This study showed machine learning algorithms can be applied to identify subclinical KC using a minimal parameter set that are routinely collected during clinical eye examination. Translational Relevance: Machine learning algorithms can be built using routinely collected clinical parameters that will assist in the objective detection of subclinical KC.
- Published
- 2020
40. Quality Matters: Biocuration Experts on the Impact of Duplication and Other Data Quality Issues in Biological Databases.
- Author
-
Chen, Q, Britto, R, Erill, I, Jeffery, CJ, Liberzon, A, Magrane, M, Onami, J-I, Robinson-Rechavi, M, Sponarova, J, Zobel, J, Verspoor, K, Chen, Q, Britto, R, Erill, I, Jeffery, CJ, Liberzon, A, Magrane, M, Onami, J-I, Robinson-Rechavi, M, Sponarova, J, Zobel, J, and Verspoor, K
- Abstract
Biological databases represent an extraordinary collective volume of work. Diligently built up over decades and comprising many millions of contributions from the biomedical research community, biological databases provide worldwide access to a massive number of records (also known as entries) [1]. Starting from individual laboratories, genomes are sequenced, assembled, annotated, and ultimately submitted to primary nucleotide databases such as GenBank [2], European Nucleotide Archive (ENA) [3], and DNA Data Bank of Japan (DDBJ) [4] (collectively known as the International Nucleotide Sequence Database Collaboration, INSDC). Protein records, which are the translations of these nucleotide records, are deposited into central protein databases such as the UniProt KnowledgeBase (UniProtKB) [5] and the Protein Data Bank (PDB) [6]. Sequence records are further accumulated into different databases for more specialized purposes: RFam [7] and PFam [8] for RNA and protein families, respectively; DictyBase [9] and PomBase [10] for model organisms; as well as ArrayExpress [11] and Gene Expression Omnibus (GEO) [12] for gene expression profiles. These databases are selected as examples; the list is not intended to be exhaustive. However, they are representative of biological databases that have been named in the “golden set” of the 24th Nucleic Acids Research database issue (in 2016). The introduction of that issue highlights the databases that “consistently served as authoritative, comprehensive, and convenient data resources widely used by the entire community and offer some lessons on what makes a successful database” [13]. In addition, the associated information about sequences is also propagated into non-sequence databases, such as PubMed (https://www.ncbi.nlm.nih.gov/pubmed/) for scientific literature or Gene Ontology (GO) [14] for function annotations. These databases in turn benefit individual studies, many of which use these publicly available records as the basis for th
- Published
- 2020
41. Improved Topic Representations of Medical Documents to Assist COVID-19 Literature Exploration
- Author
-
Otmakhova, Y, Verspoor, K, Baldwin, T, Šuster, S, Otmakhova, Y, Verspoor, K, Baldwin, T, and Šuster, S
- Published
- 2020
42. Testing Contextualized Word Embeddings to Improve NER in Spanish Clinical Case Narratives
- Author
-
Akhtyamova, L, Martinez, P, Verspoor, K, Cardiff, J, Akhtyamova, L, Martinez, P, Verspoor, K, and Cardiff, J
- Abstract
In the Big Data era, there is an increasing need to fully exploit and analyze the huge quantity of information available about health. Natural Language Processing (NLP) technologies can contribute by extracting relevant information from unstructured data contained in Electronic Health Records (EHR) such as clinical notes, patients’ discharge summaries and radiology reports. The extracted information can help in health-related decision making processes. The Named Entity Recognition (NER) task, which detects important concepts in texts (e.g., diseases, symptoms, drugs, etc.), is crucial in the information extraction process yet has received little attention in languages other than English. In this work, we develop a deep learning-based NLP pipeline for biomedical entity extraction in Spanish clinical narratives. We explore the use of contextualized word embeddings, which incorporate context variation into word representations, to enhance named entity recognition in Spanish language clinical text, particularly of pharmacological substances, compounds, and proteins. Various combinations of word and sense embeddings were tested on the evaluation corpus of the PharmacoNER 2019 task, the Spanish Clinical Case Corpus (SPACCC). This data set consists of clinical case sections extracted from open access Spanish-language medical publications. Our study shows that our deep-learning-based system with domain-specific contextualized embeddings coupled with stacking of complementary embeddings yields superior performance over a system with integrated standard and general-domain word embeddings. With this system, we achieve performance competitive with the state-of-the-art.
- Published
- 2020
43. From Evaluating to Forecasting Performance: How to Turn Information Retrieval, Natural Language Processing and Recommender Systems into Predictive Sciences (Dagstuhl Perspectives Workshop 17442)
- Author
-
Ferro N., Fuhr N., Grefenstette G., Konstan J. A., Castells P., Daly E. M., Declerck T., Ekstrand M. D., Geyer W., Gonzalo J., Kuflik T., Lind'En K., Magnini B., Nie J. Y., Perego R., Shapira B., Soboroff I., Tintarev N., Verspoor K., Willemsen M. C., Zobel J., and Human Technology Interaction
- Subjects
000 Computer science, knowledge, general works ,05 social sciences ,Formal models ,User interaction ,02 engineering and technology ,020204 information systems ,Computer Science ,0202 electrical engineering, electronic engineering, information engineering ,0509 other social sciences ,Evaluation ,050904 information & library sciences ,Simulation ,Information Systems - Abstract
We describe the state-of-the-art in performance modeling and prediction for Information Retrieval (IR), Natural Language Processing (NLP) and Recommender Systems (RecSys) along with its shortcomings and strengths. We present a framework for further research, identifying five major problem areas: understanding measures, performance analysis, making underlying assumptions explicit, identifying application features determining performance, and the development of predic- tion models describing the relationship between assumptions, features and resulting performance
- Published
- 2019
44. Automated assessment of biological database assertions using the scientific literature
- Author
-
Bouadjenek, MR, Zobel, J, Verspoor, K, Bouadjenek, MR, Zobel, J, and Verspoor, K
- Abstract
Background The large biological databases such as GenBank contain vast numbers of records, the content of which is substantively based on external resources, including published literature. Manual curation is used to establish whether the literature and the records are indeed consistent. We explore in this paper an automated method for assessing the consistency of biological assertions, to assist biocurators, which we call BARC, Biocuration tool for Assessment of Relation Consistency. In this method a biological assertion is represented as a relation between two objects (for example, a gene and a disease); we then use our novel set-based relevance algorithm SaBRA to retrieve pertinent literature, and apply a classifier to estimate the likelihood that this relation (assertion) is correct. Results Our experiments on assessing gene–disease relations and protein–protein interactions using the PubMed Central collection show that BARC can be effective at assisting curators to perform data cleansing. Specifically, the results obtained showed that BARC substantially outperforms the best baselines, with an improvement of F-measure of 3.5% and 13%, respectively, on gene-disease relations and protein-protein interactions. We have additionally carried out a feature analysis that showed that all feature types are informative, as are all fields of the documents. Conclusions BARC provides a clear benefit for the biocuration community, as there are no prior automated tools for identifying inconsistent assertions in large-scale biological databases.
- Published
- 2019
45. Detecting Chemical Reactions in Patents
- Author
-
Yoshikawa, H, Nguyen, DQ, Zhai, Z, Druckenbrodt, C, Thorne, C, Akhondi, SA, Baldwin, T, Verspoor, K, Yoshikawa, H, Nguyen, DQ, Zhai, Z, Druckenbrodt, C, Thorne, C, Akhondi, SA, Baldwin, T, and Verspoor, K
- Abstract
Extracting chemical reactions from patents is a crucial task for chemists working on chemical exploration. In this paper we introduce the novel task of detecting the textual spans that describe or refer to chemical reactions within patents. We formulate this task as a paragraph-level sequence tagging problem, where the system is required to return a sequence of paragraphs that contain a description of a reaction. To address this new task, we construct an annotated dataset from an existing proprietary database of chemical reactions manually extracted from patents. We introduce several baseline methods for the task and evaluate them over our dataset. Through error analysis, we discuss what makes the task complex and challenging, and suggest possible directions for future research.
- Published
- 2019
46. From POS tagging to dependency parsing for biomedical event extraction
- Author
-
Dat, QN, Verspoor, K, Dat, QN, and Verspoor, K
- Abstract
BACKGROUND: Given the importance of relation or event extraction from biomedical research publications to support knowledge capture and synthesis, and the strong dependency of approaches to this information extraction task on syntactic information, it is valuable to understand which approaches to syntactic processing of biomedical text have the highest performance. RESULTS: We perform an empirical study comparing state-of-the-art traditional feature-based and neural network-based models for two core natural language processing tasks of part-of-speech (POS) tagging and dependency parsing on two benchmark biomedical corpora, GENIA and CRAFT. To the best of our knowledge, there is no recent work making such comparisons in the biomedical context; specifically no detailed analysis of neural models on this data is available. Experimental results show that in general, the neural models outperform the feature-based models on two benchmark biomedical corpora GENIA and CRAFT. We also perform a task-oriented evaluation to investigate the influences of these models in a downstream application on biomedical event extraction, and show that better intrinsic parsing performance does not always imply better extrinsic event extraction performance. CONCLUSION: We have presented a detailed empirical study comparing traditional feature-based and neural network-based models for POS tagging and dependency parsing in the biomedical context, and also investigated the influence of parser selection for a biomedical event extraction downstream task. AVAILABILITY OF DATA AND MATERIALS: We make the retrained models available at https://github.com/datquocnguyen/BioPosDep .
- Published
- 2019
47. Overview of the BioCreative VI Precision Medicine Track: mining protein interactions and mutations for precision medicine
- Author
-
Dogan, RI, Kim, S, Chatr-aryamontri, A, Wei, C-H, Comeau, DC, Antunes, R, Matos, S, Chen, Q, Elangovan, A, Panyam, NC, Verspoor, K, Liu, H, Wang, Y, Liu, Z, Altinel, B, Husunbeyi, ZM, Ozgur, A, Fergadis, A, Wang, C-K, Dai, H-J, Tran, T, Kavuluru, R, Luo, L, Steppi, A, Zhang, J, Qu, J, Lu, Z, Dogan, RI, Kim, S, Chatr-aryamontri, A, Wei, C-H, Comeau, DC, Antunes, R, Matos, S, Chen, Q, Elangovan, A, Panyam, NC, Verspoor, K, Liu, H, Wang, Y, Liu, Z, Altinel, B, Husunbeyi, ZM, Ozgur, A, Fergadis, A, Wang, C-K, Dai, H-J, Tran, T, Kavuluru, R, Luo, L, Steppi, A, Zhang, J, Qu, J, and Lu, Z
- Abstract
The Precision Medicine Initiative is a multicenter effort aiming at formulating personalized treatments leveraging on individual patient data (clinical, genome sequence and functional genomic data) together with the information in large knowledge bases (KBs) that integrate genome annotation, disease association studies, electronic health records and other data types. The biomedical literature provides a rich foundation for populating these KBs, reporting genetic and molecular interactions that provide the scaffold for the cellular regulatory systems and detailing the influence of genetic variants in these interactions. The goal of BioCreative VI Precision Medicine Track was to extract this particular type of information and was organized in two tasks: (i) document triage task, focused on identifying scientific literature containing experimentally verified protein–protein interactions (PPIs) affected by genetic mutations and (ii) relation extraction task, focused on extracting the affected interactions (protein pairs). To assist system developers and task participants, a large-scale corpus of PubMed documents was manually annotated for this task. Ten teams worldwide contributed 22 distinct text-mining models for the document triage task, and six teams worldwide contributed 14 different text-mining systems for the relation extraction task. When comparing the text-mining system predictions with human annotations, for the triage task, the best F-score was 69.06%, the best precision was 62.89%, the best recall was 98.0% and the best average precision was 72.5%. For the relation extraction task, when taking homologous genes into account, the best F-score was 37.73%, the best precision was 46.5% and the best recall was 54.1%. Submitted systems explored a wide range of methods, from traditional rule-based, statistical and machine learning systems to state-of-the-art deep learning methods. Given the level of participation and the individual team results we find the precision
- Published
- 2019
48. Exploring effective approaches for haplotype block phasing
- Author
-
Al Bkhetan, Z, Zobel, J, Kowalczyk, A, Verspoor, K, Goudey, B, Al Bkhetan, Z, Zobel, J, Kowalczyk, A, Verspoor, K, and Goudey, B
- Abstract
BACKGROUND: Knowledge of phase, the specific allele sequence on each copy of homologous chromosomes, is increasingly recognized as critical for detecting certain classes of disease-associated mutations. One approach for detecting such mutations is through phased haplotype association analysis. While the accuracy of methods for phasing genotype data has been widely explored, there has been little attention given to phasing accuracy at haplotype block scale. Understanding the combined impact of the accuracy of phasing tool and the method used to determine haplotype blocks on the error rate within the determined blocks is essential to conduct accurate haplotype analyses. RESULTS: We present a systematic study exploring the relationship between seven widely used phasing methods and two common methods for determining haplotype blocks. The evaluation focuses on the number of haplotype blocks that are incorrectly phased. Insights from these results are used to develop a haplotype estimator based on a consensus of three tools. The consensus estimator achieved the most accurate phasing in all applied tests. Individually, EAGLE2, BEAGLE and SHAPEIT2 alternate in being the best performing tool in different scenarios. Determining haplotype blocks based on linkage disequilibrium leads to more correctly phased blocks compared to a sliding window approach. We find that there is little difference between phasing sections of a genome (e.g. a gene) compared to phasing entire chromosomes. Finally, we show that the location of phasing error varies when the tools are applied to the same data several times, a finding that could be important for downstream analyses. CONCLUSIONS: The choice of phasing and block determination algorithms and their interaction impacts the accuracy of phased haplotype blocks. This work provides guidance and evidence for the different design choices needed for analyses using haplotype blocks. The study highlights a number of issues that may have limited the rep
- Published
- 2019
49. Using natural language processing and VetCompass to understand antimicrobial usage patterns in Australia
- Author
-
Hur, B, Hardefeldt, LY, Verspoor, K, Baldwin, T, Gilkerson, JR, Hur, B, Hardefeldt, LY, Verspoor, K, Baldwin, T, and Gilkerson, JR
- Abstract
Background Currently there is an incomplete understanding of antimicrobial usage patterns in veterinary clinics in Australia, but such knowledge is critical for the successful implementation and monitoring of antimicrobial stewardship programs. Methods VetCompass Australia collects medical records from 181 clinics in Australia (as of May 2018). These records contain detailed information from individual consultations regarding the medications dispensed. One unique aspect of VetCompass Australia is its focus on applying natural language processing (NLP) and machine learning techniques to analyse the records, similar to efforts conducted in other medical studies. Results The free text fields of 4,394,493 veterinary consultation records of dogs and cats between 2013 and 2018 were collated by VetCompass Australia and NLP techniques applied to enable the querying of the antimicrobial usage within these consultations. Conclusion The NLP algorithms developed matched antimicrobial in clinical records with 96.7% accuracy and an F1 Score of 0.85, as evaluated relative to expert annotations. This dataset can be readily queried to demonstrate the antimicrobial usage patterns of companion animal practices throughout Australia.
- Published
- 2019
50. Using natural language processing and VetCompass to understand antimicrobial usage patterns in Australia
- Author
-
Hur, B, primary, Hardefeldt, LY, additional, Verspoor, K, additional, Baldwin, T, additional, and Gilkerson, JR, additional
- Published
- 2019
- Full Text
- View/download PDF
Catalog
Discovery Service for Jio Institute Digital Library
For full access to our library's resources, please sign in.