270 results on '"Verspoor K"'
Search Results
2. Generalizability of a Musculoskeletal Therapist Electronic Health Record for Modelling Outcomes to Work-Related Musculoskeletal Disorders
- Author
-
Wassell, M., Vitiello, A., Butler-Henderson, K., Verspoor, K., and Pollard, H.
- Published
- 2024
- Full Text
- View/download PDF
3. Integration of background knowledge for automatic detection of inconsistencies in gene ontology annotation
- Author
-
Chen, J, Goudey, B, Geard, N, Verspoor, K, Chen, J, Goudey, B, Geard, N, and Verspoor, K
- Abstract
MOTIVATION: Biological background knowledge plays an important role in the manual quality assurance (QA) of biological database records. One such QA task is the detection of inconsistencies in literature-based Gene Ontology Annotation (GOA). This manual verification ensures the accuracy of the GO annotations based on a comprehensive review of the literature used as evidence, Gene Ontology (GO) terms, and annotated genes in GOA records. While automatic approaches for the detection of semantic inconsistencies in GOA have been developed, they operate within predetermined contexts, lacking the ability to leverage broader evidence, especially relevant domain-specific background knowledge. This paper investigates various types of background knowledge that could improve the detection of prevalent inconsistencies in GOA. In addition, the paper proposes several approaches to integrate background knowledge into the automatic GOA inconsistency detection process. RESULTS: We have extended a previously developed GOA inconsistency dataset with several kinds of GOA-related background knowledge, including GeneRIF statements, biological concepts mentioned within evidence texts, GO hierarchy and existing GO annotations of the specific gene. We have proposed several effective approaches to integrate background knowledge as part of the automatic GOA inconsistency detection process. The proposed approaches can improve automatic detection of self-consistency and several of the most prevalent types of inconsistencies. UNLABELLED: This is the first study to explore the advantages of utilizing background knowledge and to propose a practical approach to incorporate knowledge in automatic GOA inconsistency detection. We establish a new benchmark for performance on this task. Our methods may be applicable to various tasks that involve incorporating biological background knowledge. AVAILABILITY AND IMPLEMENTATION: https://github.com/jiyuc/de-inconsistency.
- Published
- 2024
4. Focused Contrastive Loss for Classification With Pre-Trained Language Models
- Author
-
He, J, Li, Y, Zhai, Z, Fang, B, Thorne, C, Druckenbrodt, C, Akhondi, S, Verspoor, K, He, J, Li, Y, Zhai, Z, Fang, B, Thorne, C, Druckenbrodt, C, Akhondi, S, and Verspoor, K
- Published
- 2024
5. Advancing delirium classification: A clinical notes-based natural language processing-supported machine learning model
- Author
-
Amjad, S, Holmes, NE, Kishore, K, Young, M, Bailey, J, Bellomo, R, Verspoor, K, Amjad, S, Holmes, NE, Kishore, K, Young, M, Bailey, J, Bellomo, R, and Verspoor, K
- Published
- 2024
6. Electrocardiogram arrhythmia detection with novel signal processing and persistent homology-derived predictors
- Author
-
Verspoor, K, Dlugas, H, Verspoor, K, and Dlugas, H
- Abstract
Many approaches to computer-aided electrocardiogram (ECG) arrhythmia detection have been performed, several of which combine persistent homology and machine learning. We present a novel ECG signal processing pipeline and method of constructing predictor variables for use in statistical models. Specifically, we introduce an isoelectric baseline to yield non-trivial topological features corresponding to the P, Q, S, and T-waves (if they exist) and utilize the N-most persistent 1-dimensional homological features and their corresponding area-minimal cycle representatives to construct predictor variables derived from the persistent homology of the ECG signal for some choice of N. The binary classification of (1) Atrial Fibrillation vs. Non-Atrial Fibrillation, (2) Arrhythmia vs. Normal Sinus Rhythm, and (3) Arrhythmias with Morphological Changes vs. Sinus Rhythm with Bradycardia and Tachycardia Treated as Non-Arrhythmia was performed using Logistic Regression, Linear Discriminant Analysis, Quadratic Discriminant Analysis, Naive Bayes, Random Forest, Gradient Boosted Decision Tree, K-Nearest Neighbors, and Support Vector Machine with a linear, radial, and polynomial kernel Models with stratified 5-fold cross validation. The Gradient Boosted Decision Tree Model attained the best results with a mean F1-score and mean Accuracy of ( 0.967 , 0.946 ), ( 0.839 , 0.946 ), and ( 0.943 , 0.921 ) across the five folds for binary classifications of (1), (2), and (3), respectively.
- Published
- 2024
7. Graph embedding-based link prediction for literature-based discovery in Alzheimer's Disease
- Author
-
Pu, Y, Beck, D, Verspoor, K, Pu, Y, Beck, D, and Verspoor, K
- Abstract
OBJECTIVE: We explore the framing of literature-based discovery (LBD) as link prediction and graph embedding learning, with Alzheimer's Disease (AD) as our focus disease context. The key link prediction setting of prediction window length is specifically examined in the context of a time-sliced evaluation methodology. METHODS: We propose a four-stage approach to explore literature-based discovery for Alzheimer's Disease, creating and analyzing a knowledge graph tailored to the AD context, and predicting and evaluating new knowledge based on time-sliced link prediction. The first stage is to collect an AD-specific corpus. The second stage involves constructing an AD knowledge graph with identified AD-specific concepts and relations from the corpus. In the third stage, 20 pairs of training and testing datasets are constructed with the time-slicing methodology. Finally, we infer new knowledge with graph embedding-based link prediction methods. We compare different link prediction methods in this context. The impact of limiting prediction evaluation of LBD models in the context of short-term and longer-term knowledge evolution for Alzheimer's Disease is assessed. RESULTS: We constructed an AD corpus of over 16 k papers published in 1977-2021, and automatically annotated it with concepts and relations covering 11 AD-specific semantic entity types. The knowledge graph of Alzheimer's Disease derived from this resource consisted of ∼11 k nodes and ∼394 k edges, among which 34% were genotype-phenotype relationships, 57% were genotype-genotype relationships, and 9% were phenotype-phenotype relationships. A Structural Deep Network Embedding (SDNE) model consistently showed the best performance in terms of returning the most confident set of link predictions as time progresses over 20 years. A huge improvement in model performance was observed when changing the link prediction evaluation setting to consider a more distant future, reflecting the time required for knowledge accum
- Published
- 2023
8. Attention-based multimodal fusion with contrast for robust clinical prediction in the face of missing modalities
- Author
-
Liu, J, Capurro, D, Nguyen, A, Verspoor, K, Liu, J, Capurro, D, Nguyen, A, and Verspoor, K
- Abstract
OBJECTIVE: With the increasing amount and growing variety of healthcare data, multimodal machine learning supporting integrated modeling of structured and unstructured data is an increasingly important tool for clinical machine learning tasks. However, it is non-trivial to manage the differences in dimensionality, volume, and temporal characteristics of data modalities in the context of a shared target task. Furthermore, patients can have substantial variations in the availability of data, while existing multimodal modeling methods typically assume data completeness and lack a mechanism to handle missing modalities. METHODS: We propose a Transformer-based fusion model with modality-specific tokens that summarize the corresponding modalities to achieve effective cross-modal interaction accommodating missing modalities in the clinical context. The model is further refined by inter-modal, inter-sample contrastive learning to improve the representations for better predictive performance. We denote the model as Attention-based cRoss-MOdal fUsion with contRast (ARMOUR). We evaluate ARMOUR using two input modalities (structured measurements and unstructured text), six clinical prediction tasks, and two evaluation regimes, either including or excluding samples with missing modalities. RESULTS: Our model shows improved performances over unimodal or multimodal baselines in both evaluation regimes, including or excluding patients with missing modalities in the input. The contrastive learning improves the representation power and is shown to be essential for better results. The simple setup of modality-specific tokens enables ARMOUR to handle patients with missing modalities and allows comparison with existing unimodal benchmark results. CONCLUSION: We propose a multimodal model for robust clinical prediction to achieve improved performance while accommodating patients with missing modalities. This work could inspire future research to study the effective incorporation of multi
- Published
- 2023
9. Benchmarks for measurement of duplicate detection methods in nucleotide databases
- Author
-
Chen, Q, Zobel, J, Verspoor, K, Chen, Q, Zobel, J, and Verspoor, K
- Abstract
UNLABELLED: Duplication of information in databases is a major data quality challenge. The presence of duplicates, implying either redundancy or inconsistency, can have a range of impacts on the quality of analyses that use the data. To provide a sound basis for research on this issue in databases of nucleotide sequences, we have developed new, large-scale validated collections of duplicates, which can be used to test the effectiveness of duplicate detection methods. Previous collections were either designed primarily to test efficiency, or contained only a limited number of duplicates of limited kinds. To date, duplicate detection methods have been evaluated on separate, inconsistent benchmarks, leading to results that cannot be compared and, due to limitations of the benchmarks, of questionable generality. In this study, we present three nucleotide sequence database benchmarks, based on information drawn from a range of resources, including information derived from mapping to two data sections within the UniProt Knowledgebase (UniProtKB), UniProtKB/Swiss-Prot and UniProtKB/TrEMBL. Each benchmark has distinct characteristics. We quantify these characteristics and argue for their complementary value in evaluation. The benchmarks collectively contain a vast number of validated biological duplicates; the largest has nearly half a billion duplicate pairs (although this is probably only a tiny fraction of the total that is present). They are also the first benchmarks targeting the primary nucleotide databases. The records include the 21 most heavily studied organisms in molecular biology research. Our quantitative analysis shows that duplicates in the different benchmarks, and in different organisms, have different characteristics. It is thus unreliable to evaluate duplicate detection methods against any single benchmark. For example, the benchmark derived from UniProtKB/Swiss-Prot mappings identifies more diverse types of duplicates, showing the importance of expert cu
- Published
- 2023
10. Automating Quality Assessment of Medical Evidence in Systematic Reviews: Model Development and Validation Study
- Author
-
Suster, S, Baldwin, T, Lau, JH, Yepes, AJ, Iraola, DM, Otmakhova, Y, Verspoor, K, Suster, S, Baldwin, T, Lau, JH, Yepes, AJ, Iraola, DM, Otmakhova, Y, and Verspoor, K
- Abstract
BACKGROUND: Assessment of the quality of medical evidence available on the web is a critical step in the preparation of systematic reviews. Existing tools that automate parts of this task validate the quality of individual studies but not of entire bodies of evidence and focus on a restricted set of quality criteria. OBJECTIVE: We proposed a quality assessment task that provides an overall quality rating for each body of evidence (BoE), as well as finer-grained justification for different quality criteria according to the Grading of Recommendation, Assessment, Development, and Evaluation formalization framework. For this purpose, we constructed a new data set and developed a machine learning baseline system (EvidenceGRADEr). METHODS: We algorithmically extracted quality-related data from all summaries of findings found in the Cochrane Database of Systematic Reviews. Each BoE was defined by a set of population, intervention, comparison, and outcome criteria and assigned a quality grade (high, moderate, low, or very low) together with quality criteria (justification) that influenced that decision. Different statistical data, metadata about the review, and parts of the review text were extracted as support for grading each BoE. After pruning the resulting data set with various quality checks, we used it to train several neural-model variants. The predictions were compared against the labels originally assigned by the authors of the systematic reviews. RESULTS: Our quality assessment data set, Cochrane Database of Systematic Reviews Quality of Evidence, contains 13,440 instances, or BoEs labeled for quality, originating from 2252 systematic reviews published on the internet from 2002 to 2020. On the basis of a 10-fold cross-validation, the best neural binary classifiers for quality criteria detected risk of bias at 0.78 F1 (P=.68; R=0.92) and imprecision at 0.75 F1 (P=.66; R=0.86), while the performance on inconsistency, indirectness, and publication bias criteria was l
- Published
- 2023
11. Detecting evidence of invasive fungal infections in cytology and histopathology reports enriched with concept-level annotations
- Author
-
Rozova, V, Khanina, A, Teng, JC, Teh, JSK, Worth, LJ, Slavin, MA, Thursky, KA, Verspoor, K, Rozova, V, Khanina, A, Teng, JC, Teh, JSK, Worth, LJ, Slavin, MA, Thursky, KA, and Verspoor, K
- Abstract
Invasive fungal infections (IFIs) are particularly dangerous to high-risk patients with haematological malignancies and are responsible for excessive mortality and delays in cancer therapy. Surveillance of IFI in clinical settings offers an opportunity to identify potential risk factors and evaluate new therapeutic strategies. However, manual surveillance is both time- and resource-intensive. As part of a broader project aimed to develop a system for automated IFI surveillance by leveraging electronic medical records, we present our approach to detecting evidence of IFI in the key diagnostic domain of histopathology. Using natural language processing (NLP), we analysed cytology and histopathology reports to identify IFI-positive reports. We compared a conventional bag-of-words classification model to a method that relies on concept-level annotations. Although the investment to prepare data supporting concept annotations is substantial, extracting targeted information specific to IFI as a pre-processing step increased the performance of the classifier from the PR AUC of 0.84 to 0.92 and enabled model interpretability. We have made publicly available the annotated dataset of 283 reports, the Cytology and Histopathology IFI Reports corpus (CHIFIR), to allow the clinical NLP research community to further build on our results.
- Published
- 2023
12. Analysis of predictive performance and reliability of classifiers for quality assessment of medical evidence revealed important variation by medical area
- Author
-
Suster, S, Baldwin, T, Verspoor, K, Suster, S, Baldwin, T, and Verspoor, K
- Abstract
OBJECTIVES: A major obstacle in deployment of models for automated quality assessment is their reliability. To analyze their calibration and selective classification performance. STUDY DESIGN AND SETTING: We examine two systems for assessing the quality of medical evidence, EvidenceGRADEr and RobotReviewer, both developed from Cochrane Database of Systematic Reviews (CDSR) to measure strength of bodies of evidence and risk of bias (RoB) of individual studies, respectively. We report their calibration error and Brier scores, present their reliability diagrams, and analyze the risk-coverage trade-off in selective classification. RESULTS: The models are reasonably well calibrated on most quality criteria (expected calibration error [ECE] 0.04-0.09 for EvidenceGRADEr, 0.03-0.10 for RobotReviewer). However, we discover that both calibration and predictive performance vary significantly by medical area. This has ramifications for the application of such models in practice, as average performance is a poor indicator of group-level performance (e.g., health and safety at work, allergy and intolerance, and public health see much worse performance than cancer, pain, and anesthesia, and Neurology). We explore the reasons behind this disparity. CONCLUSION: Practitioners adopting automated quality assessment should expect large fluctuations in system reliability and predictive performance depending on the medical area. Prospective indicators of such behavior should be further researched.
- Published
- 2023
13. Stratification of keratoconus progression using unsupervised machine learning analysis of tomographical parameters
- Author
-
Cao, K, Verspoor, K, Chan, E, Daniell, M, Sahebjada, S, Baird, PN, Cao, K, Verspoor, K, Chan, E, Daniell, M, Sahebjada, S, and Baird, PN
- Abstract
Purpose: This study aimed to stratify eyes with keratoconus (KC) based on longitudinal changes in all Pentacam parameters into clusters using unsupervised machine learning, with the broader objective of more clearly defining the characteristics of KC progression. Methods: A data-driven cluster analysis (hierarchical clustering) was undertaken on a retrospective cohort of 1017 kC eyes and 128 control eyes. Clusters were derived using 6-month tomographical change in individual eyes from analysis of the reduced dimensionality parameter space using all available Pentacam parameters (406 principal components). The optimal number of clusters was determined by the clustering's capacity to discriminate progression between KC and control eyes based on change across parameters. One-way ANOVA was used to compare parameters between inferred clusters. Complete Pentacam data changes at 6, 12 and 18-month time points provided validation datasets to determine the generalizability of the clustering model. Results: We identified three clusters in KC progression patterns. Eyes designated within cluster 3 had the most rapidly changing tomographical parameters compared to eyes in either cluster 1 or 2. Eyes designated within cluster 1 reflected minimal changes in tomographical parameters, closest to the tomographical changes of control (non-KC) eyes. Thirty-nine corneal curvature parameters were identified and associated with these stratified clusters, with each of these parameters changing significantly different between three clusters. Similar clusters were identified at the 6, 12 and 18-month follow-up. Conclusions: The clustering model developed was able to automatically detect and categorize KC tomographical features into fast, slow, or limited change at different time points. This new KC stratification tool may provide an opportunity to provide a precision medicine approach to KC.
- Published
- 2023
14. Early prediction of incident liver disease using conventional risk factors and gut-microbiome-augmented gradient boosting
- Author
-
Liu, Y, Meric, G, Havulinna, AS, Teo, SM, Aberg, F, Ruuskanen, M, Sanders, J, Zhu, Q, Tripathi, A, Verspoor, K, Cheng, S, Jain, M, Jousilahti, P, Vazquez-Baeza, Y, Loomba, R, Lahti, L, Niiranen, T, Salomaa, V, Knight, R, Inouye, M, Liu, Y, Meric, G, Havulinna, AS, Teo, SM, Aberg, F, Ruuskanen, M, Sanders, J, Zhu, Q, Tripathi, A, Verspoor, K, Cheng, S, Jain, M, Jousilahti, P, Vazquez-Baeza, Y, Loomba, R, Lahti, L, Niiranen, T, Salomaa, V, Knight, R, and Inouye, M
- Abstract
The gut microbiome has shown promise as a predictive biomarker for various diseases. However, the potential of gut microbiota for prospective risk prediction of liver disease has not been assessed. Here, we utilized shallow shotgun metagenomic sequencing of a large population-based cohort (N > 7,000) with ∼15 years of follow-up in combination with machine learning to investigate the predictive capacity of gut microbial predictors individually and in conjunction with conventional risk factors for incident liver disease. Separately, conventional and microbial factors showed comparable predictive capacity. However, microbiome augmentation of conventional risk factors using machine learning significantly improved the performance. Similarly, disease-free survival analysis showed significantly improved stratification using microbiome-augmented models. Investigation of predictive microbial signatures revealed previously unknown taxa for liver disease, as well as those previously associated with hepatic function and disease. This study supports the potential clinical validity of gut metagenomic sequencing to complement conventional risk factors for prediction of liver diseases.
- Published
- 2022
15. COVID-19 Drug Repurposing: A Network-Based Framework for Exploring Biomedical Literature and Clinical Trials for Possible Treatments
- Author
-
Hamed, AA, Fandy, TE, Tkaczuk, KL, Verspoor, K, Lee, BS, Hamed, AA, Fandy, TE, Tkaczuk, KL, Verspoor, K, and Lee, BS
- Abstract
BACKGROUND: With the Coronavirus becoming a new reality of our world, global efforts continue to seek answers to many questions regarding the spread, variants, vaccinations, and medications. Particularly, with the emergence of several strains (e.g., Delta, Omicron), vaccines will need further development to offer complete protection against the new variants. It is critical to identify antiviral treatments while the development of vaccines continues. In this regard, the repurposing of already FDA-approved drugs remains a major effort. In this paper, we investigate the hypothesis that a combination of FDA-approved drugs may be considered as a candidate for COVID-19 treatment if (1) there exists an evidence in the COVID-19 biomedical literature that suggests such a combination, and (2) there is match in the clinical trials space that validates this drug combination. METHODS: We present a computational framework that is designed for detecting drug combinations, using the following components (a) a Text-mining module: to extract drug names from the abstract section of the biomedical publications and the intervention/treatment sections of clinical trial records. (b) a network model constructed from the drug names and their associations, (c) a clique similarity algorithm to identify candidate drug treatments. RESULT AND CONCLUSIONS: Our framework has identified treatments in the form of two, three, or four drug combinations (e.g., hydroxychloroquine, doxycycline, and azithromycin). The identifications of the various treatment candidates provided sufficient evidence that supports the trustworthiness of our hypothesis.
- Published
- 2022
16. Accuracy of Machine Learning Assisted Detection of Keratoconus: A Systematic Review and Meta-Analysis
- Author
-
Cao, K, Verspoor, K, Sahebjada, S, Baird, PN, Cao, K, Verspoor, K, Sahebjada, S, and Baird, PN
- Abstract
(1) Background: The objective of this review was to synthesize available data on the use of machine learning to evaluate its accuracy (as determined by pooled sensitivity and specificity) in detecting keratoconus (KC), and measure reporting completeness of machine learning models in KC based on TRIPOD (the transparent reporting of multivariable prediction models for individual prognosis or diagnosis) statement. (2) Methods: Two independent reviewers searched the electronic databases for all potential articles on machine learning and KC published prior to 2021. The TRIPOD 29-item checklist was used to evaluate the adherence to reporting guidelines of the studies, and the adherence rate to each item was computed. We conducted a meta-analysis to determine the pooled sensitivity and specificity of machine learning models for detecting KC. (3) Results: Thirty-five studies were included in this review. Thirty studies evaluated machine learning models for detecting KC eyes from controls and 14 studies evaluated machine learning models for detecting early KC eyes from controls. The pooled sensitivity for detecting KC was 0.970 (95% CI 0.949-0.982), with a pooled specificity of 0.985 (95% CI 0.971-0.993), whereas the pooled sensitivity of detecting early KC was 0.882 (95% CI 0.822-0.923), with a pooled specificity of 0.947 (95% CI 0.914-0.967). Between 3% and 48% of TRIPOD items were adhered to in studies, and the average (median) adherence rate for a single TRIPOD item was 23% across all studies. (4) Conclusions: Application of machine learning model has the potential to make the diagnosis and monitoring of KC more efficient, resulting in reduced vision loss to the patients. This review provides current information on the machine learning models that have been developed for detecting KC and early KC. Presently, the machine learning models performed poorly in identifying early KC from control eyes and many of these research studies did not follow established reporting standa
- Published
- 2022
17. Large-scale protein-protein post-translational modification extraction with distant supervision and confidence calibrated BioBERT
- Author
-
Elangovan, A, Li, Y, Pires, DE, Davis, MJ, Verspoor, K, Elangovan, A, Li, Y, Pires, DE, Davis, MJ, and Verspoor, K
- Abstract
MOTIVATION: Protein-protein interactions (PPIs) are critical to normal cellular function and are related to many disease pathways. A range of protein functions are mediated and regulated by protein interactions through post-translational modifications (PTM). However, only 4% of PPIs are annotated with PTMs in biological knowledge databases such as IntAct, mainly performed through manual curation, which is neither time- nor cost-effective. Here we aim to facilitate annotation by extracting PPIs along with their pairwise PTM from the literature by using distantly supervised training data using deep learning to aid human curation. METHOD: We use the IntAct PPI database to create a distant supervised dataset annotated with interacting protein pairs, their corresponding PTM type, and associated abstracts from the PubMed database. We train an ensemble of BioBERT models-dubbed PPI-BioBERT-x10-to improve confidence calibration. We extend the use of ensemble average confidence approach with confidence variation to counteract the effects of class imbalance to extract high confidence predictions. RESULTS AND CONCLUSION: The PPI-BioBERT-x10 model evaluated on the test set resulted in a modest F1-micro 41.3 (P =5 8.1, R = 32.1). However, by combining high confidence and low variation to identify high quality predictions, tuning the predictions for precision, we retained 19% of the test predictions with 100% precision. We evaluated PPI-BioBERT-x10 on 18 million PubMed abstracts and extracted 1.6 million (546507 unique PTM-PPI triplets) PTM-PPI predictions, and filter [Formula: see text] (4584 unique) high confidence predictions. Of the 5700, human evaluation on a small randomly sampled subset shows that the precision drops to 33.7% despite confidence calibration and highlights the challenges of generalisability beyond the test set even with confidence calibration. We circumvent the problem by only including predictions associated with multiple papers, improving the precision to 5
- Published
- 2022
18. What does it take to bake a cake? The RecipeRef corpus and anaphora resolution in procedural text
- Author
-
Fang, B, Baldwin, T, Verspoor, K, Fang, B, Baldwin, T, and Verspoor, K
- Published
- 2022
19. Cross-linguistic comparison of linguistic feature encoding in BERT models for typologically different languages
- Author
-
Otmakhova, Y, Verspoor, K, Lau, JH, Otmakhova, Y, Verspoor, K, and Lau, JH
- Abstract
Though recently there have been an increased interest in how pre-trained language models encode different linguistic features, there is still a lack of systematic comparison between languages with different morphology and syntax. In this paper, using BERT as an example of a pre-trained model, we compare how three typologically different languages (English, Korean, and Russian) encode morphology and syntax features across different layers. In particular, we contrast languages which differ in a particular aspect, such as flexibility of word order, head directionality, morphological type, presence of grammatical gender, and morphological richness, across four different tasks.
- Published
- 2022
20. Tasks as needs: reframing the paradigm of clinical natural language processing research for real-world decision support
- Author
-
Lederman, A, Lederman, R, Verspoor, K, Lederman, A, Lederman, R, and Verspoor, K
- Abstract
Electronic medical records are increasingly used to store patient information in hospitals and other clinical settings. There has been a corresponding proliferation of clinical natural language processing (cNLP) systems aimed at using text data in these records to improve clinical decision-making, in comparison to manual clinician search and clinical judgment alone. However, these systems have delivered marginal practical utility and are rarely deployed into healthcare settings, leading to proposals for technical and structural improvements. In this paper, we argue that this reflects a violation of Friedman's "Fundamental Theorem of Biomedical Informatics," and that a deeper epistemological change must occur in the cNLP field, as a parallel step alongside any technical or structural improvements. We propose that researchers shift away from designing cNLP systems independent of clinical needs, in which cNLP tasks are ends in themselves-"tasks as decisions"-and toward systems that are directly guided by the needs of clinicians in realistic decision-making contexts-"tasks as needs." A case study example illustrates the potential benefits of developing cNLP systems that are designed to more directly support clinical needs.
- Published
- 2022
21. Uncertainty Estimation and Reduction of Pre-trained Models for Text Regression
- Author
-
Wang, Y, Beck, D, Baldwin, T, Verspoor, K, Wang, Y, Beck, D, Baldwin, T, and Verspoor, K
- Abstract
State-of-the-art classification and regression models are often not well calibrated, and cannot reliably provide uncertainty estimates, limiting their utility in safety-critical applications such as clinical decision-making. While recent work has focused on calibration of classifiers, there is almost no work in NLP on calibration in a regression setting. In this paper, we quantify the calibration of pretrained language models for text regression, both intrinsically and extrinsically. We further apply uncertainty estimates to augment training data in low-resource domains. Our experiments on three regression tasks in both self-training and active-learning settings show that uncertainty estimation can be used to increase overall performance and enhance model generalization.
- Published
- 2022
22. Exploring automatic inconsistency detection for literature-based gene ontology annotation
- Author
-
Chen, J, Goudey, B, Zobel, J, Geard, N, Verspoor, K, Chen, J, Goudey, B, Zobel, J, Geard, N, and Verspoor, K
- Abstract
MOTIVATION: Literature-based gene ontology annotations (GOA) are biological database records that use controlled vocabulary to uniformly represent gene function information that is described in the primary literature. Assurance of the quality of GOA is crucial for supporting biological research. However, a range of different kinds of inconsistencies in between literature as evidence and annotated GO terms can be identified; these have not been systematically studied at record level. The existing manual-curation approach to GOA consistency assurance is inefficient and is unable to keep pace with the rate of updates to gene function knowledge. Automatic tools are therefore needed to assist with GOA consistency assurance. This article presents an exploration of different GOA inconsistencies and an early feasibility study of automatic inconsistency detection. RESULTS: We have created a reliable synthetic dataset to simulate four realistic types of GOA inconsistency in biological databases. Three automatic approaches are proposed. They provide reasonable performance on the task of distinguishing the four types of inconsistency and are directly applicable to detect inconsistencies in real-world GOA database records. Major challenges resulting from such inconsistencies in the context of several specific application settings are reported. This is the first study to introduce automatic approaches that are designed to address the challenges in current GOA quality assurance workflows. The data underlying this article are available in Github at https://github.com/jiyuc/AutoGOAConsistency.
- Published
- 2022
23. The patient is more dead than alive: exploring the current state of the multi-document summarization of the biomedical literature
- Author
-
Otmakhova, Y, Verspoor, K, Baldwin, T, Lau, JH, Otmakhova, Y, Verspoor, K, Baldwin, T, and Lau, JH
- Published
- 2022
24. Overcoming challenges in extracting prescribing habits from veterinary clinics using big data and deep learning
- Author
-
Hur, B, Hardefeldt, LY, Verspoor, K, Baldwin, T, Gilkerson, JR, Hur, B, Hardefeldt, LY, Verspoor, K, Baldwin, T, and Gilkerson, JR
- Abstract
Understanding antimicrobial usage patterns and encouraging appropriate antimicrobial usage is a critical component of antimicrobial stewardship. Studies using VetCompass Australia and Natural Language Processing (NLP) have demonstrated antimicrobial usage patterns in companion animal practices across Australia. Doing so has highlighted the many obstacles and barriers to the task of converting raw clinical notes into a format that can be readily queried and analysed. We developed NLP systems using rules-based algorithms and machine learning to automate the extraction of data describing the key elements to assess appropriate antimicrobial use. These included the clinical indication, antimicrobial agent selection, dose and duration of therapy. Our methods were applied to over 4.4 million companion animal clinical records across Australia on all consultations with antimicrobial use to help us understand what antibiotics are being given and why on a population level. Of these, approximately only 40% recorded the reason why antimicrobials were prescribed, along with the dose and duration of treatment. NLP and deep learning might be able to overcome the difficulties of harvesting free text data from clinical records, but when the essential data are not recorded in the clinical records, then, this becomes an insurmountable obstacle.
- Published
- 2022
25. Propagation, detection and correction of errors using the sequence database network
- Author
-
Goudey, B, Geard, N, Verspoor, K, Zobel, J, Goudey, B, Geard, N, Verspoor, K, and Zobel, J
- Abstract
Nucleotide and protein sequences stored in public databases are the cornerstone of many bioinformatics analyses. The records containing these sequences are prone to a wide range of errors, including incorrect functional annotation, sequence contamination and taxonomic misclassification. One source of information that can help to detect errors are the strong interdependency between records. Novel sequences in one database draw their annotations from existing records, may generate new records in multiple other locations and will have varying degrees of similarity with existing records across a range of attributes. A network perspective of these relationships between sequence records, within and across databases, offers new opportunities to detect-or even correct-erroneous entries and more broadly to make inferences about record quality. Here, we describe this novel perspective of sequence database records as a rich network, which we call the sequence database network, and illustrate the opportunities this perspective offers for quantification of database quality and detection of spurious entries. We provide an overview of the relevant databases and describe how the interdependencies between sequence records across these databases can be exploited by network analyses. We review the process of sequence annotation and provide a classification of sources of error, highlighting propagation as a major source. We illustrate the value of a network perspective through three case studies that use network analysis to detect errors, and explore the quality and quantity of critical relationships that would inform such network analyses. This systematic description of a network perspective of sequence database records provides a novel direction to combat the proliferation of errors within these critical bioinformatics resources.
- Published
- 2022
26. Predicting Publication of Clinical Trials Using Structured and Unstructured Data: Model Development and Validation Study.
- Author
-
Wang, S, Šuster, S, Baldwin, T, Verspoor, K, Wang, S, Šuster, S, Baldwin, T, and Verspoor, K
- Abstract
BACKGROUND: Publication of registered clinical trials is a critical step in the timely dissemination of trial findings. However, a significant proportion of completed clinical trials are never published, motivating the need to analyze the factors behind success or failure to publish. This could inform study design, help regulatory decision-making, and improve resource allocation. It could also enhance our understanding of bias in the publication of trials and publication trends based on the research direction or strength of the findings. Although the publication of clinical trials has been addressed in several descriptive studies at an aggregate level, there is a lack of research on the predictive analysis of a trial's publishability given an individual (planned) clinical trial description. OBJECTIVE: We aimed to conduct a study that combined structured and unstructured features relevant to publication status in a single predictive approach. Established natural language processing techniques as well as recent pretrained language models enabled us to incorporate information from the textual descriptions of clinical trials into a machine learning approach. We were particularly interested in whether and which textual features could improve the classification accuracy for publication outcomes. METHODS: In this study, we used metadata from ClinicalTrials.gov (a registry of clinical trials) and MEDLINE (a database of academic journal articles) to build a data set of clinical trials (N=76,950) that contained the description of a registered trial and its publication outcome (27,702/76,950, 36% published and 49,248/76,950, 64% unpublished). This is the largest data set of its kind, which we released as part of this work. The publication outcome in the data set was identified from MEDLINE based on clinical trial identifiers. We carried out a descriptive analysis and predicted the publication outcome using 2 approaches: a neural network with a large domain-specific language mo
- Published
- 2022
27. Overview of ChEMU 2022 Evaluation Campaign: Information Extraction in Chemical Patents
- Author
-
Barron-Cedeno, A, DaSanMartino, G, Esposti, MD, Sebastiani, F, Macdonald, C, Pasi, G, Hanbury, A, Potthast, M, Faggioli, G, Ferro, N, Li, Y, Fang, B, He, J, Yoshikawa, H, Akhondi, SA, Druckenbrodt, C, Thorne, C, Afzal, Z, Zhai, Z, Baldwin, T, Verspoor, K, Barron-Cedeno, A, DaSanMartino, G, Esposti, MD, Sebastiani, F, Macdonald, C, Pasi, G, Hanbury, A, Potthast, M, Faggioli, G, Ferro, N, Li, Y, Fang, B, He, J, Yoshikawa, H, Akhondi, SA, Druckenbrodt, C, Thorne, C, Afzal, Z, Zhai, Z, Baldwin, T, and Verspoor, K
- Published
- 2022
28. Cross-modal Clinical Graph Transformer for Ophthalmic Report Generation
- Author
-
Li, M, Cai, W, Verspoor, K, Pan, S, Liang, X, Chang, X, Li, M, Cai, W, Verspoor, K, Pan, S, Liang, X, and Chang, X
- Abstract
Automatic generation of ophthalmic reports using datadriven neural networks has great potential in clinical practice. When writing a report, ophthalmologists make inferences with prior clinical knowledge. This knowledge has been neglected in prior medical report generation methods. To endow models with the capability of incorporating expert knowledge, we propose a Cross-modal clinical Graph Transformer (CGT) for ophthalmic report generation (ORG), in which clinical relation triples are injected into the visual features as prior knowledge to drive the decoding procedure. However, two major common Knowledge Noise (KN) issues may affect models' effectiveness. 1) Existing general biomedical knowledge bases such as the UMLS may not align meaningfully to the specific context and language of the report, limiting their utility for knowledge injection. 2) Incorporating too much knowledge may divert the visual features from their correct meaning. To overcome these limitations, we design an automatic information extraction scheme based on natural language processing to obtain clinical entities and relations directly from in-domain training reports. Given a set of ophthalmic images, our CGT first restores a sub-graph from the clinical graph and injects the restored triples into visual features. Then visible matrix is employed during the encoding procedure to limit the impact of knowledge. Finally, reports are predicted by the encoded cross-modal features via a Transformer decoder. Extensive experiments on the large-scale FFA-IR benchmark demonstrate that the proposed CGT is able to outperform previous benchmark methods and achieve state-of-the-art performances.
- Published
- 2022
29. Overcoming challenges in extracting prescribing habits from veterinary clinics using big data and deep learning
- Author
-
Hur, B, primary, Hardefeldt, LY, additional, Verspoor, K, additional, Baldwin, T, additional, and Gilkerson, JR, additional
- Published
- 2022
- Full Text
- View/download PDF
30. Use and validation of text mining and cluster algorithms to derive insights from Corona Virus Disease-2019 (COVID-19) medical literature
- Author
-
Reddy, Sandeep, Bhaskar, R, Padmanabhan, S, Verspoor, K, Mamillapalli, C, Lahoti, R, Makinen, V-P, Pradhan, S, Kushwah, P, Sinha, S, Reddy, Sandeep, Bhaskar, R, Padmanabhan, S, Verspoor, K, Mamillapalli, C, Lahoti, R, Makinen, V-P, Pradhan, S, Kushwah, P, and Sinha, S
- Published
- 2021
31. ChemTables: a dataset for semantic classification on tables in chemical patents
- Author
-
Zhai, Z, Druckenbrodt, C, Thorne, C, Akhondi, SA, Dat, QN, Cohn, T, Verspoor, K, Zhai, Z, Druckenbrodt, C, Thorne, C, Akhondi, SA, Dat, QN, Cohn, T, and Verspoor, K
- Abstract
Chemical patents are a commonly used channel for disclosing novel compounds and reactions, and hence represent important resources for chemical and pharmaceutical research. Key chemical data in patents is often presented in tables. Both the number and the size of tables can be very large in patent documents. In addition, various types of information can be presented in tables in patents, including spectroscopic and physical data, or pharmacological use and effects of chemicals. Since images of Markush structures and merged cells are commonly used in these tables, their structure also shows substantial variation. This heterogeneity in content and structure of tables in chemical patents makes relevant information difficult to find. We therefore propose a new text mining task of automatically categorising tables in chemical patents based on their contents. Categorisation of tables based on the nature of their content can help to identify tables containing key information, improving the accessibility of information in patents that is highly relevant for new inventions. For developing and evaluating methods for the table classification task, we developed a new dataset, called CHEMTABLES, which consists of 788 chemical patent tables with labels of their content type. We introduce this data set in detail. We further establish strong baselines for the table classification task in chemical patents by applying state-of-the-art neural network models developed for natural language processing, including TabNet, ResNet and Table-BERT on CHEMTABLES. The best performing model, Table-BERT, achieves a performance of 88.66 micro-averaged [Formula: see text] score on the table classification task. The CHEMTABLES dataset is publicly available at https://doi.org/10.17632/g7tjh7tbrj.3 , subject to the CC BY NC 3.0 license. Code/models evaluated in this work are in a Github repository https://github.com/zenanz/ChemTables .
- Published
- 2021
32. Use and validation of text mining and cluster algorithms to derive insights from Corona Virus Disease-2019 (COVID-19) medical literature.
- Author
-
Reddy, S, Bhaskar, R, Padmanabhan, S, Verspoor, K, Mamillapalli, C, Lahoti, R, Makinen, V-P, Pradhan, S, Kushwah, P, Sinha, S, Reddy, S, Bhaskar, R, Padmanabhan, S, Verspoor, K, Mamillapalli, C, Lahoti, R, Makinen, V-P, Pradhan, S, Kushwah, P, and Sinha, S
- Abstract
The emergence of the severe acute respiratory syndrome coronavirus 2 (SARS-CoV-2) late last year has not only led to the world-wide coronavirus disease 2019 (COVID-19) pandemic but also a deluge of biomedical literature. Following the release of the COVID-19 open research dataset (CORD-19) comprising over 200,000 scholarly articles, we a multi-disciplinary team of data scientists, clinicians, medical researchers and software engineers developed an innovative natural language processing (NLP) platform that combines an advanced search engine with a biomedical named entity recognition extraction package. In particular, the platform was developed to extract information relating to clinical risk factors for COVID-19 by presenting the results in a cluster format to support knowledge discovery. Here we describe the principles behind the development, the model and the results we obtained.
- Published
- 2021
33. Machine learning with a reduced dimensionality representation of comprehensive Pentacam tomography parameters to identify subclinical keratoconus
- Author
-
Cao, K, Verspoor, K, Chan, E, Daniell, M, Sahebjada, S, Baird, PN, Cao, K, Verspoor, K, Chan, E, Daniell, M, Sahebjada, S, and Baird, PN
- Abstract
PURPOSE: To investigate the performance of a machine learning model based on a reduced dimensionality parameter space derived from complete Pentacam parameters to identify subclinical keratoconus (KC). METHODS: All 1692 available parameters were obtained from the Pentacam imaging machine on 145 subclinical KC and 122 control eyes. We applied a principal component analysis (PCA) to the complete Pentacam dataset to reduce its parameter dimensionality. Subsequently, we investigated machine learning performance of the random forest algorithm with increasing numbers of components to identify their optimal number for detecting subclinical KC from control eyes. RESULTS: The dimensionality of the complete set of 1692 Pentacam parameters was reduced to 267 principal components using PCA. Subsequent selection of 15 of these principal components explained over 85% of the variance of the original Pentacam-derived parameters and input to train a random forest machine learning model to achieve the best accuracy of 98% in detecting subclinical KC eyes. The model established also reached a high sensitivity of 97% in identification of subclinical KC and a specificity of 98% in recognizing control eyes. CONCLUSIONS: A random forest-based model trained using a modest number of components derived from a reduced dimensionality representation of complete Pentacam system parameters allowed for high accuracy of subclinical KC identification.
- Published
- 2021
34. ChEMU 2020: Natural Language Processing Methods Are Effective for Information Extraction From Chemical Patents.
- Author
-
He, J, Nguyen, DQ, Akhondi, SA, Druckenbrodt, C, Thorne, C, Hoessel, R, Afzal, Z, Zhai, Z, Fang, B, Yoshikawa, H, Albahem, A, Cavedon, L, Cohn, T, Baldwin, T, Verspoor, K, He, J, Nguyen, DQ, Akhondi, SA, Druckenbrodt, C, Thorne, C, Hoessel, R, Afzal, Z, Zhai, Z, Fang, B, Yoshikawa, H, Albahem, A, Cavedon, L, Cohn, T, Baldwin, T, and Verspoor, K
- Abstract
Chemical patents represent a valuable source of information about new chemical compounds, which is critical to the drug discovery process. Automated information extraction over chemical patents is, however, a challenging task due to the large volume of existing patents and the complex linguistic properties of chemical patents. The Cheminformatics Elsevier Melbourne University (ChEMU) evaluation lab 2020, part of the Conference and Labs of the Evaluation Forum 2020 (CLEF2020), was introduced to support the development of advanced text mining techniques for chemical patents. The ChEMU 2020 lab proposed two fundamental information extraction tasks focusing on chemical reaction processes described in chemical patents: (1) chemical named entity recognition, requiring identification of essential chemical entities and their roles in chemical reactions, as well as reaction conditions; and (2) event extraction, which aims at identification of event steps relating the entities involved in chemical reactions. The ChEMU 2020 lab received 37 team registrations and 46 runs. Overall, the performance of submissions for these tasks exceeded our expectations, with the top systems outperforming strong baselines. We further show the methods to be robust to variations in sampling of the test data. We provide a detailed overview of the ChEMU 2020 corpus and its annotation, showing that inter-annotator agreement is very strong. We also present the methods adopted by participants, provide a detailed analysis of their performance, and carefully consider the potential impact of data leakage on interpretation of the results. The ChEMU 2020 Lab has shown the viability of automated methods to support information extraction of key information in chemical patents.
- Published
- 2021
35. Memorization vs. Generalization: Quantifying Data Leakage in NLP Performance Evaluation
- Author
-
Elangovan, A, He, J, Verspoor, K, Elangovan, A, He, J, and Verspoor, K
- Published
- 2021
36. Automatic consistency assurance for literature-based gene ontology annotation
- Author
-
Chen, J, Geard, N, Zobel, J, Verspoor, K, Chen, J, Geard, N, Zobel, J, and Verspoor, K
- Abstract
BACKGROUND: Literature-based gene ontology (GO) annotation is a process where expert curators use uniform expressions to describe gene functions reported in research papers, creating computable representations of information about biological systems. Manual assurance of consistency between GO annotations and the associated evidence texts identified by expert curators is reliable but time-consuming, and is infeasible in the context of rapidly growing biological literature. A key challenge is maintaining consistency of existing GO annotations as new studies are published and the GO vocabulary is updated. RESULTS: In this work, we introduce a formalisation of biological database annotation inconsistencies, identifying four distinct types of inconsistency. We propose a novel and efficient method using state-of-the-art text mining models to automatically distinguish between consistent GO annotation and the different types of inconsistent GO annotation. We evaluate this method using a synthetic dataset generated by directed manipulation of instances in an existing corpus, BC4GO. We provide detailed error analysis for demonstrating that the method achieves high precision on more confident predictions. CONCLUSIONS: Two models built using our method for distinct annotation consistency identification tasks achieved high precision and were robust to updates in the GO vocabulary. Our approach demonstrates clear value for human-in-the-loop curation scenarios.
- Published
- 2021
37. Early prediction of diagnostic-related groups and estimation of hospital cost by processing clinical notes
- Author
-
Liu, J, Capurro, D, Nguyen, A, Verspoor, K, Liu, J, Capurro, D, Nguyen, A, and Verspoor, K
- Abstract
As healthcare providers receive fixed amounts of reimbursement for given services under DRG (Diagnosis-Related Groups) payment, DRG codes are valuable for cost monitoring and resource allocation. However, coding is typically performed retrospectively post-discharge. We seek to predict DRGs and DRG-based case mix index (CMI) at early inpatient admission using routine clinical text to estimate hospital cost in an acute setting. We examined a deep learning-based natural language processing (NLP) model to automatically predict per-episode DRGs and corresponding cost-reflecting weights on two cohorts (paid under Medicare Severity (MS) DRG or All Patient Refined (APR) DRG), without human coding efforts. It achieved macro-averaged area under the receiver operating characteristic curve (AUC) scores of 0·871 (SD 0·011) on MS-DRG and 0·884 (0·003) on APR-DRG in fivefold cross-validation experiments on the first day of ICU admission. When extended to simulated patient populations to estimate average cost-reflecting weights, the model increased its accuracy over time and obtained absolute CMI error of 2·40 (1·07%) and 12·79% (2·31%), respectively on the first day. As the model could adapt to variations in admission time, cohort size, and requires no extra manual coding efforts, it shows potential to help estimating costs for active patients to support better operational decision-making in hospitals.
- Published
- 2021
38. ChEMU-Ref: A corpus for modeling anaphora resolution in the chemical domain
- Author
-
Fang, B, Druckenbrodt, C, Akhondi, SA, He, J, Baldwin, T, Verspoor, K, Fang, B, Druckenbrodt, C, Akhondi, SA, He, J, Baldwin, T, and Verspoor, K
- Abstract
Chemical patents contain rich coreference and bridging links, which are the target of this research. Specially, we introduce a novel annotation scheme, based on which we create the ChEMU-Ref dataset from reaction description snippets in English-language chemical patents. We propose a neural approach to anaphora resolution, which we show to achieve strong results, especially when jointly trained over coreference and bridging links.
- Published
- 2021
39. Bow-tie architecture of gene regulatory networks in species of varying complexity
- Author
-
Roy, GG, He, S, Geard, N, Verspoor, K, Roy, GG, He, S, Geard, N, and Verspoor, K
- Abstract
The gene regulatory network (GRN) architecture plays a key role in explaining the biological differences between species. We aim to understand species differences in terms of some universally present dynamical properties of their gene regulatory systems. A network architectural feature associated with controlling system-level dynamical properties is the bow-tie, identified by a strongly connected subnetwork, the core layer, between two sets of nodes, the in and the out layers. Though a bow-tie architecture has been observed in many networks, its existence has not been extensively investigated in GRNs of species of widely varying biological complexity. We analyse publicly available GRNs of several well-studied species from prokaryotes to unicellular eukaryotes to multicellular organisms. In their GRNs, we find the existence of a bow-tie architecture with a distinct largest strongly connected core layer. We show that the bow-tie architecture is a characteristic feature of GRNs. We observe an increasing trend in the relative core size with species complexity. Using studied relationships of the core size with dynamical properties like robustness and fragility, flexibility, criticality, controllability and evolvability, we hypothesize how these regulatory system properties have emerged differently with biological complexity, based on the observed differences of the GRN bow-tie architectures.
- Published
- 2021
40. The Evolution of Clinical Knowledge During COVID-19: Towards a Global Learning Health System.
- Author
-
Verspoor, K and Verspoor, K
- Abstract
OBJECTIVES: We examine the knowledge ecosystem of COVID-19, focusing on clinical knowledge and the role of health informatics as enabling technology. We argue for commitment to the model of a global learning health system to facilitate rapid knowledge translation supporting health care decision making in the face of emerging diseases. METHODS AND RESULTS: We frame the evolution of knowledge in the COVID-19 crisis in terms of learning theory, and present a view of what has occurred during the pandemic to rapidly derive and share knowledge as an (underdeveloped) instance of a global learning health system. We identify the key role of information technologies for electronic data capture and data sharing, computational modelling, evidence synthesis, and knowledge dissemination. We further highlight gaps in the system and barriers to full realisation of an efficient and effective global learning health system. CONCLUSIONS: The need for a global knowledge ecosystem supporting rapid learning from clinical practice has become more apparent than ever during the COVID-19 pandemic. Continued effort to realise the vision of a global learning health system, including establishing effective approaches to data governance and ethics to support the system, is imperative to enable continuous improvement in our clinical care.
- Published
- 2021
41. Measuring emotions in the COVID-19 real world worry dataset
- Author
-
Kleinberg, Bennett, van der Vegt, Isabelle, Mozes, Maximilian, Verspoor, K., Bretonnel Cohen, K., Dredze, M., Ferrara, E., May, J., Munro, R., Paris, C., Wallace, B., and Department of Methodology and Statistics
- Abstract
The COVID-19 pandemic is having a dramatic impact on societies and economies around the world. With various measures of lockdowns and social distancing in place, it becomes important to understand emotional responses on a large scale. In this paper, we present the first ground truth dataset of emotional responses to COVID-19. We asked participants to indicate their emotions and express these in text. This resulted in the Real World Worry Dataset of 5,000 texts (2,500 short + 2,500 long texts). Our analyses suggest that emotional responses correlated with linguistic measures. Topic modeling further revealed that people in the UK worry about their family and the economic situation. Tweet-sized texts functioned as a call for solidarity, while longer texts shed light on worries and concerns. Using predictive modeling approaches, we were able to approximate the emotional responses of participants from text within 14% of their actual value. We encourage others to use the dataset and improve how we can use automated methods to learn about emotional responses and worries about an urgent problem.
- Published
- 2020
42. PoLoBag: Polynomial Lasso Bagging for signed gene regulatory network inference from expression data.
- Author
-
Luigi Martelli, P, Roy, GG, Geard, N, Verspoor, K, He, S, Luigi Martelli, P, Roy, GG, Geard, N, Verspoor, K, and He, S
- Abstract
MOTIVATION: Inferring gene regulatory networks (GRNs) from expression data is a significant systems biology problem. A useful inference algorithm should not only unveil the global structure of the regulatory mechanisms but also the details of regulatory interactions such as edge direction (from regulator to target) and sign (activation/inhibition). Many popular GRN inference algorithms cannot infer edge signs, and those that can infer signed GRNs cannot simultaneously infer edge directions or network cycles. RESULTS: To address these limitations of existing algorithms we propose Polynomial Lasso Bagging (PoLoBag) for signed GRN inference with both edge directions and network cycles. PoLoBag is an ensemble regression algorithm in a bagging framework where Lasso weights estimated on bootstrap samples are averaged. These bootstrap samples incorporate polynomial features to capture higher order interactions. Results demonstrate that PoLoBag is consistently more accurate for signed inference than state-of-the-art algorithms on simulated and real-world expression datasets. AVAILABILITY: Algorithm and data are freely available at https://github.com/gourabghoshroy/PoLoBag. SUPPLEMENTARY INFORMATION: Supplementary data are available at Bioinformatics online.
- Published
- 2020
43. Domain Adaptation and Instance Selection for Disease Syndrome Classification over Veterinary Clinical Notes
- Author
-
Hur, B, Baldwin, T, Verspoor, K, Hardefeldt, L, Gilkerson, J, Hur, B, Baldwin, T, Verspoor, K, Hardefeldt, L, and Gilkerson, J
- Published
- 2020
44. Development of a Self-Harm Monitoring System for Victoria
- Author
-
Robinson, J, Witt, K, Lamblin, M, Spittal, MJ, Carter, G, Verspoor, K, Page, A, Rajaram, G, Rozova, V, Hill, NTM, Pirkis, J, Bleeker, C, Pleban, A, Knott, JC, Robinson, J, Witt, K, Lamblin, M, Spittal, MJ, Carter, G, Verspoor, K, Page, A, Rajaram, G, Rozova, V, Hill, NTM, Pirkis, J, Bleeker, C, Pleban, A, and Knott, JC
- Abstract
The prevention of suicide and suicide-related behaviour are key policy priorities in Australia and internationally. The World Health Organization has recommended that member states develop self-harm surveillance systems as part of their suicide prevention efforts. This is also a priority under Australia's Fifth National Mental Health and Suicide Prevention Plan. The aim of this paper is to describe the development of a state-based self-harm monitoring system in Victoria, Australia. In this system, data on all self-harm presentations are collected from eight hospital emergency departments in Victoria. A natural language processing classifier that uses machine learning to identify episodes of self-harm is currently being developed. This uses the free-text triage case notes, together with certain structured data fields, contained within the metadata of the incoming records. Post-processing is undertaken to identify primary mechanism of injury, substances consumed (including alcohol, illicit drugs and pharmaceutical preparations) and presence of psychiatric disorders. This system will ultimately leverage routinely collected data in combination with advanced artificial intelligence methods to support robust community-wide monitoring of self-harm. Once fully operational, this system will provide accurate and timely information on all presentations to participating emergency departments for self-harm, thereby providing a useful indicator for Australia's suicide prevention efforts.
- Published
- 2020
45. Artificial intelligence for clinical decision support in neurology
- Author
-
Pedersen, M, Verspoor, K, Jenkinson, M, Law, M, Abbott, DF, Jackson, GD, Pedersen, M, Verspoor, K, Jenkinson, M, Law, M, Abbott, DF, and Jackson, GD
- Abstract
Artificial intelligence is one of the most exciting methodological shifts in our era. It holds the potential to transform healthcare as we know it, to a system where humans and machines work together to provide better treatment for our patients. It is now clear that cutting edge artificial intelligence models in conjunction with high-quality clinical data will lead to improved prognostic and diagnostic models in neurological disease, facilitating expert-level clinical decision tools across healthcare settings. Despite the clinical promise of artificial intelligence, machine and deep-learning algorithms are not a one-size-fits-all solution for all types of clinical data and questions. In this article, we provide an overview of the core concepts of artificial intelligence, particularly contemporary deep-learning methods, to give clinician and neuroscience researchers an appreciation of how artificial intelligence can be harnessed to support clinical decisions. We clarify and emphasize the data quality and the human expertise needed to build robust clinical artificial intelligence models in neurology. As artificial intelligence is a rapidly evolving field, we take the opportunity to iterate important ethical principles to guide the field of medicine is it moves into an artificial intelligence enhanced future.
- Published
- 2020
46. Evaluating the Performance of Various Machine Learning Algorithms to Detect Subclinical Keratoconus
- Author
-
Cao, K, Verspoor, K, Sahebjada, S, Baird, PN, Cao, K, Verspoor, K, Sahebjada, S, and Baird, PN
- Abstract
Purpose: Keratoconus (KC) represents one of the leading causes of corneal transplantation worldwide. Detecting subclinical KC would lead to better management to avoid the need for corneal grafts, but the condition is clinically challenging to diagnose. We wished to compare eight commonly used machine learning algorithms using a range of parameter combinations by applying them to our KC dataset and build models to better differentiate subclinical KC from non-KC eyes. Methods: Oculus Pentacam was used to obtain corneal parameters on 49 subclinical KC and 39 control eyes, along with clinical and demographic parameters. Eight machine learning methods were applied to build models to differentiate subclinical KC from control eyes. Dominant algorithms were trained with all combinations of the considered parameters to select important parameter combinations. The performance of each model was evaluated and compared. Results: Using a total of eleven parameters, random forest, support vector machine and k-nearest neighbors had better performance in detecting subclinical KC. The highest area under the curve of 0.97 for detecting subclinical KC was achieved using five parameters by the random forest method. The highest sensitivity (0.94) and specificity (0.90) were obtained by the support vector machine and the k-nearest neighbor model, respectively. Conclusions: This study showed machine learning algorithms can be applied to identify subclinical KC using a minimal parameter set that are routinely collected during clinical eye examination. Translational Relevance: Machine learning algorithms can be built using routinely collected clinical parameters that will assist in the objective detection of subclinical KC.
- Published
- 2020
47. Quality Matters: Biocuration Experts on the Impact of Duplication and Other Data Quality Issues in Biological Databases.
- Author
-
Chen, Q, Britto, R, Erill, I, Jeffery, CJ, Liberzon, A, Magrane, M, Onami, J-I, Robinson-Rechavi, M, Sponarova, J, Zobel, J, Verspoor, K, Chen, Q, Britto, R, Erill, I, Jeffery, CJ, Liberzon, A, Magrane, M, Onami, J-I, Robinson-Rechavi, M, Sponarova, J, Zobel, J, and Verspoor, K
- Abstract
Biological databases represent an extraordinary collective volume of work. Diligently built up over decades and comprising many millions of contributions from the biomedical research community, biological databases provide worldwide access to a massive number of records (also known as entries) [1]. Starting from individual laboratories, genomes are sequenced, assembled, annotated, and ultimately submitted to primary nucleotide databases such as GenBank [2], European Nucleotide Archive (ENA) [3], and DNA Data Bank of Japan (DDBJ) [4] (collectively known as the International Nucleotide Sequence Database Collaboration, INSDC). Protein records, which are the translations of these nucleotide records, are deposited into central protein databases such as the UniProt KnowledgeBase (UniProtKB) [5] and the Protein Data Bank (PDB) [6]. Sequence records are further accumulated into different databases for more specialized purposes: RFam [7] and PFam [8] for RNA and protein families, respectively; DictyBase [9] and PomBase [10] for model organisms; as well as ArrayExpress [11] and Gene Expression Omnibus (GEO) [12] for gene expression profiles. These databases are selected as examples; the list is not intended to be exhaustive. However, they are representative of biological databases that have been named in the “golden set” of the 24th Nucleic Acids Research database issue (in 2016). The introduction of that issue highlights the databases that “consistently served as authoritative, comprehensive, and convenient data resources widely used by the entire community and offer some lessons on what makes a successful database” [13]. In addition, the associated information about sequences is also propagated into non-sequence databases, such as PubMed (https://www.ncbi.nlm.nih.gov/pubmed/) for scientific literature or Gene Ontology (GO) [14] for function annotations. These databases in turn benefit individual studies, many of which use these publicly available records as the basis for th
- Published
- 2020
48. Improved Topic Representations of Medical Documents to Assist COVID-19 Literature Exploration
- Author
-
Otmakhova, Y, Verspoor, K, Baldwin, T, Šuster, S, Otmakhova, Y, Verspoor, K, Baldwin, T, and Šuster, S
- Published
- 2020
49. Testing Contextualized Word Embeddings to Improve NER in Spanish Clinical Case Narratives
- Author
-
Akhtyamova, L, Martinez, P, Verspoor, K, Cardiff, J, Akhtyamova, L, Martinez, P, Verspoor, K, and Cardiff, J
- Abstract
In the Big Data era, there is an increasing need to fully exploit and analyze the huge quantity of information available about health. Natural Language Processing (NLP) technologies can contribute by extracting relevant information from unstructured data contained in Electronic Health Records (EHR) such as clinical notes, patients’ discharge summaries and radiology reports. The extracted information can help in health-related decision making processes. The Named Entity Recognition (NER) task, which detects important concepts in texts (e.g., diseases, symptoms, drugs, etc.), is crucial in the information extraction process yet has received little attention in languages other than English. In this work, we develop a deep learning-based NLP pipeline for biomedical entity extraction in Spanish clinical narratives. We explore the use of contextualized word embeddings, which incorporate context variation into word representations, to enhance named entity recognition in Spanish language clinical text, particularly of pharmacological substances, compounds, and proteins. Various combinations of word and sense embeddings were tested on the evaluation corpus of the PharmacoNER 2019 task, the Spanish Clinical Case Corpus (SPACCC). This data set consists of clinical case sections extracted from open access Spanish-language medical publications. Our study shows that our deep-learning-based system with domain-specific contextualized embeddings coupled with stacking of complementary embeddings yields superior performance over a system with integrated standard and general-domain word embeddings. With this system, we achieve performance competitive with the state-of-the-art.
- Published
- 2020
50. From Evaluating to Forecasting Performance: How to Turn Information Retrieval, Natural Language Processing and Recommender Systems into Predictive Sciences (Dagstuhl Perspectives Workshop 17442)
- Author
-
Ferro N., Fuhr N., Grefenstette G., Konstan J. A., Castells P., Daly E. M., Declerck T., Ekstrand M. D., Geyer W., Gonzalo J., Kuflik T., Lind'En K., Magnini B., Nie J. Y., Perego R., Shapira B., Soboroff I., Tintarev N., Verspoor K., Willemsen M. C., Zobel J., and Human Technology Interaction
- Subjects
000 Computer science, knowledge, general works ,05 social sciences ,Formal models ,User interaction ,02 engineering and technology ,020204 information systems ,Computer Science ,0202 electrical engineering, electronic engineering, information engineering ,0509 other social sciences ,Evaluation ,050904 information & library sciences ,Simulation ,Information Systems - Abstract
We describe the state-of-the-art in performance modeling and prediction for Information Retrieval (IR), Natural Language Processing (NLP) and Recommender Systems (RecSys) along with its shortcomings and strengths. We present a framework for further research, identifying five major problem areas: understanding measures, performance analysis, making underlying assumptions explicit, identifying application features determining performance, and the development of predic- tion models describing the relationship between assumptions, features and resulting performance
- Published
- 2019
Catalog
Discovery Service for Jio Institute Digital Library
For full access to our library's resources, please sign in.