10,305 results on '"NATURAL language processing"'
Search Results
2. Shortcut Learning of Large Language Models in Natural Language Understanding.
- Author
-
MENGNAN DU, FENGXIANG HE, NA ZOU, DACHENG TAO, and XIA HU
- Subjects
- *
LANGUAGE models , *NATURAL language processing , *ARTIFICIAL intelligence , *MACHINE learning , *ALGORITHMS , *INDUCTION (Logic) - Abstract
The article looks at the use of large language models to carry out natural language understanding (NLU) tasks. It suggests that the shortcut learning common to existing large language models based on machine learning limits how robust their performance can be because they are overly dependent on spurious correlations and incidental relationships. It discusses possible approaches to overcoming this problem in the future development of large language models.
- Published
- 2024
- Full Text
- View/download PDF
3. Show It or Tell It? Text, Visualization, and Their Combination: When communicating information, language should be considered as co-equal with visualization.
- Author
-
HEARST, MARTI A.
- Subjects
- *
DATA visualization , *LANGUAGE & languages , *COMMUNICATION in information science , *LITERACY , *NATURAL language processing , *USER interfaces - Abstract
This article emphasizes the corresponding role language should have, along with visualization, in the communication of information. Topics include the combination and balance of text and visualization, an investigation of text without visualization, necessary improvements to cognitive models, and how natural language processing can impact information communication.
- Published
- 2023
- Full Text
- View/download PDF
4. Comparison of novel natural language processing algorithm with artificial neural networks in personal assistance to decrease time consumption.
- Author
-
Dhanushravi, R. and Logu, K.
- Subjects
- *
NATURAL language processing , *ARTIFICIAL neural networks , *LARGE-scale brain networks , *ALGORITHMS - Abstract
This work is a relative investigation of novel normal language and counterfeit brain networks calculations for improving voice discovery to lessen the time utilization of time individual collaborators. Materials and Methods: Novel Natural Language Processing (NNLP) calculation and Artificial Neural Networks (ANNs) calculation strategies are reenacted by differing the NNLP boundary and mechanize voice location to upgrade the pH. Test size is determined utilizing G power 80% for two gatherings and there are 40 examples utilized in this work. Results: Based on acquired results NNLP has fundamentally diminished time utilization and the exactness has been worked on around 89.00% contrasted with ANNs with precision of 71.90%. Measurable importance contrast among NNLP and ANNs was viewed as 0.376 (p>0.05). End: NNLP calculations give improved brings about voice acknowledgment than ANNs calculations. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
5. A malicious news detection on social networks using natural language processing technique in comparison with deep learning algorithm with improved F1 score.
- Author
-
Basha, Shaik Jabeer and Logu, K.
- Subjects
- *
MACHINE learning , *DEEP learning , *SOCIAL networks , *SUPPORT vector machines , *NATURAL language processing , *FAKE news - Abstract
False news is characterized as a made-up story with a goal to bamboozle or to delude. In this paper we present the answer for the errand of fake news discovery by using Deep Learning structures. Fake news identification is carried out utilizing two AI calculations, Natural Language Processing Algorithm(N=10) and Deep Learning algorithm(N=10) calculations. False and True these two sorts of dataset is utilized for Fake news recognition, and it is gathered from kaggle.com. Dataset comprises lines and 6 fundamental boundaries that are connected with the False news that information gathered from twitter. For each gathering more than 30 examples are taken, and it is separated into preparing and testing. Accuracy for Natural Language handling calculation is 91.300% and for Support Vector Machine calculation is 77.500%. There exists an insightful critical distinction between Natural Language Processing Technique and Support Vector Machine calculations with p<0.05. Fake news location utilizing Natural Language Processing calculation seems to acquire higher precision than the Support Vector Machine calculation. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
6. Comparison of natural language processing algorithm with support vector machine for fake news identification to improve peak signal to noise ratio with classified accuracy.
- Author
-
Basha, Shaik Jabeer and Logu, K.
- Subjects
- *
SUPPORT vector machines , *SIGNAL-to-noise ratio , *NATURAL language processing , *FAKE news , *ALGORITHMS - Abstract
The primary objective is to carry out the identification of phony news discovery around the web-based entertainment with the proposed Natural Language Processing contrasted and Support Vector machine Algorithm. Counterfeit news discovery is executed utilizing two AI calculations, Natural Language Processing Algorithm(N=10) and Support Vector Machine(N=10) calculations. Phony and True these two kinds of dataset is utilized for Fake news discovery, and it is gathered from kaggle.com. Dataset comprises of columns and 6 principal boundaries that are connected with the phony news that information gathered from twitter. For each gathering 20 examples are taken, and it is partitioned into preparing and testing. Exactness for Natural Language handling calculation is 91.300% and for Support Vector Machine calculation is 72.700%. There exists a logical huge distinction between Natural Language Processing Technique and Support Vector Machine calculations with p<0.05 Fake news recognition utilizing Natural Language Processing calculation seems to acquire higher precision than the Support Vector Machine calculation. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
7. Implementation of sustainable development goals through literaku application based on Google cloud APIs to improve literacy for blind people.
- Author
-
Imam, Khairul, Amalia, Amalia, Nasution, Fitri Aulia Fadillah, Martin, Eric, Ghozali, Muhammad, and Siagian, Farhan Doli Fadhiil
- Subjects
- *
BRAILLE , *CLOUD computing , *SUSTAINABLE development , *NATURAL language processing , *AGILE software development , *SCRUM (Computer software development) - Abstract
Quality education emerges on Sustainable Development Goals or SDGs in point 4 that ensure everyone receives education inclusively and equitably. Blind refers to a condition in which the function of the sense of sight is impaired to varying degrees, ranging from mild to severe to total blindness. The main problem experienced by the visually impaired in accessing literacy is highly limited due to the high cost of producing braille books, the inability of blind individuals to read braille books, and the limited availability of alternative sources, such as audiobooks. Literaku is an Android-based application that allows blind people to independently improve their literacy through the implementation of Google Cloud APIs, which serve as a tool for running applications and have a role in receiving, processing, and executing voice commands from the end user. The Literaku application aims to optimize the use of Indonesian voice commands by understanding the meaning of the nearest word with the support of Natural Language Processing technology to aid the visually impaired in locating readings and performing all application-related tasks by commanding and listening. The method applied the Agile Software Development Life Cycle with the SCRUM framework, which was conducted in phases and iterations. The Literaku application was evaluated by conducting usability testing and surveying users' satisfaction scores. The usability test was performed twice with five blind junior high school students at SLB-A YAPENTRA Tanjung Morawa District to obtain accurate user experience feedback and ensure that the program runs as intended. As a result, the final usability testing of Literaku application reached a success rate of 100%, and the level of participant satisfaction reached 89.60%, representing that the Literaku application was accepted by users very satisfactorily. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
8. Using chatbot for teaching arabic language syntax.
- Author
-
Hussien, Nadia Mahmood, Mohialden, Yasmin Makki, Hussien, Kawakib Mahmood, and Joshi, Kapil
- Subjects
- *
CHATBOTS , *NATURAL language processing , *ARABIC language , *MACHINE learning , *KNOWLEDGE base , *SYNTAX (Grammar) - Abstract
Chatbots are being used in a wide variety of industries, ranging from industry to education. It is not as effective when using traditional ways of developing a chatbot system as it is when using machine learning (ML). Historically, they were created using finite-state machines, rule-based systems, and knowledge bases. Although these technologies had shortcomings, they were nonetheless employed to create chatbots. This is because natural language processing and neural network technology have simplified the task of conversational AI systems categorizing intentions and locating persons and places. Many people have asked us how we created an Arabic chatbot that understands real language, and we'd like to demonstrate how we achieved it. It is capable of responding, acting on behalf of the user, and retaining the context of a communication between two persons. We employed models such as FastText and BERT, which may be used in multiple languages concurrently. Additionally, we employed two pipeline components that we created specifically for this project. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
9. Voice in the Machine: Ethical Considerations for Language-Capable Robots: Parsing the promise of language-capable robots.
- Author
-
Williams, T., Matuszek, Cynthia, Jokinen, Kristiina, Korpan, Raj, Pustejovsky, James, and Scassellati, Brian
- Subjects
- *
ARTIFICIAL intelligence & ethics , *ETHICS , *COMPUTATIONAL linguistics , *NATURAL language processing , *DISCRIMINATION (Sociology) , *PREJUDICES - Abstract
The article discusses various ethical considerations for language-capable robots. These concerns include trust, influence, identity, and privacy, and will require consideration by researchers, practitioners, and the general public. Various potential negative outcomes are discussed including robot control over human morals, a default identity perception grounded in white heteropatriarchy, gendered and racialized language-capable robots, and the potential for robots to be used as mobile surveillance tools.
- Published
- 2023
- Full Text
- View/download PDF
10. A Computational Inflection for Scientific Discovery.
- Author
-
HOPE, TOM, DOWNEY, DOUG, ETZIONI, OREN, WELD, DANIEL S., and HORVITZ, ERIC
- Subjects
- *
SCIENTIFIC knowledge , *LANGUAGE models , *SCIENTIFIC method , *ARTIFICIAL intelligence , *INFORMATION retrieval , *NATURAL language processing , *COGNITION , *HUMAN-artificial intelligence interaction - Abstract
This article presents an overview on task-guided scientific knowledge retrieval as a way for researchers to overcome the limitations of human cognitive capacity that in the age of explosive digital information creates a cognitive bottleneck. Topics include prototypes of task-guided scientific knowledge retrieval, as well as a look at novel representations, tools, and services and a review of systems that aid researchers in all aspects of scientific inquiry and discovery.
- Published
- 2023
- Full Text
- View/download PDF
11. Analogous Forecasting for Predicting Sport Innovation Diffusion: From Business Analytics to Natural Language Processing.
- Author
-
Wanless, Liz and Naraine, Michael L.
- Subjects
- *
NATURAL language processing , *SPORTS forecasting , *DIFFUSION of innovations , *BUSINESS analytics , *DIFFUSION of innovations theory , *HOCKEY players , *FUTUROLOGISTS - Abstract
The purpose of this study was to analyze the diffusion of one sport innovation to forecast a second. Contextualized within the diffusion of innovations theory, this study investigated cumulative business analytics diffusion as an analog for cumulative natural language processing (NLP) diffusion in professional sport. A total of 89 teams of the 123 teams in the Big Four North American men's professional sport leagues contributed: 21 from the National Football League, 23 from the National Basketball Association, 22 from Major League Baseball, and 23 from the National Hockey League. Utilizing an analogous forecasting approach, a discrete derivation of the Bass model was applied to cumulative BA adoption data. Parameters were then extended to predict cumulative NLP adoption. Resulting BA-estimated parameters (p =.0072, q =.3644) determined a close fit to NLP diffusion (root mean square error of approximation = 3.51, mean absolute error = 2.98), thereby validating BA to predict the takeoff and full adoption of NLP. This study illuminates an ongoing and isomorphic process for diffusion of innovations in the professional sport social system and generates a novel application of diffusion of innovations theory to the sport industry. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
12. Research on factors influencing the consumer repurchase intention: Data mining of consumers' online reviews based on machine learning.
- Author
-
Zhang, Jianming, Zheng, Hao, Liu, Jie, and Shen, Wei
- Subjects
- *
CONSUMER behavior , *CONSUMERS' reviews , *DATA mining , *THEORY of reasoned action , *QUALITY of service , *NATURAL language processing - Abstract
The fierce competition in the market makes it necessary for enterprises to not only consider how to increase consumers' purchase intention but also study to maintain high customer loyalty for continuous purchases. Taking the smartphone brands on the Jingdong platform (hereafter referred to as JD) as an example, the study collected 60,000 review data and using NLP technology for data mining, factors that may affect consumers' willingness to repurchase were extracted. Based on Theory of Reasoned Action (TRA), the questionnaire was made for empirical research. The results showed that the four factors, product attributes, service quality, brand image and price significantly affect consumers' repurchase intention, while service quality had the strongest effect among them, implications of the research are discussed. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
13. A data-driven conceptual framework for understanding the nature of hazards in railway accidents.
- Author
-
Hong, Wei-Ting, Clifton, Geoffrey, and Nelson, John D.
- Subjects
- *
RAILROAD accidents , *AIRCRAFT accidents , *TRANSPORTATION safety measures , *NATURAL language processing , *HAZARD mitigation , *HAZARDS - Abstract
Hazards threaten railway safety by their potential to trigger railway accidents, resulting in significant costs and impacting the public's willingness to use railways. Whilst many prior works investigate railway hazards, few offer a holistic view of hazards across jurisdictions and time because the large number of primary sources make synthesising such learnings time consuming and potentially incomplete. The conceptual framework HazardMap is developed to overcome this gap, employing open-sourced Natural Language Processing topic modelling for the automated analysis of textual data from Rail Accident Investigation Branch (RAIB), Australian Transport Safety Bureau (ATSB), National Transportation Safety Board (NTSB) and Transportation Safety Board of Canada (TSB) railway accident reports. The topic modelling depicts the relationships between hazards, railway accidents and investigator recommendations and is further extended and integrated with the existing risk theory and epidemiological accident models. The results allow the different aspects of each hazard to be listed along with the potential combinations of hazards that could trigger railway accidents. Better understanding of the aspects of individual hazards and the relationships between hazards and previous accidents can inform more effective hazard mitigation policies including technical or regulatory interventions. A case study of the risk at level crossings is provided to illustrate how HazardMap works with real-world data. This demonstrates a high degree of coverage within the existing risk management system, indicating the capability to better inform policymaking for managing risks. The primary contributions of the framework proposed are to enable a large amount of knowledge accumulated to be summarised for an intuitive policymaking process, and to allow other railway investigators to leverage lessons learnt across jurisdictions and time with limited human intervention. Future research could apply the technique to road, aviation or maritime accidents. • A framework HazardMap is developed for mapping hazards in the railway system. • A case study of the risk at level crossing is implemented. • Opportunities for practitioners in learning across jurisdiction and time. • Enabling knowledge accumulated to be summarised for policymaking process. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
14. Impacts of regional socioeconomic statuses and global events on solid waste research reflected in six waste-focused journals.
- Author
-
Zhang, Zhibo, Wang, Jingyi, Li, Jiuwei, Wang, Yao, Yin, Ke, and Fei, Xunchang
- Subjects
- *
SOLID waste , *TECHNOLOGY assessment , *SOCIOECONOMIC status , *TECHNOLOGICAL innovations , *SOCIOECONOMIC factors - Abstract
[Display omitted] • An innovative framework for categorizing solid waste research (SWR) publications. • SWR trends based on 17,629 publications using Source Latent Dirichlet Allocation. • Correlated SWR trends and major global events since 1990. • Identified influencing socio-economic factors on SWR trends. The research pertaining to solid waste is undergoing extensive advancement, thereby necessitating a consolidation and analysis of its research trajectories. The existing biblio-studies on solid waste research (SWR) lack thorough analyses of the factors influencing its trends. This article presents an innovative categorization framework that categorizes publications from six SWR journals utilizing Source Latent Dirichlet Allocation. First analyse changes in publication numbers across main categories, subcategories, journals, and regions, providing a macro-level study of SWR. Temporal analysis of keywords supplements a micro-level study of SWR, which highlights that emerging technologies with low Technology Readiness Level receive significant attention, while studies on widespread technologies are diminishing. Additionally, this study demonstrates the substantial influence of socioeconomic factors and previous SWR publications on current and future SWR trends. Finally, the article confirms the impact of global events on SWR trends by examining the structural breakpoints of SWR and their correlation with global events. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
15. Measuring Implicit Bias in ICU Notes Using Word-Embedding Neural Network Models.
- Author
-
Cobert, Julien, Mills, Hunter, Lee, Albert, Gologorskaya, Oksana, Espejo, Edie, Jeon, Sun Young, Boscardin, W. John, Heintz, Timothy A., Kennedy, Christopher J., Ashana, Deepshikha C., Chapman, Allyson Cook, Raghunathan, Karthik, Smith, Alex K., and Lee, Sei J.
- Abstract
Language in nonmedical data sets is known to transmit human-like biases when used in natural language processing (NLP) algorithms that can reinforce disparities. It is unclear if NLP algorithms of medical notes could lead to similar transmissions of biases. Can we identify implicit bias in clinical notes, and are biases stable across time and geography? To determine whether different racial and ethnic descriptors are similar contextually to stigmatizing language in ICU notes and whether these relationships are stable across time and geography, we identified notes on critically ill adults admitted to the University of California, San Francisco (UCSF), from 2012 through 2022 and to Beth Israel Deaconess Hospital (BIDMC) from 2001 through 2012. Because word meaning is derived largely from context, we trained unsupervised word-embedding algorithms to measure the similarity (cosine similarity) quantitatively of the context between a racial or ethnic descriptor (eg, African-American) and a stigmatizing target word (eg, nonco - operative) or group of words (violence , passivity , noncompliance , nonadherence). In UCSF notes, Black descriptors were less likely to be similar contextually to violent words compared with White descriptors. Contrastingly, in BIDMC notes, Black descriptors were more likely to be similar contextually to violent words compared with White descriptors. The UCSF data set also showed that Black descriptors were more similar contextually to passivity and noncompliance words compared with Latinx descriptors. Implicit bias is identifiable in ICU notes. Racial and ethnic group descriptors carry different contextual relationships to stigmatizing words, depending on when and where notes were written. Because NLP models seem able to transmit implicit bias from training data, use of NLP algorithms in clinical prediction could reinforce disparities. Active debiasing strategies may be necessary to achieve algorithmic fairness when using language models in clinical research. [Display omitted] [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
16. Incorporation of "Artificial Intelligence" for Objective Pain Assessment: A Comprehensive Review.
- Author
-
El-Tallawy, Salah N., Pergolizzi, Joseph V., Vasiliu-Feltes, Ingrid, Ahmed, Rania S., LeQuang, JoAnn K., El-Tallawy, Hamdy N., Varrassi, Giustino, and Nagiub, Mohamed S.
- Subjects
- *
PAIN measurement , *ARTIFICIAL intelligence , *NATURAL language processing , *EVIDENCE gaps , *CHILD patients - Abstract
Pain is a significant health issue, and pain assessment is essential for proper diagnosis, follow-up, and effective management of pain. The conventional methods of pain assessment often suffer from subjectivity and variability. The main issue is to understand better how people experience pain. In recent years, artificial intelligence (AI) has been playing a growing role in improving clinical diagnosis and decision-making. The application of AI offers promising opportunities to improve the accuracy and efficiency of pain assessment. This review article provides an overview of the current state of AI in pain assessment and explores its potential for improving accuracy, efficiency, and personalized care. By examining the existing literature, research gaps, and future directions, this article aims to guide further advancements in the field of pain management. An online database search was conducted via multiple websites to identify the relevant articles. The inclusion criteria were English articles published between January 2014 and January 2024). Articles that were available as full text clinical trials, observational studies, review articles, systemic reviews, and meta-analyses were included in this review. The exclusion criteria were articles that were not in the English language, not available as free full text, those involving pediatric patients, case reports, and editorials. A total of (47) articles were included in this review. In conclusion, the application of AI in pain management could present promising solutions for pain assessment. AI can potentially increase the accuracy, precision, and efficiency of objective pain assessment. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
17. Artificial Intelligence Augmented Qualitative Analysis: The Way of the Future?
- Author
-
Hitch, Danielle
- Subjects
- *
LANGUAGE & languages , *DOCUMENTATION , *QUALITATIVE research , *DATA analysis , *ARTIFICIAL intelligence , *POST-acute COVID-19 syndrome , *NATURAL language processing , *THEMATIC analysis , *MACHINE learning , *RESEARCH ethics - Abstract
The artificial intelligence (AI) revolution is here and gathering momentum, thanks to new models of natural language processing (NLP) and rapidly increasing adoption by the public. NLP technology uses statistical analysis of language structures to analyse and generate human language, using text or speech as its source material. It can also be applied to visual mediums like images and videos. A few qualitative research early adopters are beginning to adopt this technology into their work, but our understanding of its potential remains in its infancy. This article will define and describe NLP-based AI and discuss its benefits and limitations for reflexive thematic analysis in health research. While there are many platforms available, ChatGPT is the most well-known and accessible. A worked example using ChatGPT to augment reflexive thematic analysis is provided to illustrate potential application in practice. This article is intended to inspire further conversation around the role of AI in qualitative research and offer practical guidance for researchers seeking to adopt this technology. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
18. Diagnostic Workup and Therapeutic Intervention of Hiatal Hernias Discovered as Incidental Findings on Computed Tomography.
- Author
-
Lou, Johanna, Kooragayala, Keshav, Williams, Jennifer, Kalola, Ami, Crudeli, Connor, Sandilos, Georgianna, Butchy, Margaret V., Shersher, David D., and Burg, Jennifer M.
- Subjects
- *
HIATAL hernia , *COMPUTED tomography , *NATURAL language processing , *FLUOROSCOPY , *HERNIA surgery , *HERNIA - Abstract
Background: Computed tomography imaging routinely detects incidental findings; most research focuses on malignant findings. However, benign diseases such as hiatal hernia also require identification and follow-up. Natural language algorithms can help identify these non-malignant findings. Methods: Imaging of adult trauma patients from 2010 to 2020 who underwent CT chest/abdomen/pelvis was evaluated using an open-source natural language processor to query for hiatal hernias. Patients who underwent subsequent imaging, endoscopy, fluoroscopy, or operation were retrospectively reviewed. Results: 1087(10.6%) of 10 299 patients had incidental hiatal hernias: 812 small (74.7%) and 275 moderate/large (25.3%). 224 (20.7%) had subsequent imaging or endoscopic evaluation. Compared to those with small hernias, patients with moderate/large hernias were older (66.3 ± 19.4 vs 79.6 ± 12.6 years, P <.001) and predominantly female (403[49.6%] vs 199[72.4%], P <.001). Moderate/large hernias were not more likely to grow (small vs moderate/large: 13[7.6%] vs 8[15.1%], P =.102). Patients with moderate/large hernias were more likely to have an intervention or referral (small vs moderate/large: 6[3.5%] vs 7[13.2%], P =.008). No patients underwent elective or emergent hernia repair. Three patients had surgical referral; however, only one was seen by a surgeon. One patient death was associated with a large hiatal hernia. Conclusions: We demonstrate a novel utilization of natural language processing to identify patients with incidental hiatal hernia in a large population, and found a 10.6% incidence with only 1.2%. (13/1087) of these receiving a referral for follow-up. While most incidental hiatal hernias are small, moderate/large and symptomatic hernias have high risk of loss-to-follow-up and need referral pipelines to improve patient outcomes. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
19. Beyond Language Barriers: Allowing Multiple Languages in Postsecondary Chemistry Classes Through Multilingual Machine Learning.
- Author
-
Martin, Paul P. and Graulich, Nicole
- Subjects
- *
MACHINE learning , *COMMUNICATION barriers , *MACHINE translating , *NATURAL language processing , *CHEMISTRY students , *FORMATIVE evaluation - Abstract
Students who learn the language of instruction as an additional language represent a heterogeneous group with varying linguistic and cultural backgrounds, contributing to classroom diversity. Because of the manifold challenges these students encounter while learning the language of instruction, additional barriers arise for them when engaging in chemistry classes. Adapting teaching practices to the language skills of these students, for instance, in formative assessments, is essential to promote equity and inclusivity in chemistry learning. For this reason, novel educational practices are needed to meet each student's unique set of language capabilities, irrespective of course size. In this study, we propose and validate several approaches to allow undergraduate chemistry students who are not yet fluent in the language of instruction to complete a formative assessment in their preferred language. A technically easy-to-implement option for instructors is to use translation tools to translate students' reasoning in any language into the instructor's language. Besides, instructors could also establish multilingual machine learning models capable of automatically analyzing students' reasoning regardless of the applied language. Herein, we evaluated both opportunities by comparing the reliability of three translation tools and determining the degree to which multilingual machine learning models can simultaneously assess written arguments in different languages. The findings illustrate opportunities to apply machine learning for analyzing students' reasoning in multiple languages, demonstrating the potential of such techniques in ensuring equal access for learners of the language of instruction. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
20. An investigation of teachers' perceptions of using ChatGPT as a supporting tool for teaching and learning in the digital era.
- Author
-
ElSayary, Areej
- Subjects
- *
DIGITAL technology , *HIGH schools , *QUALITATIVE research , *ARTIFICIAL intelligence , *EDUCATIONAL outcomes , *INTERVIEWING , *NATURAL language processing , *TEACHING methods , *TEACHING , *LEARNING , *DESCRIPTIVE statistics , *QUANTITATIVE research , *CHI-squared test , *SURVEYS , *COLLEGE teacher attitudes , *RESEARCH methodology , *LEARNING strategies , *DATA analysis software , *MIDDLE schools - Abstract
Background: The widespread use of information and communication technology (ICT) has led to significant changes in societal aspects, resulting in the emergence of a "knowledge society." However, students and teachers have faced challenges in adapting to this digitalization. In the United Arab Emirates (UAE), transitioning to a knowledge‐based economy is a primary national agenda goal, aligning with Sustainable Development Goal 4 (SDG4) of ensuring high‐quality education. Objectives: This research investigates teachers' perceptions of using ChatGPT as a digital supporting tool for teaching and learning practices. This includes lesson planning, teaching and learning activities, assessment and feedback and the challenges and benefits explored. Methods: This study employs an explanatory sequential mixed‐methods design involving quantitative and qualitative data collection methods. An online survey was used with closed‐ended items to collect quantitative data, while semi‐structured interviews were conducted to collect qualitative data. The study participants are middle and high school teachers (n1 = 40) from different Dubai and Abu Dhabi private schools. Results and Conclusions: The most noticeable result is that teachers feel the benefits of using ChatGPT in lesson planning, teaching and learning and less in assessment and feedback. Some challenges and benefits were highlighted in each area and recommendations were suggested. However, teachers' biggest challenge was the bias and accuracy of information received and the lack of human interaction. Takeaways: The findings provide valuable insights into the potential of ChatGPT in education and inform future research in this area. Specifically, the study provided insights into the effectiveness of ChatGPT in enhancing students' learning outcomes, engagement and motivation, as well as its impact on teaching practices and paedagogical beliefs. Lay Description: ChatGPT can be used as a digital supporting tool for teaching and learningChatGPT positively impacts student learning outcomes and performancesChatGPT helped teachers generate innovative ideas for differentiated activities [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
21. Will sentiment analysis need subculture? A new data augmentation approach.
- Author
-
Wang, Zhenhua, He, Simin, Xu, Guang, and Ren, Ming
- Subjects
- *
RESEARCH funding , *CULTURE , *EMOTIONS , *NATURAL language processing , *EMOTICONS & emojis , *DATA encryption , *EXPERIMENTAL design , *LINGUISTICS , *SENTIMENT analysis , *AFFECT (Psychology) - Abstract
Nowadays, the omnipresence of the Internet has fostered a subculture that congregates around the contemporary milieu. The subculture artfully articulates the intricacies of human feelings by ardently pursuing the allure of novelty, a fact that cannot be disregarded in the sentiment analysis. This paper aims to enrich data through the lens of subculture, to address the insufficient training data faced by sentiment analysis. To this end, a new approach of subculture‐based data augmentation (SCDA) is proposed, which engenders enhanced texts for each training text by leveraging the creation of specific subcultural expression generators. The extensive experiments attest to the effectiveness and potential of SCDA. The results also shed light on the phenomenon that disparate subcultural expressions elicit varying degrees of sentiment stimulation. Moreover, an intriguing conjecture arises, suggesting the linear reversibility of certain subcultural expressions. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
22. A paradigm shift?—On the ethics of medical large language models.
- Author
-
Grote, Thomas and Berens, Philipp
- Subjects
- *
PATIENT autonomy , *DEBATE , *INTERPROFESSIONAL relations , *MEDICAL care , *RESPONSIBILITY , *PRIVACY , *BIOETHICS , *NATURAL language processing , *PARADIGMS (Social sciences) , *ATTITUDES of medical personnel , *TRUST , *MACHINE learning , *HONESTY , *MEDICAL ethics , *DISCLOSURE - Abstract
After a wave of breakthroughs in image‐based medical diagnostics and risk prediction models, machine learning (ML) has turned into a normal science. However, prominent researchers are claiming that another paradigm shift in medical ML is imminent—due to most recent staggering successes of large language models—from single‐purpose applications toward generalist models, driven by natural language. This article investigates the implications of this paradigm shift for the ethical debate. Focusing on issues like trust, transparency, threats of patient autonomy, responsibility issues in the collaboration of clinicians and ML models, fairness, and privacy, it will be argued that the main problems will be continuous with the current debate. However, due to functioning of large language models, the complexity of all these problems increases. In addition, the article discusses some profound challenges for the clinical evaluation of large language models and threats to the reproducibility and replicability of studies about large language models in medicine due to corporate interests. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
23. ELSTM: An improved long short‐term memory network language model for sequence learning.
- Author
-
Li, Zhi, Wang, Qing, Wang, Jia‐Qiang, Qu, Han‐Bing, Dong, Jichang, and Dong, Zhi
- Subjects
- *
LANGUAGE models , *TIME complexity , *RECURRENT neural networks , *NATURAL language processing , *COMPUTATIONAL complexity - Abstract
The gated structure of the long short‐term memory (LSTM) alleviates the defects of gradient disappearance and explosion in the recurrent neural network (RNN). It has received widespread attention in sequence learning such as text analysis. Although LSTM has good performance in handling remote dependencies, information loss often occurs in long‐distance transmission. We propose a new model called ELSTM based on the computational complexity and gradient dispersion in the traditional LSTM model. This model simplifies the input gate of LSTM, reduces some time complexity by reducing some components, and improves the output gate. By introducing the exponential linear unit activation layer, the problem of gradient dispersion is alleviated. Comparing the new model with multiple existing models, when predicting language sequences, the time used by the model has been greatly reduced, and the language confusion has been reduced, showing good performance. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
24. A predictive typological content retrieval method for real‐time applications using multilingual natural language processing.
- Author
-
Baskar, S., Dhote, Sunita, Dhote, Tejas, Jayanandini, Gopalan, Akila, Duraisamy, and Doss, Srinath
- Subjects
- *
ELECTRONIC data processing , *KNOWLEDGE base , *KNOWLEDGE transfer , *SOCIAL interaction - Abstract
Natural language processing (NLP) is widely used in multi‐media real‐time applications for understanding human interactions through computer aided‐analysis. NLP is common in auto‐filling, voice recognition, typo‐checking applications, and so forth. Multilingual NLP requires vast data processing and interaction recognition features for leveraging content retrieval precision. To strengthen this concept, a predictive typological content retrieval method is introduced in this article. The proposed method maximizes and relies on distributed transfer learning for training multilingual interactions with pitch and tone features. The phonetic pronunciation and the previous content‐based predictions are forwarded using knowledge transfer. This knowledge is modelled using the training data and precise contents identified in the previous processing instances. For this purpose, the auto‐fill and error correction data are augmented with the training and multilingual processing databases. Depending on the current prediction and previous content, the knowledge base is updated, and further training relies on this feature. Therefore, the proposed method accurately identifies the content across multilingual NLP models. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
25. Natural language processing with deep learning enabled hybrid content retrieval model for digital library management.
- Author
-
Ragab, Mahmoud, Almuhammadi, Anas, Mansour, Romany F., and Kadry, Seifedine
- Subjects
- *
DEEP learning , *NATURAL language processing , *LIBRARY administration , *DIGITAL libraries , *ARTIFICIAL intelligence , *BLENDED learning - Abstract
In recent times, natural language processing (NLP) technique has received significant attention in content retrieval (CR) domain. The emergence of digital libraries, in recent years, enables people from across the globe to access and store books, documents, and literature of multiple kinds. The development of NLP models has considerably improved the performance in terms of digital library management. In this scenario, artificial intelligence‐based expert systems are required to handle massive quantities of data that exist in digital libraries and achieve effective CR performance. In this background, the current study designs NLP with deep learning enabled hybrid content retrieval (NLPDL‐HCR) model for digital library management. The aim of the presented NLPDL‐HCR is to effectually retrieve the images as well as textual data from digital libraries based on a user's query. The proposed NLPDL‐HCR model encompasses two major stages namely, text retrieval and image retrieval (IR). During text retrieval process, the proposed NLPDL‐HCR model includes term frequency inverse document frequency vectorizer with optimal gated recurrent unit (GRU) model. The hyperparameters of the GRU model are optimally adjusted with the help of RMSProp approach. Besides, the IR process involves three sub‐processes namely, densely connected networks‐based feature extraction, butterfly optimization algorithm‐based hyperparameter tuning, and Euclidean distance‐based similarity measurement. The experimental analysis results, accomplished by the proposed NLPDL‐HCR model using benchmark datasets, highlighted its superior performance over recent state‐of‐the‐art approaches. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
26. Development of Artificial Intelligence‐based chatbot for smart aquafarm practices.
- Author
-
Jasmin, Ayesha, Ramesh, Pradeep, and Tanveer, Mohammad
- Subjects
- *
ARTIFICIAL intelligence , *CHATBOTS , *SOCIAL media , *NATURAL language processing , *APPLICATION software - Abstract
This study includes the development of a farm advisory framework known as the "AquaProFAN framework," which is concerned with the study of various requirements by the aquafarmers and stakeholder community in fisheries. AquaProFAN (Aqua Professional Farm Advisory Network) advisory committee provided careful consideration to the aquafarmers' inquiries and proper clarification with many innovative ideas, and the feasibility study on chatbot implementation was also performed. The end product, called AquaGent, is an AI‐based chatbot designed for shrimp aquafarmers and support services. It was developed and deployed in social media platforms like Facebook Messenger and Telegram for research and dissemination purposes. The effectiveness of the chatbot was analysed and the findings indicated that: (i) informational requests are more satisfactory than emotional requests, (ii) the performance of chatbot is comparatively better than other software applications concerning informational requests, (iii) participants perceive 'AquaGent' chatbot to be more user friendly and time saving. It can be concluded that the implementation of this state‐of‐art technology in aquaculture sector will improve stakeholders' understanding for future efficient and profitable production. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
27. EU judicial behaviour research: a look back and a look ahead.
- Author
-
Dyevre, Arthur
- Subjects
- *
NATURAL language processing , *LEGAL judgments - Abstract
Over the last three decades, the field of EU judicial behaviour has spawned a sizeable body of work. While the efforts of EU judicial scholars have indubitably generated important insights about the operation of the Court of Justice and its interactions with domestic tribunals and litigants, EU judicial behaviour research suffers from substantial limitations at multiple levels. The first arises from the field's poor integration with theoretical advances that have emerged in other contexts. The second pertains to its methodological assumptions, which have yet to be updated to respond to the credibility crisis. The third relates to the existing datasets and the paucity of data on national courts' practices outside the preliminary ruling mechanism. To address these shortcomings, I suggest how the field may benefit from incorporating theoretical advances from research on judges in other contexts, a stronger emphasis on smart designs and experimental and quasi-experimental methods and the deployment of data-crawling and Natural Language Processing (NLP) techniques. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
28. Characteristics and Problems of Smart City Construction above the Prefecture Level in China: An Exploratory Study.
- Author
-
Chu, Jinhua, Zhong, Anyuan, and Zhang, Wenkun
- Subjects
- *
SMART cities , *GAUSSIAN mixture models , *CITIES & towns , *URBAN policy , *MUNICIPAL government - Abstract
A smart city is regarded as a new form of urban development proposed by humans to cope with urban problems. Smart cities are of great interest to theoretical circles and are favored by city governments. However, in China, research on common problems and solution strategies in the process of smart city construction is still insufficient. To address this insufficiency, we utilized Python language and the Selenium library to realize a web crawler and construct a smart city corpus background library. Then, the word2vec model was used for word vectorization to divide policy texts into phrases, and 132 policies from 85 cities above the prefecture level in China were analyzed using the Gaussian mixture model (GMM). The results showed many problems in the smart city construction process in China, including the homogenization of content, the wide topic coverage, the influence of higher-level governments, the mismatch with the local situation, and the neglect of relationship governance. Moreover, an efficiency-governance model was designed, and it held certain theoretical and practical guiding significance for further promoting the healthy and sustainable development of smart city construction in China. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
29. Developing a single‐session outcome measure using natural language processing on digital mental health transcripts.
- Author
-
Milligan, Gregor, Bernard, Aynsley, Dowthwaite, Liz, Vallejos, Elvira Perez, Davis, Jamie, Salhi, Louisa, and Goulding, James
- Abstract
Background Methods Results Conclusion Current outcome measures in digital mental health lack granularity, especially for single‐session interventions. This study aimed to address this by utilising natural language processing (NLP) methods to create a clear and relevant outcome measure. This paper describes the development of the Adult Session Wants and Needs Outcome Measure (Adult SWAN‐OM), a novel outcome measure for the Qwell digital mental healthcare platform to understand service user (SU) needs engaging in single‐session therapy (SST).The research employs a multi‐phased approach combining NLP methods with the typical stages of outcome measures development as follows: (1) assumption definition and validation with SUs and clinicians; (2) transcript theme extraction using the RoBERTa large language model (LLM) in conjunction with topic modelling to extract themes from 254 single‐session transcripts from 192 SUs; (3) clinical item refinement focus group; (4) content validity with clinicians and SUs to improve the relevance and clarity of the items; and (5) outcome measure finalisation in a workshop held with clinicians to consolidate the final wording.Ninety‐six potential wants and needs were generated and distilled into 12 measure items. The outcome measure was shown to be relevant and clear to both SUs and clinicians when used in the context of SST.This study highlights the potential of combining NLP approaches with co‐creation methods in single‐session outcome measure development. We argue that the incorporation of clinical expertise and SU experience ensures the clarity and applicability of such measures and that this approach to capturing single‐session wants and needs promises novel insights for supporting digital mental health interventions. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
30. Temporal and Semantic Fusion for Multi-Label Crime Classification via a TCN-BERT-Coupled Approach.
- Author
-
Wang, Pei, Chen, Teng, and Wang, Yuewei
- Subjects
- *
ARTIFICIAL intelligence , *CRIME , *CLASSIFICATION , *TRANSFORMER models , *NATURAL language processing - Abstract
Artificial Intelligence (AI) techniques leverage the justice system in terms of effectiveness and efficiency. AI-empowered multi-label crime classification can facilitate the precise and expedient categorization of various legal documents. Multi-label classification within the justice system has an indispensable role in achieving accurate legal categorization that invigorates case analysis, optimizes resource distribution and refines the contours of legal processes. To support this critical function, this paper proposes a temporal convolutional network (TCN)-bidirectional encoder representations from transformers (BERT)-coupled model for multi-label crime classification. The proposed method fuses the temporal formation and semantic information in the model to obtain a high-quality result. The experimental results show that the proposed method achieved the best accuracy in comparison to existing methods on a public dataset. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
31. A Pre-Silicon Detection Based on Deep Learning Model for Hardware Trojans.
- Author
-
Ma, Pengcheng, Wang, Zhen, and Wang, Yong
- Subjects
- *
DEEP learning , *CONVOLUTIONAL neural networks , *NATURAL language processing , *ARTIFICIAL neural networks - Abstract
Several hardware Trojan (HT) detection techniques are available today to ensure the security of hardware systems. However, the existing pre-silicon HT detection techniques have problems such as difficulties in capturing HT path features and poor applicability. To address these challenges, this paper proposes a gate-level HT detection scheme based on a deep learning model. We parse the circuit gate-level netlist and develop an algorithm to extract circuit path sentences based on the signal propagation rule. Path sentences consisting of gate names are extracted as experimental datasets. We apply the theory of natural language processing (NLP) to the task of HT detection and use three neural networks to filter the length of path sentences. Then, based on the deep learning model text convolutional neural network (TextCNN), we propose PS-TextCNN for HT detection. Our approach is verified on seven benchmark circuits of the RS232-series and eight benchmark circuits of the s-series. We achieve an average true positive rate (TPR) of 88.9%. The TPR of the RS232-series reaches a high score of 99.5%. The TPR of the s-series is 79.5%, which is significantly higher than that of the existing gate-level HT detection techniques. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
32. Application of BiLSTM-CRF model with different embeddings for product name extraction in unstructured Turkish text.
- Author
-
Arslan, Serdar
- Subjects
- *
DEEP learning , *ARTIFICIAL neural networks , *NATURAL language processing , *RANDOM fields , *EVIDENCE gaps , *DISCOURSE analysis - Abstract
Named entity recognition (NER) plays a pivotal role in Natural Language Processing by identifying and classifying entities within textual data. While NER methodologies have seen significant advancements, driven by pretrained word embeddings and deep neural networks, the majority of these studies have focused on text with well-defined grammar and structure. A significant research gap exists concerning NER in informal or unstructured text, where traditional grammar rules and sentence structure are absent. This research addresses this crucial gap by focusing on the detection of product names within unstructured Turkish text. To accomplish this, we propose a deep learning-based NER model which combines a Bidirectional Long Short-Term Memory (BiLSTM) architecture with a Conditional Random Field (CRF) layer, further enhanced by FastText embeddings. To comprehensively evaluate and compare our model's performance, we explore different embedding approaches, including Word2Vec and Glove, in conjunction with the Bidirectional Long Short-Term Memory and Conditional Random Field (BiLSTM-CRF) model. Furthermore, we conduct comparisons against BERT to assess the efficacy of our approach. Our experimentation utilizes a Turkish e-commerce dataset gathered from the internet, where traditional grammatical and structural rules may not apply. The BiLSTM-CRF model with FastText embeddings achieved an F1 score value of 57.40%, a precision value of 55.78%, and a recall value of 59.12%. These results indicate promising performance in outperforming other baseline techniques. This research contributes to the field of NER by addressing the unique challenges posed by unstructured Turkish text and opens avenues for improved entity recognition in informal language settings, with potential applications across various domains. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
33. Advances in machine learning with chemical language models in molecular property and reaction outcome predictions.
- Author
-
Das, Manajit, Ghosh, Ankit, and Sunoj, Raghavan B.
- Subjects
- *
LANGUAGE models , *NATURAL language processing , *CHEMICAL models , *CHEMICAL milling , *RECURRENT neural networks , *MACHINE learning - Abstract
Molecular properties and reactions form the foundation of chemical space. Over the years, innumerable molecules have been synthesized, a smaller fraction of them found immediate applications, while a larger proportion served as a testimony to creative and empirical nature of the domain of chemical science. With increasing emphasis on sustainable practices, it is desirable that a target set of molecules are synthesized preferably through a fewer empirical attempts instead of a larger library, to realize an active candidate. In this front, predictive endeavors using machine learning (ML) models built on available data acquire high timely significance. Prediction of molecular property and reaction outcome remain one of the burgeoning applications of ML in chemical science. Among several methods of encoding molecular samples for ML models, the ones that employ language like representations are gaining steady popularity. Such representations would additionally help adopt well‐developed natural language processing (NLP) models for chemical applications. Given this advantageous background, herein we describe several successful chemical applications of NLP focusing on molecular property and reaction outcome predictions. From relatively simpler recurrent neural networks (RNNs) to complex models like transformers, different network architecture have been leveraged for tasks such as de novo drug design, catalyst generation, forward and retro‐synthesis predictions. The chemical language model (CLM) provides promising avenues toward a broad range of applications in a time and cost‐effective manner. While we showcase an optimistic outlook of CLMs, attention is also placed on the persisting challenges in reaction domain, which would optimistically be addressed by advanced algorithms tailored to chemical language and with increased availability of high‐quality datasets. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
34. Advancing automatic text summarization: Unleashing enhanced binary multi-objective grey wolf optimization with mutation.
- Author
-
Sheikh, Muhammad Ayyaz, Bashir, Maryam, and Sudddle, Mehtab Kiran
- Subjects
- *
TEXT summarization , *GREY Wolf Optimizer algorithm , *NATURAL language processing , *WOLVES , *EVOLUTIONARY algorithms , *RESEARCH personnel - Abstract
Automatic Text Summarization (ATS) is gaining popularity as there is a growing demand for a system capable of processing extensive textual content and delivering a concise, yet meaningful, relevant, and useful summary. Manual summarization is both expensive and time-consuming, making it impractical for humans to handle vast amounts of data. Consequently, the need for ATS systems has become evident. These systems encounter challenges such as ensuring comprehensive content coverage, determining the appropriate length of the summary, addressing redundancy, and maintaining coherence in the generated summary. Researchers are actively addressing these challenges by employing Natural Language Processing (NLP) techniques. While traditional methods exist for generating summaries, they often fall short of addressing multiple aspects simultaneously. To overcome this limitation, recent advancements have introduced multi-objective evolutionary algorithms for ATS. This study proposes an enhancement to the performance of ATS through the utilization of an improved version of the Binary Multi-Objective Grey Wolf Optimizer (BMOGWO), incorporating mutation. The performance of this enhanced algorithm is assessed by comparing it with state-of-the-art algorithms using the DUC2002 dataset. Experimental results demonstrate that the proposed algorithm significantly outperforms the compared approaches. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
35. Soutcom: Real‐time sentiment analysis of Arabic text for football fan satisfaction using a bidirectional LSTM.
- Author
-
Alfarhood, Sultan
- Abstract
In the last few years, various topics, including sports, have seen social media platforms emerge as significant sources of information and viewpoints. Football fans use social media to express their opinions and sentiments about their favourite teams and players. Analysing these opinions can provide valuable information on the satisfaction of football fans with their teams. In this article, we present Soutcom, a scalable real‐time system that estimates the satisfaction of football fans with their teams. Our approach leverages the power of social media platforms to gather real‐time opinions and emotions of football fans and applies state‐of‐the‐art machine learning‐based sentiment analysis techniques to accurately predict the sentiment of Arabic posts. Soutcom is designed as a cloud‐based scalable system integrated with the X (formerly known as Twitter) API and a football data service to retrieve live posts and match data. The Arabic posts are analysed using our proposed bidirectional LSTM (biLSTM) model, which we trained on a custom dataset specifically tailored for the sports domain. Our evaluation shows that the proposed model outperforms other machine learning models such as Random Forest, XGBoost and Convolutional Neural Networks (CNNs) in terms of accuracy and F1‐score with values of 0.83 and 0.82, respectively. Furthermore, we analyse the inference time of our proposed model and suggest that there is a trade‐off between performance and efficiency when selecting a model for sentiment analysis on Arabic posts. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
36. Generative design of compounds with desired potency from target protein sequences using a multimodal biochemical language model.
- Author
-
Chen, Hengwei and Bajorath, Jürgen
- Subjects
- *
LANGUAGE models , *DEEP learning , *BIOCHEMICAL models , *AMINO acid sequence , *NATURAL language processing , *MACHINE translating , *NEUROLINGUISTICS - Abstract
Deep learning models adapted from natural language processing offer new opportunities for the prediction of active compounds via machine translation of sequential molecular data representations. For example, chemical language models are often derived for compound string transformation. Moreover, given the principal versatility of language models for translating different types of textual representations, off-the-beaten-path design tasks might be explored. In this work, we have investigated generative design of active compounds with desired potency from target sequence embeddings, representing a rather provoking prediction task. Therefore, a dual-component conditional language model was designed for learning from multimodal data. It comprised a protein language model component for generating target sequence embeddings and a conditional transformer for predicting new active compounds with desired potency. To this end, the designated "biochemical" language model was trained to learn mappings of combined protein sequence and compound potency value embeddings to corresponding compounds, fine-tuned on individual activity classes not encountered during model derivation, and evaluated on compound test sets that were structurally distinct from training sets. The biochemical language model correctly reproduced known compounds with different potency for all activity classes, providing proof-of-concept for the approach. Furthermore, the conditional model consistently reproduced larger numbers of known compounds as well as more potent compounds than an unconditional model, revealing a substantial effect of potency conditioning. The biochemical language model also generated structurally diverse candidate compounds departing from both fine-tuning and test compounds. Overall, generative compound design based on potency value-conditioned target sequence embeddings yielded promising results, rendering the approach attractive for further exploration and practical applications. Scientific contribution: The approach introduced herein combines protein language model and chemical language model components, representing an advanced architecture, and is the first methodology for predicting compounds with desired potency from conditioned protein sequence data. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
37. Analysis of addiction craving onset through natural language processing of the online forum Reddit.
- Author
-
Kramer, Thea, Groh, Georg, Stüben, Nathalie, and Soyka, Michael
- Subjects
- *
NATURAL language processing , *INTERNET forums , *ALCOHOLISM , *ALCOHOLIC beverages , *DESIRE , *VIRTUAL communities , *ARTIFICIAL intelligence - Abstract
Aims: Alcohol cravings are considered a major factor in relapse among individuals with alcohol use disorder (AUD). This study aims to investigate the frequency and triggers of cravings in the daily lives of people with alcohol-related issues. Large amounts of data are analyzed with Artificial Intelligence (AI) methods to identify possible groupings and patterns. Methods: For the analysis, posts from the online forum "stopdrinking" on the Reddit platform were used as the dataset from April 2017 to April 2022. The posts were filtered for craving content and processed using the word2vec method to map them into a multi-dimensional vector space. Statistical analyses were conducted to calculate the nature and frequency of craving contexts and triggers (location, time, social environment, and emotions) using word similarity scores. Additionally, the themes of the craving-related posts were semantically grouped using a Latent Dirichlet Allocation (LDA) topic model. The accuracy of the results was evaluated using two manually created test datasets. Results: Approximately 16% of the forum posts discuss cravings. The number of craving-related posts decreases exponentially with the number of days since the author's last alcoholic drink. The topic model confirms that the majority of posts involve individual factors and triggers of cravings. The context analysis aligns with previous craving trigger findings related to the social environment, locations and emotions. Strong semantic craving similarities were found for the emotions boredom, stress and the location airport. The results for each method were successfully validated on test datasets. Conclusions: This exploratory approach is the first to analyze alcohol cravings in the daily lives of over 24,000 individuals, providing a foundation for further AI-based craving analyses. The analysis confirms commonly known craving triggers and even discovers new important craving contexts. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
38. Blending Strategic Expertise and Technology: A Case Study for Practice Analysis.
- Author
-
Belwalkar, Bharati B., Schultz, Matthew, Curnow, Christina, and Setzer, J. Carl
- Abstract
There is a growing integration of technology in the workplace (World Economic Forum), and with it, organizations are increasingly relying on advanced technological approaches for improving their human capital processes to stay relevant and competitive in complex environments. All professions must keep up with this transition and begin integrating technology into their tools and processes. This paper centers on how advanced technological approaches (such as natural language processing (NLP) and data mining) have complemented a traditional practice analysis of the accounting profession. We also discuss strategic selection and use of subject‐matter experts (SMEs) for more efficient practice analysis. The authors have adopted a triangulation process—gathering information from traditional practice analysis, using selected SMEs, and confirming findings with a novel NLP‐based approach. These methods collectively contributed to the revision of the Uniform CPA Exam blueprint and in understanding accounting trends. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
39. A reduced proteomic signature in critically ill Covid-19 patients determined with plasma antibody micro-array and machine learning.
- Author
-
Patel, Maitray A., Daley, Mark, Van Nynatten, Logan R., Slessarev, Marat, Cepinskas, Gediminas, and Fraser, Douglas D.
- Subjects
- *
COVID-19 , *MACHINE learning , *CRITICALLY ill , *PROTEOMICS , *NATURAL language processing , *BLOOD coagulation factors - Abstract
Background: COVID-19 is a complex, multi-system disease with varying severity and symptoms. Identifying changes in critically ill COVID-19 patients' proteomes enables a better understanding of markers associated with susceptibility, symptoms, and treatment. We performed plasma antibody microarray and machine learning analyses to identify novel proteins of COVID-19. Methods: A case-control study comparing the concentration of 2000 plasma proteins in age- and sex-matched COVID-19 inpatients, non-COVID-19 sepsis controls, and healthy control subjects. Machine learning was used to identify a unique proteome signature in COVID-19 patients. Protein expression was correlated with clinically relevant variables and analyzed for temporal changes over hospitalization days 1, 3, 7, and 10. Expert-curated protein expression information was analyzed with Natural language processing (NLP) to determine organ- and cell-specific expression. Results: Machine learning identified a 28-protein model that accurately differentiated COVID-19 patients from ICU non-COVID-19 patients (accuracy = 0.89, AUC = 1.00, F1 = 0.89) and healthy controls (accuracy = 0.89, AUC = 1.00, F1 = 0.88). An optimal nine-protein model (PF4V1, NUCB1, CrkL, SerpinD1, Fen1, GATA-4, ProSAAS, PARK7, and NET1) maintained high classification ability. Specific proteins correlated with hemoglobin, coagulation factors, hypertension, and high-flow nasal cannula intervention (P < 0.01). Time-course analysis of the 28 leading proteins demonstrated no significant temporal changes within the COVID-19 cohort. NLP analysis identified multi-system expression of the key proteins, with the digestive and nervous systems being the leading systems. Conclusions: The plasma proteome of critically ill COVID-19 patients was distinguishable from that of non-COVID-19 sepsis controls and healthy control subjects. The leading 28 proteins and their subset of 9 proteins yielded accurate classification models and are expressed in multiple organ systems. The identified COVID-19 proteomic signature helps elucidate COVID-19 pathophysiology and may guide future COVID-19 treatment development. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
40. Identification of patients' smoking status using an explainable AI approach: a Danish electronic health records case study.
- Author
-
Ebrahimi, Ali, Henriksen, Margrethe Bang Høstgaard, Brasen, Claus Lohman, Hilberg, Ole, Hansen, Torben Frøstrup, Jensen, Lars Henrik, Peimankar, Abdolrahman, and Wiil, Uffe Kock
- Subjects
- *
ELECTRONIC health records , *MACHINE learning , *NATURAL language processing , *SMOKING , *ARTIFICIAL intelligence - Abstract
Background: Smoking is a critical risk factor responsible for over eight million annual deaths worldwide. It is essential to obtain information on smoking habits to advance research and implement preventive measures such as screening of high-risk individuals. In most countries, including Denmark, smoking habits are not systematically recorded and at best documented within unstructured free-text segments of electronic health records (EHRs). This would require researchers and clinicians to manually navigate through extensive amounts of unstructured data, which is one of the main reasons that smoking habits are rarely integrated into larger studies. Our aim is to develop machine learning models to classify patients' smoking status from their EHRs. Methods: This study proposes an efficient natural language processing (NLP) pipeline capable of classifying patients' smoking status and providing explanations for the decisions. The proposed NLP pipeline comprises four distinct components, which are; (1) considering preprocessing techniques to address abbreviations, punctuation, and other textual irregularities, (2) four cutting-edge feature extraction techniques, i.e. Embedding, BERT, Word2Vec, and Count Vectorizer, employed to extract the optimal features, (3) utilization of a Stacking-based Ensemble (SE) model and a Convolutional Long Short-Term Memory Neural Network (CNN-LSTM) for the identification of smoking status, and (4) application of a local interpretable model-agnostic explanation to explain the decisions rendered by the detection models. The EHRs of 23,132 patients with suspected lung cancer were collected from the Region of Southern Denmark during the period 1/1/2009-31/12/2018. A medical professional annotated the data into 'Smoker' and 'Non-Smoker' with further classifications as 'Active-Smoker', 'Former-Smoker', and 'Never-Smoker'. Subsequently, the annotated dataset was used for the development of binary and multiclass classification models. An extensive comparison was conducted of the detection performance across various model architectures. Results: The results of experimental validation confirm the consistency among the models. However, for binary classification, BERT method with CNN-LSTM architecture outperformed other models by achieving precision, recall, and F1-scores between 97% and 99% for both Never-Smokers and Active-Smokers. In multiclass classification, the Embedding technique with CNN-LSTM architecture yielded the most favorable results in class-specific evaluations, with equal performance measures of 97% for Never-Smoker and measures in the range of 86 to 89% for Active-Smoker and 91–92% for Never-Smoker. Conclusion: Our proposed NLP pipeline achieved a high level of classification performance. In addition, we presented the explanation of the decision made by the best performing detection model. Future work will expand the model's capabilities to analyze longer notes and a broader range of categories to maximize its utility in further research and screening applications. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
41. Automating Fault Test Cases Generation and Execution for Automotive Safety Validation via NLP and HIL Simulation.
- Author
-
Amyan, Ayman, Abboush, Mohammad, Knieke, Christoph, and Rausch, Andreas
- Subjects
- *
LANGUAGE models , *SAFETY standards , *NATURAL language processing , *FAULT location (Engineering) - Abstract
The complexity and the criticality of automotive electronic implanted systems are steadily advancing and that is especially the case for automotive software development. ISO 26262 describes requirements for the development process to confirm the safety of such complex systems. Among these requirements, fault injection is a reliable technique to assess the effectiveness of safety mechanisms and verify the correct implementation of the safety requirements. However, the method of injecting the fault in the system under test in many cases is still manual and depends on an expert, requiring a high level of knowledge of the system. In complex systems, it consumes time, is difficult to execute, and takes effort, because the testers limit the fault injection experiments and inject the minimum number of possible test cases. Fault injection enables testers to identify and address potential issues with a system under test before they become actual problems. In the automotive industry, failures can have serious hazards. In these systems, it is essential to ensure that the system can operate safely even in the presence of faults. We propose an approach using natural language processing (NLP) technologies to automatically derive the fault test cases from the functional safety requirements (FSRs) and execute them automatically by hardware-in-the-loop (HIL) in real time according to the black-box concept and the ISO 26262 standard. The approach demonstrates effectiveness in automatically identifying fault injection locations and conditions, simplifying the testing process, and providing a scalable solution for various safety-critical systems. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
42. Design of a large language model for improving customer service in telecom operators.
- Author
-
Xiaoliang, Ma, RuQiang, Zhao, Ying, Liu, Congjian, Deng, and Dequan, Du
- Abstract
Telecommunications operators are tasked with enhancing service quality, reducing operational costs, and preserving customer privacy. This study presents an innovative application of large language models (LLMs) integrated with the LangChain technology framework, aimed at revolutionizing customer service in the telecom sector. The LangChain framework features a Knowledge Organizing Module and a Knowledge Search Module, both designed to refine customer support operations. The research develops an LLM‐based approach to improve the segmentation and organization of knowledge bases, tailored for the telecommunications industry. This approach ensures seamless integration with existing LLMs while preserving distinct knowledge domains, crucial for search accuracy. Additionally, the framework includes an advanced information security protocol with a robust filtering system that effectively excludes sensitive data from the model's outputs, enhancing data security. Empirical findings indicate that the ChatGLM2‐6B+LangChain model outperforms the baseline ChatGLM2, demonstrating heightened effectiveness in telecom‐specific tasks and outstripping even more sophisticated models like GPT‐4. The implementation of this LLM‐based framework within telecom customer service systems has significantly sharpened the precision of knowledge recommendations, as reflected by a dramatic increase in acceptance rates from 15% to 70%. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
43. Mechanisms upholding the persistence of stigma across 100 years of historical text.
- Author
-
Charlesworth, Tessa E. S. and Hatzenbuehler, Mark L.
- Subjects
- *
SOCIAL stigma , *SOCIAL groups , *NATURAL language processing - Abstract
Today, many social groups face negative stereotypes. Is such negativity a stable feature of society and, if so, what mechanisms maintain stability both within and across group targets? Answering these theoretically and practically important questions requires data on dozens of group stereotypes examined simultaneously over historical and societal scales, which is only possible through recent advances in Natural Language Processing. Across two studies, we use word embeddings from millions of English-language books over 100 years (1900–2000) and extract stereotypes for 58 stigmatized groups. Study 1 examines aggregate, societal-level trends in stereotype negativity by averaging across these groups. Results reveal striking persistence in aggregate negativity (no meaningful slope), suggesting that society maintains a stable level of negative stereotypes. Study 2 introduces and tests a new framework identifying potential mechanisms upholding stereotype negativity over time. We find evidence of two key sources of this aggregate persistence: within-group "reproducibility" (e.g., stereotype negativity can be maintained by using different traits with the same underlying meaning) and across-group "replacement" (e.g., negativity from one group is transferred to other related groups). These findings provide novel historical evidence of mechanisms upholding stigmatization in society and raise new questions regarding the possibility of future stigma change. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
44. The public attitude towards ChatGPT on reddit: A study based on unsupervised learning from sentiment analysis and topic modeling.
- Author
-
Xu, Zhaoxiang, Fang, Qingguo, Huang, Yanbo, and Xie, Mingjian
- Subjects
- *
CHATGPT , *PUBLIC opinion , *SENTIMENT analysis , *NATURAL language processing , *SOCIAL media , *PUBLIC spaces , *ONLINE comments - Abstract
ChatGPT has demonstrated impressive abilities and impacted various aspects of human society since its creation, gaining widespread attention from different social spheres. This study aims to comprehensively assess public perception of ChatGPT on Reddit. The dataset was collected via Reddit, a social media platform, and includes 23,733 posts and comments related to ChatGPT. Firstly, to examine public attitudes, this study conducts content analysis utilizing topic modeling with the Latent Dirichlet Allocation (LDA) algorithm to extract pertinent topics. Furthermore, sentiment analysis categorizes user posts and comments as positive, negative, or neutral using Textblob and Vader in natural language processing. The result of topic modeling shows that seven topics regarding ChatGPT are identified, which can be grouped into three themes: user perception, technical methods, and impacts on society. Results from the sentiment analysis show that 61.6% of the posts and comments hold favorable opinions on ChatGPT. They emphasize ChatGPT's ability to prompt and engage in natural conversations with users, without relying on complex natural language processing. It provides suggestions for ChatGPT developers to enhance its usability design and functionality. Meanwhile, stakeholders, including users, should comprehend the advantages and disadvantages of ChatGPT in human society to promote ethical and regulated implementation of the system. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
45. Cities for citizens! Public value spheres for understanding conflicts in urban planning.
- Author
-
Herzog, Rico H, Gonçalves, Juliana E, Slingerland, Geertje, Kleinhans, Reinout, Prang, Holger, Brazier, Frances, and Verma, Trivik
- Subjects
- *
PUBLIC value , *PUBLIC spaces , *CITIES & towns , *URBAN planning , *PUBLIC sphere , *NATURAL language processing - Abstract
Identifying the diverse and often competing values of citizens, and resolving the consequent public value conflicts, are of significant importance for inclusive and integrated urban development. Scholars have highlighted that relational, value-laden urban space gives rise to many diverse conflicts that vary both spatially and temporally. Although notions of public value conflicts have been conceived in theory, there are few empirical studies that identify such values and their conflicts in urban space. Building on public value theory and using a case-study mixed-methods approach, this paper proposes a new approach to empirically investigate public value conflicts in urban space. Using unstructured participatory data of 4528 citizen contributions from a Public Participation Geographic Information Systems in Hamburg, Germany, natural language processing and spatial clustering techniques are used to identify areas of potential value conflicts. Four expert interviews assess and interpret these quantitative findings. By integrating quantitative assessments with the qualitative findings of the interviews, we identify 19 general public values and nine archetypical conflicts. On the basis of these results, this paper proposes a new conceptual model of 'Public Value Spheres' that extends the understanding of public value conflicts and helps to further account for the value-laden nature of urban space. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
46. Multi-strategy text data augmentation for enhanced aspect-based sentiment analysis in resource-limited scenarios.
- Author
-
Zhao, Chuanjun, Sun, Xuzhuang, and Feng, Rong
- Subjects
- *
DATA augmentation , *SENTIMENT analysis , *NATURAL language processing , *DEEP learning - Abstract
Aspect-based sentiment analysis (ABSA) constitutes a significant field within natural language processing (NLP). This study proposes a multi-strategy text data augmentation methodology to overcome challenges such as limited dataset sizes and the absence of comprehensive, high-quality corpora in aspect-level sentiment classification (ASC). Specifically, it expands the SemEval 2014 Restaurant and Laptop training datasets from 3017 to 4646 instances and from 1864 to 3693 instances, respectively. The methodology encompasses both word-level and sentence-level augmentations. Evaluations using advanced deep learning techniques, including Att-LSTM, SVM, CNN, and Bi-LSTM, were conducted on the enhanced SemEval 2014 Task 4 restaurant dataset and laptops dataset. The test dataset sizes are 920 and 668, respectively. The Att-LSTM, demonstrating superior performance, recorded F1 score improvements of 5.0% and 4.4% on the restaurant and laptop datasets, respectively, following the application of the multi-strategy augmentation methodology compared to others. This approach significantly enlarges the dataset and improves performance in ASC tasks. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
47. Hybrid Deep Learning Model Based on Sparse Recurrent Architecture.
- Author
-
Wu, Yutao and Liu, Min
- Subjects
- *
ARTIFICIAL neural networks , *NATURAL language processing , *DEEP learning , *IMAGE recognition (Computer vision) , *TRANSFORMER models , *NETWORK performance - Abstract
Deep neural network has made surprising achievements in natural language processing, image pattern classification recognition, and other domains in the last few years. It is still tough to apply to hardware-constrained or mobile equipment because of the huge number of parameters, high storage as well as computing costs. In this paper, a new sparse iteration neural network architecture is proposed. First, the pruning method is used to compress the model size and make the network sparse. Then the architecture is iterated on the sparse network model, and the network performance is improved without adding additional parameters. Finally, the hybrid deep learning model was carried out on CV tasks and NLP tasks on ANN, CNN, and Transformer. Compared with the sparse network architecture, we finally found that the accuracy of the MINST, CIFAR10, PASCAL VOC 2012, and SQuAD datasets is improved by 0.47%, 0.64%, 3.75%, and 15.06%, respectively. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
48. Evaluating the strengths and weaknesses of large language models in answering neurophysiology questions.
- Author
-
Shojaee-Mend, Hassan, Mohebbati, Reza, Amiri, Mostafa, and Atarodi, Alireza
- Subjects
- *
LANGUAGE models , *NATURAL language processing , *GEMINI (Chatbot) , *NEUROPHYSIOLOGY , *NEUROLINGUISTICS , *PROCESS capability , *CHATGPT - Abstract
Large language models (LLMs), like ChatGPT, Google's Bard, and Anthropic's Claude, showcase remarkable natural language processing capabilities. Evaluating their proficiency in specialized domains such as neurophysiology is crucial in understanding their utility in research, education, and clinical applications. This study aims to assess and compare the effectiveness of Large Language Models (LLMs) in answering neurophysiology questions in both English and Persian (Farsi) covering a range of topics and cognitive levels. Twenty questions covering four topics (general, sensory system, motor system, and integrative) and two cognitive levels (lower-order and higher-order) were posed to the LLMs. Physiologists scored the essay-style answers on a scale of 0–5 points. Statistical analysis compared the scores across different levels such as model, language, topic, and cognitive levels. Performing qualitative analysis identified reasoning gaps. In general, the models demonstrated good performance (mean score = 3.87/5), with no significant difference between language or cognitive levels. The performance was the strongest in the motor system (mean = 4.41) while the weakest was observed in integrative topics (mean = 3.35). Detailed qualitative analysis uncovered deficiencies in reasoning, discerning priorities, and knowledge integrating. This study offers valuable insights into LLMs' capabilities and limitations in the field of neurophysiology. The models demonstrate proficiency in general questions but face challenges in advanced reasoning and knowledge integration. Targeted training could address gaps in knowledge and causal reasoning. As LLMs evolve, rigorous domain-specific assessments will be crucial for evaluating advancements in their performance. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
49. Documentation of social determinants of health across individuals from different racial and ethnic groups in home healthcare.
- Author
-
Hobensack, Mollie, Scharp, Danielle, Song, Jiyoun, and Topaz, Maxim
- Abstract
Introduction Design Methods Results Conclusion Clinical Relevance Home healthcare (HHC) enables patients to receive healthcare services within their homes to manage chronic conditions and recover from illnesses. Recent research has identified disparities in HHC based on race or ethnicity. Social determinants of health (SDOH) describe the external factors influencing a patient's health, such as access to care and social support. Individuals from racially or ethnically minoritized communities are known to be disproportionately affected by SDOH. Existing evidence suggests that SDOH are documented in clinical notes. However, no prior study has investigated the documentation of SDOH across individuals from different racial or ethnic backgrounds in the HHC setting. This study aimed to (1) describe frequencies of SDOH documented in clinical notes by race or ethnicity and (2) determine associations between race or ethnicity and SDOH documentation.Retrospective data analysis.We conducted a cross‐sectional secondary data analysis of 86,866 HHC episodes representing 65,693 unique patients from one large HHC agency in New York collected between January 1, 2015, and December 31, 2017. We reported the frequency of six SDOH (physical environment, social environment, housing and economic circumstances, food insecurity, access to care, and education and literacy) documented in clinical notes across individuals reported as Asian/Pacific Islander, Black, Hispanic, multi‐racial, Native American, or White. We analyzed differences in SDOH documentation by race or ethnicity using logistic regression models.Compared to patients reported as White, patients across other racial or ethnic groups had higher frequencies of SDOH documented in their clinical notes. Our results suggest that race or ethnicity is associated with SDOH documentation in HHC.As the study of SDOH in HHC continues to evolve, our results provide a foundation to evaluate social information in the HHC setting and understand how it influences the quality of care provided.The results of this exploratory study can help clinicians understand the differences in SDOH across individuals from different racial and ethnic groups and serve as a foundation for future research aimed at fostering more inclusive HHC documentation practices. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
50. Preoperative prediction model for risk of readmission after total joint replacement surgery: a random forest approach leveraging NLP and unfairness mitigation for improved patient care and cost-effectiveness.
- Author
-
Digumarthi, Varun, Amin, Tapan, Kanu, Samuel, Mathew, Joshua, Edwards, Bryan, Peterson, Lisa A, Lundy, Matthew E, and Hegarty, Karen E
- Subjects
- *
RISK assessment , *PREOPERATIVE period , *RANDOM forest algorithms , *PREDICTION models , *PATIENT readmissions , *DESCRIPTIVE statistics , *NATURAL language processing , *NURSING care facilities , *ARTIFICIAL joints , *ELECTRONIC health records , *DATA analysis software , *MACHINE learning , *ALGORITHMS - Abstract
Background: The Center for Medicare and Medicaid Services (CMS) imposes payment penalties for readmissions following total joint replacement surgeries. This study focuses on total hip, knee, and shoulder arthroplasty procedures as they account for most joint replacement surgeries. Apart from being a burden to healthcare systems, readmissions are also troublesome for patients. There are several studies which only utilized structured data from Electronic Health Records (EHR) without considering any gender and payor bias adjustments. Methods: For this study, dataset of 38,581 total knee, hip, and shoulder replacement surgeries performed from 2015 to 2021 at Novant Health was gathered. This data was used to train a random forest machine learning model to predict the combined endpoint of emergency department (ED) visit or unplanned readmissions within 30 days of discharge or discharge to Skilled Nursing Facility (SNF) following the surgery. 98 features of laboratory results, diagnoses, vitals, medications, and utilization history were extracted. A natural language processing (NLP) model finetuned from Clinical BERT was used to generate an NLP risk score feature for each patient based on their clinical notes. To address societal biases, a feature bias analysis was performed in conjunction with propensity score matching. A threshold optimization algorithm from the Fairlearn toolkit was used to mitigate gender and payor biases to promote fairness in predictions. Results: The model achieved an Area Under the Receiver Operating characteristic Curve (AUROC) of 0.738 (95% confidence interval, 0.724 to 0.754) and an Area Under the Precision-Recall Curve (AUPRC) of 0.406 (95% confidence interval, 0.384 to 0.433). Considering an outcome prevalence of 16%, these metrics indicate the model's ability to accurately discriminate between readmission and non-readmission cases within the context of total arthroplasty surgeries while adjusting patient scores in the model to mitigate bias based on patient gender and payor. Conclusion: This work culminated in a model that identifies the most predictive and protective features associated with the combined endpoint. This model serves as a tool to empower healthcare providers to proactively intervene based on these influential factors without introducing bias towards protected patient classes, effectively mitigating the risk of negative outcomes and ultimately improving quality of care regardless of socioeconomic factors. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
Catalog
Discovery Service for Jio Institute Digital Library
For full access to our library's resources, please sign in.