6,511 results on '"Supervised Machine Learning"'
Search Results
2. Age-stratified predictions of suicide attempts using machine learning in middle and late adolescence.
- Author
-
Kusuma, Karen, Larsen, Mark, Quiroz, Juan C., and Torok, Michelle
- Subjects
- *
MACHINE learning , *SUICIDE risk assessment , *ATTEMPTED suicide , *SUICIDE risk factors , *BOOSTING algorithms - Abstract
Prevalence of suicidal behaviour increases rapidly in middle to late adolescence. Predicting suicide attempts across different ages would enhance our understanding of how suicidal behaviour manifests in this period of rapid development. This study aimed to develop separate models to predict suicide attempts within a cohort at middle and late adolescence. It also sought to examine differences between the models derived across both developmental stages. This study used data from the nationally representative Longitudinal Study of Australian Children (N = 2266). We selected over 700 potential suicide attempt predictors measured via self-report questionnaires, and linked healthcare and education administrative datasets. Logistic regression, random forests, and gradient boosting algorithms were developed to predict suicide attempts across two stages (mid-adolescence: 14–15 years; late adolescence: 18–19 years) using predictors sampled two years prior (mid-adolescence: 12–13 years; late adolescence: 16–17 years). The late adolescence models (AUROC = 0.77–0.88, F1-score = 0.22–0.28, Sensitivity = 0.54–0.64) performed better than the mid-adolescence models (AUROC = 0.70–0.76, F1-score = 0.12–0.19, Sensitivity = 0.40–0.64). The most important features for predicting suicide attempts in mid-adolescence were mostly school-related, while the most important features in late adolescence included measures of prior suicidality, psychosocial health, and future plans. To date, this is the first study to use machine learning models to predict suicide attempts at different ages. Our findings suggest that the optimal suicide risk prediction model differs by stage of adolescence. Future research and interventions should consider that risk presentations can change rapidly during adolescence. • Suicide risk profiles differ according to demographic characteristics such as age. • We used machine learning to predict suicide attempts in middle and late adolescence. • The mid and late adolescence models performed differently. • Suicide risk formulations should consider the adolescent's developmental stage. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
3. Hybrid logistic regression support vector model to enhance prediction of bipolar disorder.
- Author
-
Agnihotri, Nisha and Prasad, Sanjeev Kumar
- Subjects
SUPERVISED learning ,SUPPORT vector machines ,MEDICAL personnel ,BIPOLAR disorder ,LOGISTIC regression analysis ,FAMILY history (Medicine) - Abstract
Bipolar disorder has become one of the major mental health issues due to stressed life around the world. This is the major reason for suicides these days as these people are unable to convey their feeling and emotions to others. This proposed research shows the logistic regression and support vector machine hybrid model to predict bipolar disorder in patients is to develop an accurate and reliable model that can effectively predict the presence of bipolar disorder in patients based on their clinical and demographic data. The purpose is to make a framework that can help healthcare professionals diagnose bipolar disorder early, thereby enabling timely and appropriate treatment to be provided. The model should take into account various patient-specific features, such as age, gender, family history, medication use, and other medical conditions, in addition to relevant clinical and demographic variables. The aim is to create a model that can accurately classify patients with bipolar disorder and non-bipolar disorder patients while minimizing false-positive and false-negative classifications. The work shows improvement in evaluation detection in performance with hybrid logistic support vector regression (LSVR) to detect disorder and protect them to avoid worse situation. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
4. Medial prefrontal cortex neurotransmitter abnormalities in posttraumatic stress disorder with and without comorbidity to major depression.
- Author
-
Swanberg, Kelley M., Prinsen, Hetty, Averill, Christopher L., Campos, Leonardo, Kurada, Abhinav V., Krystal, John H., Petrakis, Ismene L., Averill, Lynnette A., Rothman, Douglas L., Abdallah, Chadi G., and Juchem, Christoph
- Subjects
POST-traumatic stress disorder ,MENTAL depression ,PREFRONTAL cortex ,BRAIN abnormalities ,GLUTAMINE - Abstract
Posttraumatic stress disorder (PTSD) is a chronic psychiatric condition that follows exposure to a traumatic stressor. Though previous in vivo proton (1H) MRS) research conducted at 4 T or lower has identified alterations in glutamate metabolism associated with PTSD predisposition and/or progression, no prior investigations have been conducted at higher field strength. In addition, earlier studies have not extensively addressed the impact of psychiatric comorbidities such as major depressive disorder (MDD) on PTSD‐associated 1H‐MRS‐visible brain metabolite abnormalities. Here we employ 7 T 1H MRS to examine concentrations of glutamate, glutamine, GABA, and glutathione in the medial prefrontal cortex (mPFC) of PTSD patients with MDD (PTSD+MDD+; N = 6) or without MDD (PTSD+MDD−; N = 5), as well as trauma‐unmatched controls without PTSD but with MDD (PTSD−MDD+; N = 9) or without MDD (PTSD−MDD−; N = 18). Participants with PTSD demonstrated decreased ratios of GABA to glutamine relative to healthy PTSD−MDD− controls but no single‐metabolite abnormalities. When comorbid MDD was considered, however, MDD but not PTSD diagnosis was significantly associated with increased mPFC glutamine concentration and decreased glutamate:glutamine ratio. In addition, all participants with PTSD and/or MDD collectively demonstrated decreased glutathione relative to healthy PTSD−MDD− controls. Despite limited findings in single metabolites, patterns of abnormality in prefrontal metabolite concentrations among individuals with PTSD and/or MDD enabled supervised classification to separate them from healthy controls with 80+% sensitivity and specificity, with glutathione, glutamine, and myoinositol consistently among the most informative metabolites for this classification. Our findings indicate that MDD can be an important factor in mPFC glutamate metabolism abnormalities observed using 1H MRS in cohorts with PTSD. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
5. Critical comparison of potential machine learning methods for lightning thermal damage assessment of composite laminates.
- Author
-
Lee, Juhyeong, Millen, Scott L.J., and Xu, Xiaodong
- Subjects
- *
SUPERVISED learning , *DAMAGE models , *FIBROUS composites , *ELECTRIC conductivity , *MACHINE learning , *LAMINATED materials - Abstract
The present study assesses the potential of using machine learning (ML) methods to predict the extent of lightning thermal damage in fiber-reinforced composite laminates using three supervised machine learning (SML) algorithms: (1) linear regression (LR), (2) decision tree (DT)-based, and (3) MLP models. These models were based on the 10 most significant factors that influence the severity of lightning damage, including three current waveform parameters, four material configurations, and three orthogonal electrical conductivities of each composite. All models demonstrated good performance with coefficient of determination (R2) values between 0.84 ~ 0.96. The multilayer perceptron (MLP) regression model trained with the lightning matrix damage dataset showed the most promising results (R2 > 0.94). Additional hyperparameter optimization was performed to improve the prediction performance of the baseline MLP model. The hyperparameter optimization (Adam optimizer, tanh activation function, and three hidden layers with 234 neurons) slightly improved the performance of the baseline MLP model by ~0.02, but achieved faster convergence. This result suggests that the baseline MLP model trained with the lightning matrix damage dataset is sufficiently accurate and robust. This paper highlights that ML-informed regression models can serve as an efficient first pass-estimator of lightning matrix damage in composite laminates, potentially reducing the amount of extremely time-consuming and expensive laboratory-scale lightning tests or streamlining the development of complex lightning damage models for future design. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
6. Supervised machine learning of outbred mouse genotypes to predict hepatic immunological tolerance of individuals.
- Author
-
Morita-Nakagawa, Miwa, Okamura, Kohji, Nakabayashi, Kazuhiko, Inanaga, Yukiko, Shimizu, Seiichi, Guo, Wen-Zhi, Fujino, Masayuki, and Li, Xiao-Kang
- Subjects
- *
CONVOLUTIONAL neural networks , *SUPERVISED learning , *LIVER transplantation , *IMMUNOLOGICAL tolerance , *MACHINE learning , *DEEP learning , *MICE - Abstract
It is essential to elucidate the molecular mechanisms underlying liver transplant tolerance and rejection. In cases of mouse liver transplantation between inbred strains, immunological rejection of the allograft is reduced with spontaneous apoptosis without immunosuppressive drugs, which differs from the actual clinical result. This may be because inbred strains are genetically homogeneous and less heterogeneous than others. We exploited outbred CD1 mice, which show highly heterogeneous genotypes among individuals, to search for biomarkers related to immune responses and to construct a model for predicting the outcome of liver allografting. Of the 36 mice examined, 18 died within 3 weeks after transplantation, while the others survived for more than 6 weeks. Whole-exome sequencing of the 36 donors revealed more than 9 million variants relative to the C57BL/6 J reference. We selected 6517 single-nucleotide and indel variants and performed machine learning to determine whether or not we could predict the prognosis of each genotype. Models were built by both deep learning with a one-dimensional convolutional neural network and linear classification and evaluated by leave-one-out cross-validation. Given that one short-lived mouse died early in an accident, the models perfectly predicted the outcome of all individuals, suggesting the importance of genotype collection. In addition, linear classification models provided a list of loci potentially responsible for these responses. The present methods as well as results is likely to be applicable to liver transplantation in humans. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
7. Construction of Fatigue Criteria Through Positive‐Unlabeled Learning.
- Author
-
Coudray, Olivier, Bristiel, Philippe, Dinis, Miguel, Keribin, Christine, and Pamphile, Patrick
- Subjects
- *
SUPERVISED learning - Abstract
The reliability of vehicles is a significant issue for automotive manufacturers, with mechanical fatigue representing a pivotal aspect of the design process. To accelerate the development of novel mechanical components, automotive manufacturers are increasingly employing numerical simulations with the objective of markedly reducing the number of prototype validation tests. This requires the utilization of efficient fatigue criteria capable of accurately identifying critical zones in numerical models. However, the current fatigue criteria frequently fail to demonstrate a satisfactory correlation with the results obtained from fatigue test rigs. In response, this paper proposes a probabilistic Dang Van criterion that accounts for variability in fatigue results within a multiaxial context. Additionally, a fatigue database comprising numerical results and test reports on automotive chassis components is introduced. A novel approach utilizing positive‐unlabeled learning is developed to enhance the predictive accuracy of the fatigue criterion. Its effectiveness is demonstrated through application to the fatigue database. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
8. Analyzing Unimproved Drinking Water Sources and Their Determinants Using Supervised Machine Learning: Evidence from the Somaliland Demographic Health Survey 2020.
- Author
-
Ismail, Hibak M., Muse, Abdisalam Hassan, Hassan, Mukhtar Abdi, Muse, Yahye Hassan, and Nadarajah, Saralees
- Abstract
Access to clean and safe drinking water is a fundamental human right. Despite global efforts, including the UN's "Water for Life" program, a significant portion of the population in developing countries, including Somaliland, continues to rely on unimproved water sources. These unimproved sources contribute to poor health outcomes, particularly for children. This study aimed to investigate the factors associated with the use of unimproved drinking water sources in Somaliland by employing supervised machine learning models to predict patterns and determinants based on data from the 2020 Somaliland Demographic and Health Survey (SHDS). Secondary data from SHDS 2020 were used, encompassing 8384 households across Somaliland. A multilevel logistic regression model was applied to analyze the individual- and community-level factors influencing the use of unimproved water sources. In addition, machine learning models, including logistic regression, decision tree, random forest, support vector machine (SVM), and K-nearest neighbor (KNN), were compared in terms of accuracy, sensitivity, specificity, and other metrics using cross-validation techniques. This study uses supervised machine learning models to analyze unimproved drinking water sources in Somaliland, providing data-driven insights into the complex determinants of water access. This enhances predictive accuracy and informs targeted interventions, offering a robust framework for addressing water-related public health issues in Somaliland. The analysis identified key determinants of unimproved water source usage, including socioeconomic status, education, region, and household characteristics. The random forest model performed the best with an accuracy of 93.57% and an area under the curve (AUC) score of 98%. Decision tree and KNN also exhibited strong performance, while SVM had the lowest predictive accuracy. This study highlights the role of socioeconomic and community factors in determining access to clean drinking water in Somali Land. Factors such as age, education, gender, household wealth, media access, urban or rural residence, poverty level, and literacy level significantly influenced access. Local policies and resource availability also contribute to variations in access. These findings suggest that targeted interventions aimed at improving education, infrastructure, and community water management practices can significantly reduce reliance on unimproved water sources and improve the overall public health. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
9. Enhancing microgrid energy management through solar power uncertainty mitigation using supervised machine learning.
- Author
-
Elazab, Rasha, Dahab, Ahmed Abo, Adma, Maged Abo, and Hassan, Hany Abdo
- Subjects
SUPERVISED learning ,SOLAR radiation ,SUPPORT vector machines ,KRIGING ,RENEWABLE energy sources ,SOLAR energy - Abstract
This study addresses the inherent challenges associated with the limited flexibility of power systems, specifically emphasizing uncertainties in solar power due to dynamic regional and seasonal fluctuations in photovoltaic (PV) potential. The research introduces a novel supervised machine learning model that focuses on regression methods specifically tailored for advanced microgrid energy management within a 100% PV microgrid, i.e. a microgrid system that is powered entirely by solar energy, with no reliance on other energy sources such as fossil fuels or grid electricity. In this context, "PV" specifically denotes photovoltaic solar panels that convert sunlight into electricity. A distinctive feature of the model is its exclusive reliance on current solar radiation as an input parameter to minimize prediction errors, justified by the unique advantages of supervised learning. The performance of four well-established supervised machine learning models—Neural Networks (NN), Gaussian Process Regression (GPR), Support Vector Machines (SVM), and Linear Regression (LR)—known for effectively addressing short-term uncertainty in solar radiation, is thoroughly evaluated. Results underscore the superiority of the NN approach in accurately predicting solar irradiance across diverse geographical sites, including Cairo, Egypt; Riyadh, Saudi Arabia; Yuseong-gu, Daejeon, South Korea; and Berlin, Germany. The comprehensive analysis covers both Global Horizontal Irradiance (GHI) and Direct Normal Irradiance (DNI), demonstrating the model's efficacy in various solar environments. Additionally, the study emphasizes the practical implementation of the model within an Energy Management System (EMS) using Hybrid Optimization of Multiple Electric Renewables (HOMER) software, showcasing high accuracy in microgrid energy management. This validation attests to the economic efficiency and reliability of the proposed model. The calculated range of error, as the median error for cost analysis, varies from 2 to 6%, affirming the high accuracy of the proposed model. Highlights: Introduction of a novel approach: This study presents a pioneering methodology focusing exclusively on supervised machine learning techniques for short-term uncertainty management within 100% PV microgrids, aimed at optimizing energy management efficiency. Comprehensive comparative analysis: Through meticulous comparative investigation, various regression techniques for predicting solar radiation within PV microgrids are scrutinized, providing valuable insights into their efficacy across diverse environmental conditions. Validation across diverse locations: The validation of findings across four distinct geographical locations, encompassing both PV and concentrated PV (CPV) systems, substantially enhances the generalizability of the results, advancing the understanding of solar radiation prediction dynamics for renewable energy integration strategies. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
10. Categorizing political campaign messages on social media using supervised machine learning.
- Author
-
Stromer-Galley, Jennifer and Rossini, Patricia
- Subjects
- *
SUPERVISED learning , *POLITICAL campaigns , *MACHINE learning , *CONTENT analysis , *ALGORITHMS - Abstract
Scholars have access to a rich source of political discourse via social media. Although computational approaches to understand this communication are being used, they tend to be unsupervised and off-the-shelf algorithms to describe a corpus of messages. This article details our approach at using human-supervised machine learning to study political campaign messages. Although some declare this technique too labor-intensive, it provides theoretically informed classification, making it more accurate and reliable. This article describes the design decisions and accuracy of our algorithms, and the applicability of the approach to classifying messages from Facebook and Twitter across two cultures and to advertisements. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
11. Machine learning-based hyperspectral wavelength selection and classification of spider mite-infested cucumber leaves.
- Author
-
Mandrapa, Boris, Spohrer, Klaus, Wuttke, Dominik, Ruttensperger, Ute, Dieckhoff, Christine, and Müller, Joachim
- Subjects
SUPERVISED learning ,TWO-spotted spider mite ,MACHINE learning ,FEATURE selection ,SPIDER mites ,CUCUMBERS - Abstract
Two-spotted spider mite (Tetranychus urticae) is an important greenhouse pest. In cucumbers, heavy infestations lead to the complete loss of leaf assimilation surface, resulting in plant death. Symptoms caused by spider mite feeding alter the light reflection of leaves and could therefore be optically detected. Machine learning methods have already been employed to analyze spectral information in order to differentiate between healthy and spider mite-infested leaves of crops such as tomatoes or cotton. In this study, machine learning methods were applied to cucumbers. Hyperspectral data of leaves were recorded under controlled conditions. Effective wavelengths were identified using three feature selection methods. Subsequently, three supervised machine learning algorithms were used to classify healthy and spider mite-infested leaves. All combinations of feature selection and classification methods yielded accuracy of over 80%, even when using ten or five wavelengths. These results suggest that machine learning methods are a powerful tool for image-based detection of spider mites in cucumbers. In addition, due to the limited number of wavelengths, there is also substantial potential for practical application. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
12. Optimizing Learning: Predicting Research Competency via Statistical Proficiency.
- Author
-
Wongvorachan, Tarid, Srisuttiyakorn, Siwachoat, and Sriklaub, Kanit
- Subjects
HIGHER education ,CRITICAL thinking ,MACHINE learning ,EDUCATION research ,TEST scoring - Abstract
In higher education, the cultivation of research competency is pivotal for students' critical thinking development and their subsequent transition into the professional workforce. While statistics plays a fundamental role in supporting the completion of a research project, it is often perceived as challenging, particularly by students in majors outside mathematics or statistics. The connection between students' statistical proficiency and their research competency remains unexplored despite its significance. To address this gap, we utilize the supervised machine learning approach to predict students' research competency as represented by their performance in a research methods class, with predictors of students' proficiency in statistical topics. Predictors relating to students' learning behavior in a statistics course such as assignment completion and academic dishonesty are also included as auxiliary variables. Results indicate that the three primary categories of statistical skills—namely, the understanding of statistical concepts, proficiency in selecting appropriate statistical methods, and statistics interpretation skills—can be used to predict students' research competency as demonstrated by their final course scores and letter grades. This study advocates for strategic emphasis on the identified influential topics to enhance efficiency in developing students' research competency. The findings could inform instructors in adopting a strategic approach to teaching the statistical component of research for enhanced efficiency. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
13. Classification of Coastal Benthic Substrates Using Supervised and Unsupervised Machine Learning Models on North Shore of the St. Lawrence Maritime Estuary (Canada).
- Author
-
Labbé-Morissette, Guillaume, Leclercq, Théau, Charron-Morneau, Patrick, Gonthier, Dominic, Doiron, Dany, Chouaer, Mohamed-Ali, and Munang, Dominic Ndeh
- Subjects
MACHINE learning ,SUPERVISED learning ,GAUSSIAN mixture models ,ARTIFICIAL intelligence ,COASTAL mapping - Abstract
Classification of benthic substrates is a core necessity in many scientific fields like biology, ecology, or geology, with applications branching out to a variety of industries, from fisheries to oil and gas. In the first part, a comparative analysis of supervised learning algorithms has been conducted using geomorphometric features to generate benthic substrate maps of the coastal regions of the North Shore of Quebec in order to establish a quantitative assessment of performance to serve as a benchmark. In the second part, a new method using Gaussian mixture models is showcased on the same dataset. Finally, a side-by-side comparison of both methods is featured to provide a qualitative assessment of the new algorithm's ability to match human intuition. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
14. CRISP-DM-Based Data-Driven Approach for Building Energy Prediction Utilizing Indoor and Environmental Factors.
- Author
-
Elkabalawy, Moaaz, Al-Sakkaf, Abobakr, Mohammed Abdelkader, Eslam, and Alfalah, Ghasan
- Abstract
The significant energy consumption associated with the built environment demands comprehensive energy prediction modelling. Leveraging their ability to capture intricate patterns without extensive domain knowledge, supervised data-driven approaches present a marked advantage in adaptability over traditional physical-based building energy models. This study employs various machine learning models to predict energy consumption for an office building in Berkeley, California. To enhance the accuracy of these predictions, different feature selection techniques, including principal component analysis (PCA), decision tree regression (DTR), and Pearson correlation analysis, were adopted to identify key attributes of energy consumption and address collinearity. The analyses yielded nine influential attributes: heating, ventilation, and air conditioning (HVAC) system operating parameters, indoor and outdoor environmental parameters, and occupancy. To overcome missing occupancy data in the datasets, we investigated the possibility of occupancy-based Wi-Fi prediction using different machine learning algorithms. The results of the occupancy prediction modelling indicate that Wi-Fi can be used with acceptable accuracy in predicting occupancy count, which can be leveraged to analyze occupant comfort and enhance the accuracy of building energy models. Six machine learning models were tested for energy prediction using two different datasets: one before and one after occupancy prediction. Using a 10-fold cross-validation with an 8:2 training-to-testing ratio, the Random Forest algorithm emerged superior, exhibiting the highest R
2 value of 0.92 and the lowest RMSE of 3.78 when occupancy data were included. Additionally, an error propagation analysis was conducted to assess the impact of the occupancy-based Wi-Fi prediction model's error on the energy prediction model. The results indicated that Wi-Fi-based occupancy prediction can improve the data inputs for building energy models, leading to more accurate energy consumption predictions. The findings underscore the potential of integrating the developed energy prediction models with fault detection systems, model predictive controllers, and energy load shape analysis, ultimately enhancing energy management practices. [ABSTRACT FROM AUTHOR]- Published
- 2024
- Full Text
- View/download PDF
15. Supervised machine learning of outbred mouse genotypes to predict hepatic immunological tolerance of individuals
- Author
-
Miwa Morita-Nakagawa, Kohji Okamura, Kazuhiko Nakabayashi, Yukiko Inanaga, Seiichi Shimizu, Wen-Zhi Guo, Masayuki Fujino, and Xiao-Kang Li
- Subjects
Transplantation tolerance ,Supervised machine learning ,Liver transplantation ,Outbred mouse ,Medicine ,Science - Abstract
Abstract It is essential to elucidate the molecular mechanisms underlying liver transplant tolerance and rejection. In cases of mouse liver transplantation between inbred strains, immunological rejection of the allograft is reduced with spontaneous apoptosis without immunosuppressive drugs, which differs from the actual clinical result. This may be because inbred strains are genetically homogeneous and less heterogeneous than others. We exploited outbred CD1 mice, which show highly heterogeneous genotypes among individuals, to search for biomarkers related to immune responses and to construct a model for predicting the outcome of liver allografting. Of the 36 mice examined, 18 died within 3 weeks after transplantation, while the others survived for more than 6 weeks. Whole-exome sequencing of the 36 donors revealed more than 9 million variants relative to the C57BL/6 J reference. We selected 6517 single-nucleotide and indel variants and performed machine learning to determine whether or not we could predict the prognosis of each genotype. Models were built by both deep learning with a one-dimensional convolutional neural network and linear classification and evaluated by leave-one-out cross-validation. Given that one short-lived mouse died early in an accident, the models perfectly predicted the outcome of all individuals, suggesting the importance of genotype collection. In addition, linear classification models provided a list of loci potentially responsible for these responses. The present methods as well as results is likely to be applicable to liver transplantation in humans.
- Published
- 2024
- Full Text
- View/download PDF
16. Towards accurate screening and prevention for PTSD (2-ASAP): protocol of a longitudinal prospective cohort study
- Author
-
Jeanet F. Karchoud, Chris M. Hoeboer, Greta Piwanski, Juanita A. Haagsma, Miranda Olff, Rens van de Schoot, and Mirjam van Zuiden
- Subjects
PTSD ,Trauma ,Longitudinal ,Supervised machine learning ,Sex ,Gender ,Psychiatry ,RC435-571 - Abstract
Abstract Background Effective preventive interventions for PTSD rely on early identification of individuals at risk for developing PTSD. To establish early post-trauma who are at risk, there is a need for accurate prognostic risk screening instruments for PTSD that can be widely implemented in recently trauma-exposed adults. Achieving such accuracy and generalizability requires external validation of machine learning classification models. The current 2-ASAP cohort study will perform external validation on both full and minimal feature sets of supervised machine learning classification models assessing individual risk to follow an adverse PTSD symptom trajectory over the course of 1 year. We will derive these models from the TraumaTIPS cohort, separately for men and women. Method The 2-ASAP longitudinal cohort will include N = 863 adults (N = 436 females, N = 427 males) who were recently exposed to acute civilian trauma. We will include civilian victims of accidents, crime and calamities at Victim Support Netherlands; and who were presented for medical evaluation of (suspected) traumatic injuries by emergency transportation to the emergency department. The baseline assessment within 2 months post-trauma will include self-report questionnaires on demographic, medical and traumatic event characteristics; potential risk and protective factors for PTSD; PTSD symptom severity and other adverse outcomes; and current best-practice PTSD screening instruments. Participants will be followed at 3, 6, 9, and 12 months post-trauma, assessing PTSD symptom severity and other adverse outcomes via self-report questionnaires. Discussion The ultimate goal of our study is to improve accurate screening and prevention for PTSD in recently trauma-exposed civilians. To enable future large-scale implementation, we will use self-report data to inform the prognostic models; and we will derive a minimal feature set of the classification models. This can be transformed into a short online screening instrument that is user-friendly for recently trauma-exposed adults to fill in. The eventual short online screening instrument will classify early post-trauma which adults are at risk for developing PTSD. Those at risk can be targeted and may subsequently benefit from preventive interventions, aiming to reduce PTSD and relatedly improve psychological, functional and economic outcomes.
- Published
- 2024
- Full Text
- View/download PDF
17. Enhancing microgrid energy management through solar power uncertainty mitigation using supervised machine learning
- Author
-
Rasha Elazab, Ahmed Abo Dahab, Maged Abo Adma, and Hany Abdo Hassan
- Subjects
Solar radiation prediction ,Supervised machine learning ,Neural networks ,Energy management system ,HOMER ,Microgrid ,Energy industries. Energy policy. Fuel trade ,HD9502-9502.5 - Abstract
Abstract This study addresses the inherent challenges associated with the limited flexibility of power systems, specifically emphasizing uncertainties in solar power due to dynamic regional and seasonal fluctuations in photovoltaic (PV) potential. The research introduces a novel supervised machine learning model that focuses on regression methods specifically tailored for advanced microgrid energy management within a 100% PV microgrid, i.e. a microgrid system that is powered entirely by solar energy, with no reliance on other energy sources such as fossil fuels or grid electricity. In this context, “PV” specifically denotes photovoltaic solar panels that convert sunlight into electricity. A distinctive feature of the model is its exclusive reliance on current solar radiation as an input parameter to minimize prediction errors, justified by the unique advantages of supervised learning. The performance of four well-established supervised machine learning models—Neural Networks (NN), Gaussian Process Regression (GPR), Support Vector Machines (SVM), and Linear Regression (LR)—known for effectively addressing short-term uncertainty in solar radiation, is thoroughly evaluated. Results underscore the superiority of the NN approach in accurately predicting solar irradiance across diverse geographical sites, including Cairo, Egypt; Riyadh, Saudi Arabia; Yuseong-gu, Daejeon, South Korea; and Berlin, Germany. The comprehensive analysis covers both Global Horizontal Irradiance (GHI) and Direct Normal Irradiance (DNI), demonstrating the model’s efficacy in various solar environments. Additionally, the study emphasizes the practical implementation of the model within an Energy Management System (EMS) using Hybrid Optimization of Multiple Electric Renewables (HOMER) software, showcasing high accuracy in microgrid energy management. This validation attests to the economic efficiency and reliability of the proposed model. The calculated range of error, as the median error for cost analysis, varies from 2 to 6%, affirming the high accuracy of the proposed model.
- Published
- 2024
- Full Text
- View/download PDF
18. Tailoring pretext tasks to improve self-supervised learning in histopathologic subtype classification of lung adenocarcinomas.
- Author
-
Ding, Ruiwen, Yadav, Anil, Rodriguez, Erika, Araujo Lemos da Silva, Ana, and Hsu, William
- Subjects
Histologic subtype classification ,Histopathology ,Lung adenocarcinoma ,Pretext task ,Self-supervised learning ,Humans ,Adenocarcinoma of Lung ,Lung Neoplasms ,Supervised Machine Learning ,Neural Networks ,Computer ,Image Interpretation ,Computer-Assisted - Abstract
Lung adenocarcinoma (LUAD) is a morphologically heterogeneous disease with five predominant histologic subtypes. Fully supervised convolutional neural networks can improve the accuracy and reduce the subjectivity of LUAD histologic subtyping using hematoxylin and eosin (H&E)-stained whole slide images (WSIs). However, developing supervised models with good prediction accuracy usually requires extensive manual data annotation, which is time-consuming and labor-intensive. This work proposes three self-supervised learning (SSL) pretext tasks to reduce labeling effort. These tasks not only leverage the multi-resolution nature of the H&E WSIs but also explicitly consider the relevance to the downstream task of classifying the LUAD histologic subtypes. Two tasks involve predicting the spatial relationship between tiles cropped from lower and higher magnification WSIs. We hypothesize that these tasks induce the model to learn to distinguish different tissue structures presented in the images, thus benefiting the downstream classification. The third task involves predicting the eosin stain from the hematoxylin stain, inducing the model to learn cytoplasmic features relevant to LUAD subtypes. The effectiveness of the three proposed SSL tasks and their ensemble was demonstrated by comparison with other state-of-the-art pretraining and SSL methods using three publicly available datasets. Our work can be extended to any other cancer type where tissue architectural information is important. The model could be used to expedite and complement the process of routine pathology diagnosis tasks. The code is available at https://github.com/rina-ding/ssl_luad_classification.
- Published
- 2023
19. Analysis of data-driven approaches for radar target classification
- Author
-
Coşkun, Aysu and Bilicz, Sándor
- Published
- 2024
- Full Text
- View/download PDF
20. Middle ear-acquired cholesteatoma diagnosis based on CT scan image mining using supervised machine learning models
- Author
-
Naouar Ouattassi, Mustapha Maaroufi, Hajar Slaoui, Taha Benateya Andaloussi, Arsalane Zarghili, and Mohamed Nouredine El Amine El Alami
- Subjects
Middle ear cholesteatoma ,Chronic suppurative otitis media ,Supervised machine learning ,Image mining ,Medicine (General) ,R5-920 ,Science - Abstract
Abstract Background Distinguishing between middle ear cholesteatoma and chronic suppurative otitis media (CSOM) is an ongoing challenge. While temporal bone computed tomography (CT) scan is highly accurate for diagnosing middle ear conditions, its specificity in discerning between cholesteatoma and CSOM is only moderate. To address this issue, we utilized trained machine learning models to enhance the specificity of temporal bone CT scan in diagnosing middle ear cholesteatoma. Our database consisted of temporal bone CT scan native images from 122 patients diagnosed with middle ear cholesteatoma and a control group of 115 patients diagnosed with CSOM, with both groups labeled based on surgical findings. We preprocessed the native images to isolate the region of interest and then utilized the Inception V3 convolutional neural network for image embedding into data vectors. Classification was performed using machine learning models including support vector machine (SVM), k-nearest neighbors (k-NN), random forest, and neural network. Statistical metrics employed to interpret the results included classification accuracy, precision, recall, F1 score, confusion matrix, area under the receiver operating characteristic curve (AUC), and FreeViz diagram. Results Our training dataset comprised 5390 images, and the testing dataset included 125 different images. The neural network, k-NN, and SVM models demonstrated significantly higher relevance in terms of classification accuracy, precision, and recall compared to the random forest model. For instance, the F1 scores were 0.974, 0.987, and 0.897, respectively, for the former three models, in contrast to 0.661 for the random forest model. Conclusion The performance metrics of the presented trained machine learning models hold promising prospects as potentially clinically useful aids.
- Published
- 2024
- Full Text
- View/download PDF
21. Enhancing machine learning-based seismic facies classification through attribute selection: application to 3D seismic data from the Malay and Sabah Basins, offshore Malaysia
- Author
-
Ismailalwali Babikir, Abdul Halim Abdul Latiff, Mohamed Elsaadany, Hadyan Pratama, Muhammad Sajid, Salbiah Mad Sahad, Muhammad Anwar Ishak, and Carolan Laudon
- Subjects
Seismic attributes ,Seismic facies classification ,Attribute selection ,Supervised machine learning ,Model performance ,Geophysics. Cosmic physics ,QC801-809 - Abstract
Abstract Over the past few years, the use of machine learning has gained considerable momentum in many industries, including exploration seismic. While supervised machine learning is increasingly being used in seismic data analysis, some obstacles hinder its widespread application. Seismic facies classification—a crucial aspect in this field—particularly faces challenges such as the selection of appropriate input attributes. Plethora of seismic attributes have been created over the years, and new ones are still coming out. Yet, several have been deemed redundant or geologically meaningless. In the context of machine learning, it is crucial to avoid these redundant and irrelevant attributes as they can result in overfitting, building unnecessary complex models, and prolonging computational time. The current study incorporates an attribute selection approach to seismic facies classification and evaluates the importance of several available seismic attributes. Two datasets from the AN Field and the Dangerous Grounds region offshore Malaysia were utilized. Several attribute selection techniques were evaluated, with most of them yielding perfect attribute subsets for the AN dataset. However, only the wrapper and embedded methods could produce optimal subsets for the more complex Dangerous Grounds dataset. In both datasets, distinguishing the targeted seismic facies was mainly dependent on amplitude, spectral, and gray-level co-occurrence matrix attributes. Furthermore, spectral magnitude components played a significant role in classifying the facies of the Dangerous Grounds broadband data. The study demonstrated the importance of attribute selection, established a workflow, and identified significant attributes that could enhance seismic facies classification in Malaysian basins and similar geologic settings.
- Published
- 2024
- Full Text
- View/download PDF
22. Optimizing Learning: Predicting Research Competency via Statistical Proficiency
- Author
-
Tarid Wongvorachan, Siwachoat Srisuttiyakorn, and Kanit Sriklaub
- Subjects
research competency ,statistics ,supervised machine learning ,Education (General) ,L7-991 ,Theory and practice of education ,LB5-3640 - Abstract
In higher education, the cultivation of research competency is pivotal for students’ critical thinking development and their subsequent transition into the professional workforce. While statistics plays a fundamental role in supporting the completion of a research project, it is often perceived as challenging, particularly by students in majors outside mathematics or statistics. The connection between students’ statistical proficiency and their research competency remains unexplored despite its significance. To address this gap, we utilize the supervised machine learning approach to predict students’ research competency as represented by their performance in a research methods class, with predictors of students’ proficiency in statistical topics. Predictors relating to students’ learning behavior in a statistics course such as assignment completion and academic dishonesty are also included as auxiliary variables. Results indicate that the three primary categories of statistical skills—namely, the understanding of statistical concepts, proficiency in selecting appropriate statistical methods, and statistics interpretation skills—can be used to predict students’ research competency as demonstrated by their final course scores and letter grades. This study advocates for strategic emphasis on the identified influential topics to enhance efficiency in developing students’ research competency. The findings could inform instructors in adopting a strategic approach to teaching the statistical component of research for enhanced efficiency.
- Published
- 2024
- Full Text
- View/download PDF
23. Machine learning allows robust classification of visceral fat in women with obesity using common laboratory metrics
- Author
-
Flavio Palmieri, Nidà Farooq Akhtar, Adriana Pané, Amanda Jiménez, Romina Paula Olbeyra, Judith Viaplana, Josep Vidal, Ana de Hollanda, Pau Gama-Perez, Josep C. Jiménez-Chillarón, and Pablo M. Garcia-Roves
- Subjects
Obesity ,Women ,Visceral adipose tissue ,Classification ,Supervised machine learning ,Medicine ,Science - Abstract
Abstract The excessive accumulation and malfunctioning of visceral adipose tissue (VAT) is a major determinant of increased risk of obesity-related comorbidities. Thus, risk stratification of people living with obesity according to their amount of VAT is of clinical interest. Currently, the most common VAT measurement methods include mathematical formulae based on anthropometric dimensions, often biased by human measurement errors, bio-impedance, and image techniques such as X-ray absorptiometry (DXA) analysis, which requires specialized equipment. However, previous studies showed the possibility of classifying people living with obesity according to their VAT through blood chemical concentrations by applying machine learning techniques. In addition, most of the efforts were spent on men living with obesity while little was done for women. Therefore, this study aims to compare the performance of the multilinear regression model (MLR) in estimating VAT and six different supervised machine learning classifiers, including logistic regression (LR), support vector machine and decision tree-based models, to categorize 149 women living with obesity. For clustering, the study population was categorized into classes 0, 1, and 2 according to their VAT and the accuracy of each MLR and classification model was evaluated using DXA-data (DXAdata), blood chemical concentrations (BLDdata), and both DXAdata and BLDdata together (ALLdata). Estimation error and $$\hbox {R}^{2}$$ R 2 were computed for MLR, while receiver operating characteristic (ROC) and precision-recall curves (PR) area under the curve (AUC) were used to assess the performance of every classification model. MLR models showed a poor ability to estimate VAT with mean absolute error $$\ge 401.40$$ ≥ 401.40 and $$\hbox {R}^{2} \le 0.62$$ R 2 ≤ 0.62 in all the datasets. The highest accuracy was found for LR with values of 0.57, 0.63, and 0.53 for ALLdata, DXAdata, and BLDdata, respectively. The ROC AUC showed a poor ability of both ALLdata and DXAdata to distinguish class 1 from classes 0 and 2 (AUC = 0.31, 0.71, and 0.85, respectively) as also confirmed by PR (AUC = 0.24, 0.57, and 0.73, respectively). However, improved performances were obtained when applying LR model to BLDdata (ROC AUC $$\ge $$ ≥ 0.61 and PR AUC $$\ge $$ ≥ 0.42), especially for class 1. These results seem to suggest that, while a direct and reliable estimation of VAT was not possible in our cohort, blood sample-derived information can robustly classify women living with obesity by machine learning-based classifiers, a fact that could benefit the clinical practice, especially in those health centres where medical imaging devices are not available. Nonetheless, these promising findings should be further validated over a larger population.
- Published
- 2024
- Full Text
- View/download PDF
24. Mathematical modeling and machine learning-based optimization for enhancing biofiltration efficiency of volatile organic compounds
- Author
-
Muhammad Sulaiman, Osamah Ibrahim Khalaf, Naveed Ahmad Khan, Fahad Sameer Alshammari, and Habib Hamam
- Subjects
Mathematical modeling ,Reaction mechanism ,Volatile organic compounds ,Michaelis-Menten kinetics ,Supervised machine learning ,Elman neural networks ,Medicine ,Science - Abstract
Abstract Biofiltration is a method of pollution management that utilizes a bioreactor containing live material to absorb and destroy pollutants biologically. In this paper, we investigate mathematical models of biofiltration for mixing volatile organic compounds (VOCs) for instance hydrophilic (methanol) and hydrophobic ( $$\alpha$$ α -pinene). The system of nonlinear diffusion equations describes the Michaelis-Menten kinetics of the enzymic chemical reaction. These models represent the chemical oxidation in the gas phase and mass transmission within the air-biofilm junction. Furthermore, for the numerical study of the saturation of $$\alpha$$ α -pinene and methanol in the biofilm and gas state, we have developed an efficient supervised machine learning algorithm based on the architecture of Elman neural networks (ENN). Moreover, the Levenberg-Marquardt (LM) optimization paradigm is used to find the parameters/ neurons involved in the ENN architecture. The approximation to a solutions found by the ENN-LM technique for methanol saturation and $$\alpha$$ α -pinene under variations in different physical parameters are allegorized with the numerical results computed by state-of-the-art techniques. The graphical and statistical illustration of indications of performance relative to the terms of absolute errors, mean absolute deviations, computational complexity, and mean square error validates that our results perfectly describe the real-life situation and can further be used for problems arising in chemical engineering.
- Published
- 2024
- Full Text
- View/download PDF
25. Analysis of organic and mineral nitrogen, total organic carbon and humic fractions in Ferralsols: an approach using Vis-NIR-SWIR, MIR and X-ray fluorescence spectroscopy
- Author
-
Bruna Coelho de Lima, Carlos H. dos Santos, Carlos S. Tiritan, José A. M. Demattê, Andres M. R. Gomez, Heidy S. R. Albarracín, and Bruno A. Bartsch
- Subjects
Supervised machine learning ,Spectroscopic techniques ,Accurate fertilization Sandy soil ,Ammonium and nitrate quantification ,Organic matter quantification ,Soil health ,Environmental sciences ,GE1-350 - Abstract
Abstract This work aimed to develop suitable predictive models for ammonium, nitrate, total nitrogen, total organic carbon and soil humic fractions, for Ferralsols, using Vis-NIR-SWIR, MIR and X-ray fluorescence spectroscopic techniques in conjunction with machine learning algorithms, Cubist, PLSR, Random Forest and Support Vector Machine. Chemical analyzes were carried out to determine nitrate, total nitrogen, total organic carbon and chemical fractionation of soil organic matter, as well as spectral analyzes using Vis-NIR-SWIR spectroscopy, MIR and X-ray fluorescence. The spectroscopy results were processed using RStudio v. 4.1.3, applying Cusbist, PLSR, Random Forest and Support Vector Machine machine learning algorithms to create predictive models and describe spectral curves and Pearson correlation. Of the prediction models developed for nitrogen, total organic carbon and humic fractions, the PLSR and Support Vector Machine algorithms presented the best predictive performances. The descriptive analysis of the spectra identified the main absorption bands and the location of the bands sensitive to the attributes of interest. The correlation analysis proposed that the use of Vis-NIR-SWIR, MIR and XRF spectroscopic techniques were effective in predicting the contents of nitrogen, total organic carbon and humic fractions in soil with a medium sandy texture. However, it is important to highlight that each technique has its characteristic mechanism of action, Vis-NIR-SWIR and MIR detect the element based on overtones and fundamental tones, while XRF is based on the atomic number of the elements or elemental association.
- Published
- 2024
- Full Text
- View/download PDF
26. Prediction of precancerous cervical cancer lesions among women living with HIV on antiretroviral therapy in Uganda: a comparison of supervised machine learning algorithms
- Author
-
Florence Namalinzi, Kefas Rimamnuskeb Galadima, Robinah Nalwanga, Isaac Sekitoleko, and Leon Fidele Ruganzu Uwimbabazi
- Subjects
Cervical cancer ,Supervised machine learning ,Women living with HIV ,Gynecology and obstetrics ,RG1-991 ,Public aspects of medicine ,RA1-1270 - Abstract
Abstract Background Cervical cancer (CC) is among the most prevalent cancer types among women with the highest prevalence in low- and middle-income countries (LMICs). It is a curable disease if detected early. Machine learning (ML) techniques can aid in early detection and prediction thus reducing screening and treatment costs. This study focused on women living with HIV (WLHIV) in Uganda. Its aim was to identify the best predictors of CC and the supervised ML model that best predicts CC among WLHIV. Methods Secondary data that included 3025 women from three health facilities in central Uganda was used. A multivariate binary logistic regression and recursive feature elimination with random forest (RFERF) were used to identify the best predictors. Five models; logistic regression (LR), random forest (RF), K-Nearest neighbor (KNN), support vector machine (SVM), and multi-layer perceptron (MLP) were applied to identify the out-performer. The confusion matrix and the area under the receiver operating characteristic curve (AUC/ROC) were used to evaluate the models. Results The results revealed that duration on antiretroviral therapy (ART), WHO clinical stage, TPT status, Viral load status, and family planning were commonly selected by the two techniques and thus highly significant in CC prediction. The RF from the RFERF-selected features outperformed other models with the highest scores of 90% accuracy and 0.901 AUC. Conclusion Early identification of CC and knowledge of the risk factors could help control the disease. The RF outperformed other models applied regardless of the selection technique used. Future research can be expanded to include ART-naïve women in predicting CC.
- Published
- 2024
- Full Text
- View/download PDF
27. Classification of Coastal Benthic Substrates Using Supervised and Unsupervised Machine Learning Models on North Shore of the St. Lawrence Maritime Estuary (Canada)
- Author
-
Guillaume Labbé-Morissette, Théau Leclercq, Patrick Charron-Morneau, Dominic Gonthier, Dany Doiron, Mohamed-Ali Chouaer, and Dominic Ndeh Munang
- Subjects
benthic habitat mapping ,benthic habitat classification ,supervised machine learning ,unsupervised machine learning ,artificial intelligence ,Geology ,QE1-996.5 - Abstract
Classification of benthic substrates is a core necessity in many scientific fields like biology, ecology, or geology, with applications branching out to a variety of industries, from fisheries to oil and gas. In the first part, a comparative analysis of supervised learning algorithms has been conducted using geomorphometric features to generate benthic substrate maps of the coastal regions of the North Shore of Quebec in order to establish a quantitative assessment of performance to serve as a benchmark. In the second part, a new method using Gaussian mixture models is showcased on the same dataset. Finally, a side-by-side comparison of both methods is featured to provide a qualitative assessment of the new algorithm’s ability to match human intuition.
- Published
- 2024
- Full Text
- View/download PDF
28. Optimizing the geometry of hunchbacked block-type gravity quay walls using non-linear dynamic analyses and supervised machine learning technique
- Author
-
B. Ebrahimian and َA.R. Zarnousheh Farahani
- Subjects
gravity quay wall ,broken-back geometry ,geometrical optimization ,non-linear dynamic analysis ,supervised machine learning ,Building construction ,TH1-9745 - Abstract
In the present study, the seismic behavior of hunchbacked block-type gravity quay walls rested on non-liquefiable dense seabed soil layer is investigated, and the optimal geometries for these wall types are proposed by performing non-linear time history dynamic analyses using Lagrangian explicit finite difference method. For this purpose, first, a reference numerical model of the hunchbacked quay wall is developed, and its seismic response is validated against the well-documented physical model tests. Then, the optimal hunch angles corresponding to the minimum horizontal displacement and zero rotation of the hunchbacked quay wall are estimated through the sensitivity analyses on the hunch angle of the wall, the friction angle of the backfill, and the ratio of hunch height to wall height. Subsequently, the statistical relationships are presented to predict the optimal hunch angle of the walls using multiple non-linear regressions based on the supervised machine learning technique. The results of non-linear dynamic analyses show that the deformation pattern, the movement mechanism, and, consequently, the seismic response of the hunchbacked quay wall change considerably with the variation of the hunch angle of the wall. In this regard, the rotation angle of the wall towards the seaside due to seismic loading decreases, and the deformation pattern and the movement mechanism of the hunchbacked quay wall alter from overturning towards the seaside to overturning towards the landside with an increase of the hunch angle. For all considered values of the ratio of hunch height to wall height and the backfill friction angle, increasing the hunch angle in the range of 25 to 35 degrees leads to a significant decrease in wall deformation. While increasing the hunch angle in the range of 35 to 50 degrees has less influence on reducing the wall deformation. For hunch angle values greater than 50 degrees, increasing the hunch angle has the opposite effect on improving the seismic performance of the hunchbacked quay wall and its seismic-induced deformations increase. Additionally, in the ratio of hunch height to wall height equal to 0.7, the optimal hunch angles corresponding to the zero wall rotation and the maximum reduction in the horizontal displacement of the wall decrease from 42.7 to 9. 23 degrees and from 53 to 34.5 degrees, respectively, with an increase of the friction angle of the backfill soil from 15 to 45 degrees.
- Published
- 2024
- Full Text
- View/download PDF
29. A Machine Learning Model for Predicting Sleep and Wakefulness Based on Accelerometry, Skin Temperature and Contextual Information
- Author
-
Logacjov A, Skarpsno ES, Kongsvold A, Bach K, and Mork PJ
- Subjects
actigraphy ,epidemiology ,sedentary behaviors ,sleep quality ,supervised machine learning ,support vector machines ,Psychiatry ,RC435-571 ,Neurophysiology and neuropsychology ,QP351-495 - Abstract
Aleksej Logacjov,1 Eivind Schjelderup Skarpsno,2,3 Atle Kongsvold,2 Kerstin Bach,1 Paul Jarle Mork2 1Department of Computer Science, Norwegian University of Science and Technology, Trondheim, Norway; 2Department of Public Health and Nursing, Norwegian University of Science and Technology, Trondheim, Norway; 3Department of Neurology and Clinical Neurophysiology, St. Olavs Hospital, Trondheim, NorwayCorrespondence: Aleksej Logacjov, Department of Computer Science, Faculty of Information Technology and Electrical Engineering, Norwegian University of Science and Technology, Trondheim, 7034, Norway, Email aleksej.logacjov@ntnu.noPurpose: Body-worn accelerometers are commonly used to estimate sleep duration in population-based studies. However, since accelerometry-based sleep/wake-scoring relies on detecting body movements, the prediction of sleep duration remains a challenge. The aim was to develop and evaluate the performance of a machine learning (ML) model to predict accelerometry-based sleep duration and to explore if this prediction can be improved by adding skin temperature data, circadian rhythm based on the estimated midpoint of sleep, and cyclic time features to the model.Patients and Methods: Twenty-nine adults (17 females), mean (SD) age 40.2 (15.0) years (range 17– 70) participated in the study. Overnight polysomnography (PSG) was recorded in a sleep laboratory or at home along with body movement by two accelerometers with an embedded skin temperature sensor (AX3, Axivity, UK) positioned at the low back and thigh. The PSG scoring of sleep/wake was used as ground truth for training the ML model.Results: Based on pure accelerometer data input to the ML model, the specificity and sensitivity for predicting sleep/wake was 0.52 (SD 0.24) and 0.95 (SD 0.03), respectively. Adding skin temperature data and contextual information to the ML model improved the specificity to 0.72 (SD 0.20), while sensitivity remained unchanged at 0.95 (SD 0.05). Correspondingly, sleep overestimation was reduced from 54 min (228 min, limits of agreement range [LoAR]) to 19 min (154 min LoAR).Conclusion: An ML model can predict sleep/wake periods with excellent sensitivity and moderate specificity based on a dual-accelerometer set-up when adding skin temperature data and contextual information to the model.Keywords: actigraphy, epidemiology, sedentary behaviors, sleep quality, supervised machine learning, support vector machines
- Published
- 2024
30. Comparison of the Accuracy of Ground Reaction Force Component Estimation between Supervised Machine Learning and Deep Learning Methods Using Pressure Insoles.
- Author
-
Kammoun, Amal, Ravier, Philippe, and Buttelli, Olivier
- Subjects
- *
ARTIFICIAL neural networks , *SUPERVISED learning , *CONVOLUTIONAL neural networks , *GROUND reaction forces (Biomechanics) , *DEEP learning - Abstract
The three Ground Reaction Force (GRF) components can be estimated using pressure insole sensors. In this paper, we compare the accuracy of estimating GRF components for both feet using six methods: three Deep Learning (DL) methods (Artificial Neural Network, Long Short-Term Memory, and Convolutional Neural Network) and three Supervised Machine Learning (SML) methods (Least Squares, Support Vector Regression, and Random Forest (RF)). Data were collected from nine subjects across six activities: normal and slow walking, static with and without carrying a load, and two Manual Material Handling activities. This study has two main contributions: first, the estimation of GRF components (Fx, Fy, and Fz) during the six activities, two of which have never been studied; second, the comparison of the accuracy of GRF component estimation between the six methods for each activity. RF provided the most accurate estimation for static situations, with mean RMSE values of RMSE_Fx = 1.65 N, RMSE_Fy = 1.35 N, and RMSE_Fz = 7.97 N for the mean absolute values measured by the force plate (reference) RMSE_Fx = 14.10 N, RMSE_Fy = 3.83 N, and RMSE_Fz = 397.45 N. In our study, we found that RF, an SML method, surpassed the experimented DL methods. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
31. The stochastic multi-gradient algorithm for multi-objective optimization and its application to supervised machine learning.
- Author
-
Liu, S. and Vicente, L. N.
- Subjects
- *
SUPERVISED learning , *APPROXIMATION algorithms , *DECISION making , *ALGORITHMS , *A priori - Abstract
Optimization of conflicting functions is of paramount importance in decision making, and real world applications frequently involve data that is uncertain or unknown, resulting in multi-objective optimization (MOO) problems of stochastic type. We study the stochastic multi-gradient (SMG) method, seen as an extension of the classical stochastic gradient method for single-objective optimization. At each iteration of the SMG method, a stochastic multi-gradient direction is calculated by solving a quadratic subproblem, and it is shown that this direction is biased even when all individual gradient estimators are unbiased. We establish rates to compute a point in the Pareto front, of order similar to what is known for stochastic gradient in both convex and strongly convex cases. The analysis handles the bias in the multi-gradient and the unknown a priori weights of the limiting Pareto point. The SMG method is framed into a Pareto-front type algorithm for calculating an approximation of the entire Pareto front. The Pareto-front SMG algorithm is capable of robustly determining Pareto fronts for a number of synthetic test problems. One can apply it to any stochastic MOO problem arising from supervised machine learning, and we report results for logistic binary classification where multiple objectives correspond to distinct-sources data groups. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
32. Quantitative Monitoring Method for Conveyor Belt Deviation Status Based on Attention Guidance.
- Author
-
Zhang, Xi, Yang, Zihao, Zhang, Mengchao, Yu, Yan, Zhou, Manshan, and Zhang, Yuan
- Subjects
CONVEYOR belts ,BELT conveyors ,QUANTITATIVE research ,BELTS (Clothing) ,FORECASTING - Abstract
The efficient monitoring of the belt deviation state will help to reduce unnecessary abnormal wear and the risk of belt tear. This paper proposes a coupling characterization method involving the prediction box features of the target detection network and the linear features of the conveyor belt edge to achieve the quantitative monitoring of conveyor belt deviations. The impacts of the type, location, and number of attention mechanisms on the detection effect are fully discussed. Compared with traditional image-processing-based methods, the proposed method is more efficient, eliminating the tedious process of threshold setting and improving the detection efficiency. In detail, the improved practice and tests are carried out based on the Yolov5 network, and the Grad-CAM technique is also used to explore the effect of attention mechanisms in improving the detection accuracy. The experiments show that the detection accuracy of the proposed method can reach 99%, with a detection speed of 67.7 FPS on a self-made dataset. It is also proven to have a good anti-interference ability and can effectively resist the influence of the conveying material flow, lighting conditions, and other factors on the detection accuracy. This research is of great significance in improving the intelligent operation and maintenance level of belt conveyors and ensuring their safe operation. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
33. Middle ear-acquired cholesteatoma diagnosis based on CT scan image mining using supervised machine learning models.
- Author
-
Ouattassi, Naouar, Maaroufi, Mustapha, Slaoui, Hajar, Benateya Andaloussi, Taha, Zarghili, Arsalane, and El Amine El Alami, Mohamed Nouredine
- Subjects
MACHINE learning ,SUPERVISED learning ,CONVOLUTIONAL neural networks ,RECEIVER operating characteristic curves ,K-nearest neighbor classification ,MIDDLE ear - Abstract
Background: Distinguishing between middle ear cholesteatoma and chronic suppurative otitis media (CSOM) is an ongoing challenge. While temporal bone computed tomography (CT) scan is highly accurate for diagnosing middle ear conditions, its specificity in discerning between cholesteatoma and CSOM is only moderate. To address this issue, we utilized trained machine learning models to enhance the specificity of temporal bone CT scan in diagnosing middle ear cholesteatoma. Our database consisted of temporal bone CT scan native images from 122 patients diagnosed with middle ear cholesteatoma and a control group of 115 patients diagnosed with CSOM, with both groups labeled based on surgical findings. We preprocessed the native images to isolate the region of interest and then utilized the Inception V3 convolutional neural network for image embedding into data vectors. Classification was performed using machine learning models including support vector machine (SVM), k-nearest neighbors (k-NN), random forest, and neural network. Statistical metrics employed to interpret the results included classification accuracy, precision, recall, F1 score, confusion matrix, area under the receiver operating characteristic curve (AUC), and FreeViz diagram. Results: Our training dataset comprised 5390 images, and the testing dataset included 125 different images. The neural network, k-NN, and SVM models demonstrated significantly higher relevance in terms of classification accuracy, precision, and recall compared to the random forest model. For instance, the F1 scores were 0.974, 0.987, and 0.897, respectively, for the former three models, in contrast to 0.661 for the random forest model. Conclusion: The performance metrics of the presented trained machine learning models hold promising prospects as potentially clinically useful aids. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
34. Applying data science approach to predicting diseases and recommending drugs in healthcare using machine learning models – A cardio disease case study.
- Author
-
Lambay, Muhib Anwar and Mohideen, S. Pakkir
- Subjects
CLINICAL decision support systems ,SUPERVISED learning ,DATA science ,SUPPORT vector machines ,FEATURE selection ,MACHINE learning - Abstract
Cardiovascular diseases are causing more deaths across the globe. With innovations in Artificial Intelligence (AI) predicting such diseases early is very important research area. With learning based approaches that exploit knowledge from given samples, it is possible to improve disease prediction process. There are many aspects to proper healthcare such as preventing diseases with suitable diet and lifestyle, early detection of diseases if any and efficient treatment. Data is being accumulated in every domain. However, the healthcare industry is on top of the list as it provides large volumes of data pertaining to human health, diet and drug aspects. The existing literature has not shown adequate research in this direction. The Healthcare industry has an unprecedented impact on the well-being of people across the globe. In the recent observations by World Health Organization (WHO), data science approach towards disease prediction greatly complements existing Clinical Decision Support Systems (CDSSs).This research paper presents a comprehensive study on the application of data science techniques for disease prediction and drug recommendation in healthcare, focusing on a case study involving cardiovascular diseases. The primary objective of this study is to develop a robust predictive model that identifies the likelihood of cardiovascular diseases in patients, and subsequently recommends drug interventions for optimal treatment outcomes. Here we propose Disease Prediction and Drug Recommendation Framework (DPDRF). The framework is realized by defining an algorithm known as Cardio Disease Prediction and Drug Recommendation (CDP-DR). The Disease Prediction and Drug Recommendation algorithm in turn uses different supervised machine learning (ML) algorithms such as Random Forest (RF), Logistic Regression (LR), Support Vector Machine (SVM), Decision Tree (DT), Stochastic Gradient Descent (SGD), Gradient Boosting, and Extreme Gradient Boosting (XGB). Another algorithm known as Entropy and Gain based Hybrid Feature Selection (EG-HFS) is defined to leverage quality of training leading to performance enhancement of prediction models. The experimental results with cardio disease prediction as a case study revealed that the proposed framework is useful in disease prediction and drug recommendations by using different prediction models. Highest accuracy achieved by the proposed system is 96.23%. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
35. Molecular dynamics simulation study of post‐transition state bifurcation: A case study on the ambimodal transition state of dipolar/Diels–Alder cycloaddition.
- Author
-
Murakami, Tatsuhiro, Kikuma, Yuya, Hayashi, Daiki, Ibuki, Shunichi, Nakagawa, Shoto, Ueno, Hinami, and Takayanagi, Toshiyuki
- Subjects
- *
QUASI-classical trajectory method , *MOLECULAR dynamics , *POTENTIAL energy surfaces , *RING formation (Chemistry) , *GAS phase reactions , *TRANSITION state theory (Chemistry) , *SURFACE reactions , *SOLVATION - Abstract
The potential energy surfaces for the reactions of 1,3‐butadiene with 2‐hydroxythioacrolein and 2‐aminoacrolein exhibit ambimodal transition states leading to both dipolar (4 + 3) and Diels–Alder (4 + 2) cycloaddition products, thereby demonstrating a post transition state bifurcation feature. We have investigated the bifurcation dynamics of these reactions using three molecular dynamics (MD) methods: quasi‐classical trajectory, classical MD, and ring‐polymer MD simulations. The trajectory calculations were performed with the semiempirical GFN2‐xTB method with the element‐specific parameters optimized to reproduce the density‐functional theory calculations. The effect of water solvation was examined using an implicit solvation model, revealing significant differences in bifurcation dynamic between gas‐phase and solution‐phase reactions. Nuclear quantum effects were found to play a crucial role in the proton‐transfer process from the (4 + 3) intermediate to the (4 + 3) product in the case of the 2‐aminoacrolein reaction. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
36. The Augmented Social Scientist: Using Sequential Transfer Learning to Annotate Millions of Texts with Human-Level Accuracy.
- Author
-
Do, Salomé, Ollion, Étienne, and Shen, Rubing
- Subjects
- *
SUPERVISED learning , *LANGUAGE transfer (Language learning) , *NATURAL language processing , *SEQUENTIAL learning , *RESEARCH questions - Abstract
The last decade witnessed a spectacular rise in the volume of available textual data. With this new abundance came the question of how to analyze it. In the social sciences, scholars mostly resorted to two well-established approaches, human annotation on sampled data on the one hand (either performed by the researcher, or outsourced to microworkers), and quantitative methods on the other. Each approach has its own merits - a potentially very fine-grained analysis for the former, a very scalable one for the latter - but the combination of these two properties has not yielded highly accurate results so far. Leveraging recent advances in sequential transfer learning, we demonstrate via an experiment that an expert can train a precise, efficient automatic classifier in a very limited amount of time. We also show that, under certain conditions, expert-trained models produce better annotations than humans themselves. We demonstrate these points using a classic research question in the sociology of journalism, the rise of a "horse race" coverage of politics. We conclude that recent advances in transfer learning help us augment ourselves when analyzing unstructured data. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
37. Assessing and predicting green gentrification susceptibility using an integrated machine learning approach.
- Author
-
Assaad, Rayan H. and Jezzini, Yasser
- Subjects
- *
ENVIRONMENTAL gentrification , *GENTRIFICATION , *ARTIFICIAL neural networks , *MACHINE learning , *GREEN infrastructure , *K-means clustering - Abstract
Greenery initiatives, such as green infrastructures (GIs), create sustainable and climate-resilient environments. However, they can also have unintended consequences, such as displacement and gentrification in low-income areas. This paper proposes an integrated machine learning (ML) approach that combines both unsupervised and supervised ML algorithms. First, 35 indicators that contribute to green gentrification were identified and categorised into 7 categories: social, economic, demographic, housing, household, amenities, and GIs. Second, data was collected for all census tracts in New York City. Third, the green gentrification susceptibility was modelled into 6 levels using k-means clustering analysis, which is an unsupervised ML model. Fourth, the Technique for Order of Preference by Similarity to the Ideal Solution (TOPSIS) was used to map the census tracts to their green gentrification susceptibility level. Finally, different supervised ML algorithms were trained and tested to predict the green gentrification susceptibility. The results showed that the artificial neural network (ANN) model is the most accurate in classifying and predicting the green gentrification susceptibility with an overall accuracy of 96%. Moreover, the outcomes showed that the Normal Difference Vegetation Index (NDVI), the proximity to GIs, the GIs frequency, and the total area of GIs were identified as the most important indicators to predict green gentrification susceptibility. Ultimately, the proposed approach allows practitioners and researchers to perform micro-level (i.e. on the census-tracts level) predictions and inferences about green gentrification susceptibility. This allows more focused and targeted mitigation actions to be designed and implemented in the most affected communities, thus promoting environmental justice. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
38. Holistic Review of UAV-Centric Situational Awareness: Applications, Limitations, and Algorithmic Challenges.
- Author
-
MahmoudZadeh, Somaiyeh, Yazdani, Amirmehdi, Kalantari, Yashar, Ciftler, Bekir, Aidarus, Fathi, and Al Kadri, Mhd Omar
- Subjects
SUPERVISED learning ,ARTIFICIAL intelligence ,REAL-time computing ,SITUATIONAL awareness ,DETERMINISTIC algorithms - Abstract
This paper presents a comprehensive survey of UAV-centric situational awareness (SA), delineating its applications, limitations, and underlying algorithmic challenges. It highlights the pivotal role of advanced algorithmic and strategic insights, including sensor integration, robust communication frameworks, and sophisticated data processing methodologies. The paper critically analyzes multifaceted challenges such as real-time data processing demands, adaptability in dynamic environments, and complexities introduced by advanced AI and machine learning techniques. Key contributions include a detailed exploration of UAV-centric SA's transformative potential in industries such as precision agriculture, disaster management, and urban infrastructure monitoring, supported by case studies. In addition, the paper delves into algorithmic approaches for path planning and control, as well as strategies for multi-agent cooperative SA, addressing their respective challenges and future directions. Moreover, this paper discusses forthcoming technological advancements, such as energy-efficient AI solutions, aimed at overcoming current limitations. This holistic review provides valuable insights into the UAV-centric SA, establishing a foundation for future research and practical applications in this domain. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
39. Fine-Tuning Cyber Security Defenses: Evaluating Supervised Machine Learning Classifiers for Windows Malware Detection.
- Author
-
Zada, Islam, Alatawi, Mohammed Naif, Saqlain, Syed Muhammad, Alshahrani, Abdullah, Alshamran, Adel, Imran, Kanwal, and Alfraihi, Hessa
- Subjects
SUPERVISED learning ,CYBERTERRORISM ,EVIDENCE gaps ,REQUIREMENTS engineering ,DECISION trees - Abstract
Malware attacks on Windows machines pose significant cybersecurity threats, necessitating effective detection and prevention mechanisms. Supervised machine learning classifiers have emerged as promising tools for malware detection. However, there remains a need for comprehensive studies that compare the performance of different classifiers specifically for Windows malware detection. Addressing this gap can provide valuable insights for enhancing cybersecurity strategies. While numerous studies have explored malware detection using machine learning techniques, there is a lack of systematic comparison of supervised classifiers for Windows malware detection. Understanding the relative effectiveness of these classifiers can inform the selection of optimal detection methods and improve overall security measures. This study aims to bridge the research gap by conducting a comparative analysis of supervised machine learning classifiers for detecting malware on Windows systems. The objectives include Investigating the performance of various classifiers, such as Gaussian Naïve Bayes, K Nearest Neighbors (KNN), Stochastic Gradient Descent Classifier (SGDC), and Decision Tree, in detecting Windows malware. Evaluating the accuracy, efficiency, and suitability of each classifier for real-world malware detection scenarios. Identifying the strengths and limitations of different classifiers to provide insights for cybersecurity practitioners and researchers. Offering recommendations for selecting the most effective classifier for Windows malware detection based on empirical evidence. The study employs a structured methodology consisting of several phases: exploratory data analysis, data preprocessing, model training, and evaluation. Exploratory data analysis involves understanding the dataset's characteristics and identifying preprocessing requirements. Data preprocessing includes cleaning, feature encoding, dimensionality reduction, and optimization to prepare the data for training. Model training utilizes various supervised classifiers, and their performance is evaluated using metrics such as accuracy, precision, recall, and F1 score. The study's outcomes comprise a comparative analysis of supervised machine learning classifiers for Windows malware detection. Results reveal the effectiveness and efficiency of each classifier in detecting different types of malware. Additionally, insights into their strengths and limitations provide practical guidance for enhancing cybersecurity defenses. Overall, this research contributes to advancing malware detection techniques and bolstering the security posture of Windows systems against evolving cyber threats. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
40. Supervised Machine Learning for Matchmaking in Digital Business Ecosystems and Platforms.
- Author
-
Benramdane, Mustapha Kamal, Kornyshova, Elena, Bouzefrane, Samia, and Maupas, Hubert
- Subjects
SUPERVISED learning ,MACHINE learning ,DIGITAL technology ,RECOMMENDER systems ,BUSINESS ecosystems - Abstract
In the digital era, organizations belonging to the same or different market segments come together in digital platforms that allow them to exchange. These organizations are unified within a Digital Business Ecosystem. However, the rapid growth of the number of these organizations accentuates the complexity of finding economic partners, customers, suppliers, or other organizations that can share economic interests. In our research, we propose a recommendation system that is implemented on such a digital platform, and which is based on matchmaking and hybrid supervised machine learning algorithms. In this paper, we provide a detailed analysis of the functioning of this system, the challenge encountered when processing the data which made it possible to highlight the similarities between the organizations that can be associated. Thus, we seek to improve the understanding and analysis of the data for the identification of partners in an optimal way. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
41. Machine learning allows robust classification of visceral fat in women with obesity using common laboratory metrics.
- Author
-
Palmieri, Flavio, Akhtar, Nidà Farooq, Pané, Adriana, Jiménez, Amanda, Olbeyra, Romina Paula, Viaplana, Judith, Vidal, Josep, de Hollanda, Ana, Gama-Perez, Pau, Jiménez-Chillarón, Josep C., and Garcia-Roves, Pablo M.
- Subjects
- *
SUPERVISED learning , *OVERWEIGHT women , *MATHEMATICAL formulas , *RECEIVER operating characteristic curves , *SUPPORT vector machines , *FAT , *MACHINE learning , *OBESITY in women - Abstract
The excessive accumulation and malfunctioning of visceral adipose tissue (VAT) is a major determinant of increased risk of obesity-related comorbidities. Thus, risk stratification of people living with obesity according to their amount of VAT is of clinical interest. Currently, the most common VAT measurement methods include mathematical formulae based on anthropometric dimensions, often biased by human measurement errors, bio-impedance, and image techniques such as X-ray absorptiometry (DXA) analysis, which requires specialized equipment. However, previous studies showed the possibility of classifying people living with obesity according to their VAT through blood chemical concentrations by applying machine learning techniques. In addition, most of the efforts were spent on men living with obesity while little was done for women. Therefore, this study aims to compare the performance of the multilinear regression model (MLR) in estimating VAT and six different supervised machine learning classifiers, including logistic regression (LR), support vector machine and decision tree-based models, to categorize 149 women living with obesity. For clustering, the study population was categorized into classes 0, 1, and 2 according to their VAT and the accuracy of each MLR and classification model was evaluated using DXA-data (DXAdata), blood chemical concentrations (BLDdata), and both DXAdata and BLDdata together (ALLdata). Estimation error and R 2 were computed for MLR, while receiver operating characteristic (ROC) and precision-recall curves (PR) area under the curve (AUC) were used to assess the performance of every classification model. MLR models showed a poor ability to estimate VAT with mean absolute error ≥ 401.40 and R 2 ≤ 0.62 in all the datasets. The highest accuracy was found for LR with values of 0.57, 0.63, and 0.53 for ALLdata, DXAdata, and BLDdata, respectively. The ROC AUC showed a poor ability of both ALLdata and DXAdata to distinguish class 1 from classes 0 and 2 (AUC = 0.31, 0.71, and 0.85, respectively) as also confirmed by PR (AUC = 0.24, 0.57, and 0.73, respectively). However, improved performances were obtained when applying LR model to BLDdata (ROC AUC ≥ 0.61 and PR AUC ≥ 0.42), especially for class 1. These results seem to suggest that, while a direct and reliable estimation of VAT was not possible in our cohort, blood sample-derived information can robustly classify women living with obesity by machine learning-based classifiers, a fact that could benefit the clinical practice, especially in those health centres where medical imaging devices are not available. Nonetheless, these promising findings should be further validated over a larger population. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
42. Mathematical modeling and machine learning-based optimization for enhancing biofiltration efficiency of volatile organic compounds.
- Author
-
Sulaiman, Muhammad, Khalaf, Osamah Ibrahim, Khan, Naveed Ahmad, Alshammari, Fahad Sameer, and Hamam, Habib
- Subjects
- *
BIOFILTRATION , *SUPERVISED learning , *MACHINE learning , *BURGERS' equation , *MATHEMATICAL models , *VOLATILE organic compounds - Abstract
Biofiltration is a method of pollution management that utilizes a bioreactor containing live material to absorb and destroy pollutants biologically. In this paper, we investigate mathematical models of biofiltration for mixing volatile organic compounds (VOCs) for instance hydrophilic (methanol) and hydrophobic (α -pinene). The system of nonlinear diffusion equations describes the Michaelis-Menten kinetics of the enzymic chemical reaction. These models represent the chemical oxidation in the gas phase and mass transmission within the air-biofilm junction. Furthermore, for the numerical study of the saturation of α -pinene and methanol in the biofilm and gas state, we have developed an efficient supervised machine learning algorithm based on the architecture of Elman neural networks (ENN). Moreover, the Levenberg-Marquardt (LM) optimization paradigm is used to find the parameters/ neurons involved in the ENN architecture. The approximation to a solutions found by the ENN-LM technique for methanol saturation and α -pinene under variations in different physical parameters are allegorized with the numerical results computed by state-of-the-art techniques. The graphical and statistical illustration of indications of performance relative to the terms of absolute errors, mean absolute deviations, computational complexity, and mean square error validates that our results perfectly describe the real-life situation and can further be used for problems arising in chemical engineering. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
43. Prediction of precancerous cervical cancer lesions among women living with HIV on antiretroviral therapy in Uganda: a comparison of supervised machine learning algorithms.
- Author
-
Namalinzi, Florence, Galadima, Kefas Rimamnuskeb, Nalwanga, Robinah, Sekitoleko, Isaac, and Uwimbabazi, Leon Fidele Ruganzu
- Subjects
- *
SUPERVISED learning , *HIV-positive women , *MACHINE learning , *CERVICAL cancer , *ANTIRETROVIRAL agents - Abstract
Background: Cervical cancer (CC) is among the most prevalent cancer types among women with the highest prevalence in low- and middle-income countries (LMICs). It is a curable disease if detected early. Machine learning (ML) techniques can aid in early detection and prediction thus reducing screening and treatment costs. This study focused on women living with HIV (WLHIV) in Uganda. Its aim was to identify the best predictors of CC and the supervised ML model that best predicts CC among WLHIV. Methods: Secondary data that included 3025 women from three health facilities in central Uganda was used. A multivariate binary logistic regression and recursive feature elimination with random forest (RFERF) were used to identify the best predictors. Five models; logistic regression (LR), random forest (RF), K-Nearest neighbor (KNN), support vector machine (SVM), and multi-layer perceptron (MLP) were applied to identify the out-performer. The confusion matrix and the area under the receiver operating characteristic curve (AUC/ROC) were used to evaluate the models. Results: The results revealed that duration on antiretroviral therapy (ART), WHO clinical stage, TPT status, Viral load status, and family planning were commonly selected by the two techniques and thus highly significant in CC prediction. The RF from the RFERF-selected features outperformed other models with the highest scores of 90% accuracy and 0.901 AUC. Conclusion: Early identification of CC and knowledge of the risk factors could help control the disease. The RF outperformed other models applied regardless of the selection technique used. Future research can be expanded to include ART-naïve women in predicting CC. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
44. Low-cost and scalable detection of sparse informal settlements using machine learning in Gcuwa, Eastern Cape, South Africa.
- Author
-
Chamunorwa, Brighton, Shoko, Moreblessings, and Magidi, James
- Subjects
- *
RANDOM forest algorithms , *SUPPORT vector machines , *DATA libraries , *BOOSTING algorithms , *K-nearest neighbor classification , *ENVIRONMENTAL degradation , *RADIOACTIVE waste repositories , *MACHINE learning - Abstract
Informative and scalable cartography plays a pivotal role in curbing urban pollution, waste management, and mitigating environmental damage in the development of informal settlements. The contemporary capabilities of cloud computing facilitate streamlined access to comprehensive data repositories, computational infrastructure, and proficient tools that have rapidly advanced the execution of sprawl mapping procedures. This study tests the performance of four machine-learning algorithms, namely: Gradient Boost, K Nearest Neighbor [KNN], Random Forest [RF], and Support Vector Machine [SVM] with data extracted from cloud computing repositories for delineating informal settlements in Gcuwa, Eastern Cape, South Africa, using low-cost datasets. A systematic approach comprising iterative phases, encompassing data acquisition, the development of a training dataset, modeling, and evaluation was employed. The delineation process involved the extraction of both spectral and textural features from Sentinel-2 imagery. The Random Forest algorithm emerged as the top performer, exhibiting the highest levels of accuracy and F1 score, followed by the gradient boosting, support vector machine, and then the K-nearest neighbor algorithms. Consequently, this innovative use of machine learning algorithms with low-cost datasets and the scalable, resilient approach for detecting informal settlements offers a promising avenue for enhancing urban planning and addressing sustainable development challenges. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
45. Structural load estimation of the wheel loader for customer usage profile monitoring.
- Author
-
Cho, Jae-Hong, Na, Seon-Jun, Kim, Min-Seok, and Park, Myeong-Kwan
- Subjects
- *
SUPERVISED learning , *MACHINE learning , *CONSUMERS , *STRUCTURAL health monitoring , *WHEELS - Abstract
This paper aims to estimate the structural load applied to the attachment and frame of a wheel loader by using pressure and acceleration, which have a load correlation, instead of strain. First, we propose a data-driven modeling methodology for load estimation. Load features as model input are derived based on cylinder pressure and frame acceleration. The relationship between features and the load is expressed through correlation coefficients and generalized by using the supervised machine learning technique. Next, to experimentally collect the large data set for learning, two wheel loaders of different classes are selected to build sensor cars. Then, the measurements on both wheel loaders are performed repeatedly with four experienced operators according to the test cases. Using conventional learning algorithms, model selection is performed for five parts that make up the attachment and frame. Then, the model learning and evaluation are performed. As a result of an out-of-sample test, the average estimation error for each part is approximately 5 % and the proposed methodology is experimentally verified that it is effective in estimating load for CUP monitoring. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
46. Geospatial modeling of climate change indices at Mexico City using machine learning regression.
- Author
-
Saldana-Perez, Magdalena, Guzmán, Giovanni, Palma-Preciado, Carolina, Argüelles-Cruz, Amadeo, and Moreno-Ibarra, Marco
- Abstract
Purpose: Climate change is a problem that concerns all of us. Despite the information produced by organizations such as the Expert Team on Climate Change Detection and Indices and the United Nations, only a few cities have been planned taking into account the climate changes indices. This paper aims to study climatic variations, how climate conditions might change in the future and how these changes will affect the activities and living conditions in cities, specifically focusing on Mexico city. Design/methodology/approach: In this approach, two distinct machine learning regression models, k-Nearest Neighbors and Support Vector Regression, were used to predict variations in climate change indices within select urban areas of Mexico city. The calculated indices are based on maximum, minimum and average temperature data collected from the National Water Commission in Mexico and the Scientific Research Center of Ensenada. The methodology involves pre-processing temperature data to create a training data set for regression algorithms. It then computes predictions for each temperature parameter and ultimately assesses the performance of these algorithms based on precision metrics scores. Findings: This paper combines a geospatial perspective with computational tools and machine learning algorithms. Among the two regression algorithms used, it was observed that k-Nearest Neighbors produced superior results, achieving an R
2 score of 0.99, in contrast to Support Vector Regression, which yielded an R2 score of 0.74. Originality/value: The full potential of machine learning algorithms has not been fully harnessed for predicting climate indices. This paper also identifies the strengths and weaknesses of each algorithm and how the generated estimations can then be considered in the decision-making process. [ABSTRACT FROM AUTHOR]- Published
- 2024
- Full Text
- View/download PDF
47. Deep-learning-based multistate monitoring method of belt conveyor turning section.
- Author
-
Zhang, Mengchao, Jiang, Kai, Zhao, Shuai, Hao, Nini, and Zhang, Yuan
- Subjects
CONVEYOR belts ,BELT conveyors ,COMORBIDITY ,SUSTAINABLE transportation ,MAINTENANCE costs - Abstract
During transportation, bulk materials are susceptible to spillage due to equipment instability and environmental factors, resulting in increased maintenance costs and environmental pollution. Thus, intelligent and efficient condition monitoring is crucial for maintaining operational efficiency of transfer equipment. It facilitates the timely identification of potential safety hazards, preventing accidents from occurring or their impact from spreading, thereby minimizing production and maintenance costs. This study presents a deep-learning-based multioperation synchronous monitoring method suitable for belt conveyors that integrate target segmentation and detection networks to simultaneously diagnose belt deviation, measure conveying load, identify idlers, and do other tasks on a self-made dataset. This method effectively reduces the complexity of multistate simultaneous monitoring and monitoring costs, thereby avoiding environmental pollution caused by transportation accidents. Experimental results show that the segmentation accuracy of the proposed method can be up to 88.72%, with a detection accuracy of 91.3% and an overall inference speed of 90.9 frames per second. Furthermore, by extending the dataset, the proposed method can incorporate additional tasks, such as belt damage, scattered material, and foreign object identifications. This study has practical significance in ensuring the normal and eco-friendly operation of bulk material transportation. Our source dataset is available at https://github.com/zhangzhangzhang1618/dataset-for-turnning-section [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
48. A Machine Learning Approach to Cyberbullying Detection in Arabic Tweets.
- Author
-
Musleh, Dhiaa, Rahman, Atta, Alkherallah, Mohammed Abbas, Al-Bohassan, Menhal Kamel, Alawami, Mustafa Mohammed, Alsebaa, Hayder Ali, Alnemer, Jawad Ali, Al-Mutairi, Ghazi Fayez, Aldossary, May Issa, Aldowaihi, Dalal A., and Alhaidari, Fahd
- Subjects
SUPERVISED learning ,NATURAL language processing ,MACHINE learning ,LITERATURE reviews ,SUPPORT vector machines - Abstract
With the rapid growth of internet usage, a new situation has been created that enables practicing bullying. Cyberbullying has increased over the past decade, and it has the same adverse effects as face-to-face bullying, like anger, sadness, anxiety, and fear. With the anonymity people get on the internet, they tend to be more aggressive and express their emotions freely without considering the effects, which can be a reason for the increase in cyberbullying and it is the main motive behind the current study. This study presents a thorough background of cyberbullying and the techniques used to collect, preprocess, and analyze the datasets. Moreover, a comprehensive review of the literature has been conducted to figure out research gaps and effective techniques and practices in cyberbullying detection in various languages, and it was deduced that there is significant room for improvement in the Arabic language. As a result, the current study focuses on the investigation of shortlisted machine learning algorithms in natural language processing (NLP) for the classification of Arabic datasets duly collected from Twitter (also known as X). In this regard, support vector machine (SVM), Naïve Bayes (NB), Random Forest (RF), Logistic regression (LR), Bootstrap aggregating (Bagging), Gradient Boosting (GBoost), Light Gradient Boosting Machine (LightGBM), Adaptive Boosting (AdaBoost), and eXtreme Gradient Boosting (XGBoost) were shortlisted and investigated due to their effectiveness in the similar problems. Finally, the scheme was evaluated by well-known performance measures like accuracy, precision, Recall, and F1-score. Consequently, XGBoost exhibited the best performance with 89.95% accuracy, which is promising compared to the state-of-the-art. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
49. Secured Frank Wolfe learning and Dirichlet Gaussian Vicinity based authentication for IoT edge computing.
- Author
-
Sangeethapriya, J., Arock, Michael, and Reddy, U. Srinivasulu
- Subjects
EDGE computing ,SUPERVISED learning ,TIME complexity ,INTERNET of things ,COMPUTER systems ,ERROR rates ,LOGISTIC functions (Mathematics) - Abstract
With the evolution of the Internet of Things (IoT) several users take part in different applications via sensors. The foremost confront here remains in selecting the most confidential users or sensors in the edge computing system of the IoT. Here, both the end-users and the edge servers are likely to be malicious or compromised sensors. Several works have been contributed to identifying and isolating the malicious end-users or edge servers. Our work concentrates on the security aspects of edge servers of IoT. The Frank-Wolfe Optimal Service Requests (FWOSR) algorithm is utilized to evaluate the boundaries or limits of the logistic regression model, in which the convex problem under a linear approximation is solved for weight sparsity (i.e. several user requests competing for closest edge server) to avoid over-fitting in the supervised machine learning process. We design a Frank Wolfe Supervised Machine Learning (FWSL) technique to choose an optimal edge server and further minimize the computational and communication costs between the user requests and the edge server. Next, Dirichlet Gaussian Blocked Gibbs Vicinity-based Authentication model for location-based services in Cloud networks is proposed. Here, the vicinity-based authentication is implemented based on Received Signal Strength Indicators (RSSI), MAC address and packet arrival time. With this, the authentication accuracy is improved by introducing the Gaussian function in the vicinity test and provides flexible vicinity range control by taking into account multiple locations. Simulation and experiment are also conducted to validate the computational cost, communication cost, time complexity and detection error rate. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
50. Automated quantification of photoreceptor outer segments in developing and degenerating retinas on microscopy images across scales.
- Author
-
Seidemann, Suse, Salomon, Florian, Hoffmann, Karl B., Kurth, Thomas, Sbalzarini, Ivo F., Haase, Robert, and Ader, Marius
- Subjects
PHOTORECEPTORS ,RETINA ,SUPERVISED learning ,MICROSCOPY ,RETINAL degeneration ,IMAGE segmentation ,RETINAL imaging - Abstract
The functionality of photoreceptors, rods, and cones is highly dependent on their outer segments (POS), a cellular compartment containing highly organized membranous structures that generate biochemical signals from incident light. While POS formation and degeneration are qualitatively assessed on microscopy images, reliable methodology for quantitative analyses is still limited. Here, we developed methods to quantify POS (QuaPOS) maturation and quality on retinal sections using automated image analyses. POS formation was examined during the development and in adulthood of wild-type mice via light microscopy (LM) and transmission electron microscopy (TEM). To quantify the number, size, shape, and fluorescence intensity of POS, retinal cryosections were immunostained for the cone POS marker S-opsin. Fluorescence images were used to train the robust classifier QuaPOS-LM based on supervised machine learning for automated image segmentation. Characteristic features of segmentation results were extracted to quantify the maturation of cone POS. Subsequently, this quantification method was applied to characterize POS degeneration in "cone photoreceptor function loss 1" mice. TEM images were used to establish the ultrastructural quantification method QuaPOS-TEM for the alignment of POS membranes. Images were analyzed using a customwritten MATLAB code to extract the orientation of membranes from the image gradient and their alignment (coherency). This analysis was used to quantify the POS morphology of wild-type and two inherited retinal degeneration ("retinal degeneration 19" and "rhodopsin knock-out") mouse lines. Both automated analysis technologies provided robust characterization and quantification of POS based on LM or TEM images. Automated image segmentation by the classifier QuaPOS-LM and analysis of the orientation of membrane stacks by QuaPOS-TEM using fluorescent or TEM images allowed quantitative evaluation of POS formation and quality. The assessments showed an increase in POS number, volume, and membrane coherency during wild-type postnatal development, while a decrease in all three observables was detected in different retinal degeneration mouse models. All the code used for the presented analysis is open source, including example datasets to reproduce the findings. Hence, the QuaPOS quantification methods are useful for in-depth characterization of POS on retinal sections in developmental studies, for disease modeling, or after therapeutic interventions affecting photoreceptors. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
Catalog
Discovery Service for Jio Institute Digital Library
For full access to our library's resources, please sign in.