898 results on '"Data-driven models"'
Search Results
2. Modeling of Digital Twins
- Author
-
Gupta, Sunil, Iyer, Ravi S., Kumar, Sanjeev, Gupta, Sunil, Iyer, Ravi S., and Kumar, Sanjeev
- Published
- 2025
- Full Text
- View/download PDF
3. A framework based on physics-informed graph neural ODE: for continuous spatial-temporal pandemic prediction.
- Author
-
Cheng, Haodong, Mao, Yingchi, and Jia, Xiao
- Subjects
ORDINARY differential equations ,PARTIAL differential equations ,INVERSE problems ,TIME series analysis ,PREDICTION models - Abstract
Physics-informed spatial-temporal discrete sequence learning networks have great potential in solving partial differential equations and time series prediction compared to traditional fully connected PINN algorithms, and can serve as the foundation for data-driven sequence prediction modeling and inverse problem analysis. However, such existing models are unable to deal with inverse problem scenarios in which the parameters of the physical process are time-varying and unknown, while usually failing to make predictions in continuous time. In this paper, we propose a continuous time series prediction algorithm constructed by the physics-informed graph neural ordinary differential equation (PGNODE). Proposed parameterized GNODE-GRU and physics-informed loss constraints are used to explicitly characterize and solve unknown time-varying hyperparameters. The GNODE solver integrates this physical parameter to predict the sequence value at any time. This paper uses epidemic prediction tasks as a case study, and experimental results demonstrate that the proposed algorithm can effectively improve the prediction accuracy of the spread of epidemics in the future continuous time. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
4. Assessment of data‐driven modeling approaches for chromatographic separation processes.
- Author
-
Michalopoulou, Foteini and Papathanasiou, Maria M.
- Subjects
ARTIFICIAL neural networks ,ENGINEERING models ,PARTIAL differential equations ,SYSTEMS engineering ,COMPUTER systems - Abstract
Chromatographic separation processes are described by nonlinear partial differential and algebraic equations, which may result in high computational cost, hindering further online applications. To decrease the computational burden, different data‐driven modeling approaches can be implemented. In this work, we investigate different strategies of data‐driven modeling for chromatographic processes, using artificial neural networks to predict pseudo‐dynamic elution profiles, without the use of explicit temporal information. We assess the performance of the surrogates trained on different dataset sizes, achieving good predictions with a minimum of 3400 data points. Different activation functions are used and evaluated against the original high‐fidelity model, using accuracy, interpolation, and simulation time as performance metrics. Based on these metrics, the best performing data‐driven models are implemented in a process optimization framework. The results indicate that data‐driven models can capture the nonlinear profile of the process and that can be considered as reliable surrogates used to aid process development. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
5. Forecasting operation of a chiller plant facility using data-driven models.
- Author
-
Salimian Rizi, Behzad, Faramarzi, Afshin, Pertzborn, Amanda, and Heidarinejad, Mohammad
- Subjects
- *
MOVING average process , *INTELLIGENT agents , *CONSUMPTION (Economics) , *PREDICTION models , *ENERGY consumption - Abstract
• Predicted power consumption and COP a chiller plant using XGBoost. • Quantified impacts of data intervals and data processing on the accuracy of models. • Improved the predictions compared to the baseline using data smoothing methods. • Demonstrated a guide to develop data-driven chiller power and COP models. In recent years, data-driven models have enabled accurate prediction of chiller power consumption and chiller coefficient of performance (COP). This study evaluates the usage of time series Extreme Gradient Boosting (XGBoost) models to predict chiller power consumption and chiller COP of a water-cooled chiller plant. The 10-second measured data used in this study are from the Intelligent Building Agents Laboratory (IBAL), which includes two water-cooled chillers. Preprocessing, data selection, noise analysis, and data smoothing methods influence the accuracy of these data-driven predictions. The data intervals were changed to 30 s, 60 s, and 180 s using down-sampling and averaging strategies to investigate the impact of data preprocessing methods and data resolutions on the accuracy of chiller COP and power consumption models. To overcome the effect of noise on the accuracy of the models of chiller power consumption and COP, two data smoothing methods, the moving average window strategy and the Savitzky-Golay (SG) filter, are applied. The results show that both methods improve the predictions compared to the baseline, with the SG filter slightly outperforming the moving average. Particularly, the mean absolute percentage error of the chiller COP and power consumption models improved from 4.8 to 4.9 for the baseline to 1.9 and 2.3 with the SG filter, respectively. Overall, this study provides a practical guide to developing XGBoost data-driven chiller power consumption and COP prediction models. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
6. Quantifying Streamflow Prediction Uncertainty Through Process‐Aware Data‐Driven Models.
- Author
-
Roy, Abhinanda and Kasiviswanathan, K. S.
- Subjects
RANDOM forest algorithms ,HYDROLOGIC models ,SUPPORT vector machines ,EMERGENCY management ,PARAMETRIC modeling - Abstract
The hydrological model simulation accompanied with uncertainty quantification helps enhance their overall reliability. Since uncertainty quantification including all the sources (input, model structure and parameter) is challenging, it is often limited to only addressing model parametric uncertainty, neglecting other uncertainty sources. This paper focuses on exploiting the potential of state‐of‐the‐art data‐driven models (or DDMs) in quantifying the prediction uncertainty of process‐based hydrological models. This is achieved by integrating the robust predictive ability of the DDMs with the process understanding ability of the hydrological models. The Bayesian‐based data assimilation (DA) technique is used to quantify uncertainty in process‐based hydrological models. This is accomplished by choosing two DDMs, random forest algorithm (RF) and support vector machine (SVM), which are distinctly integrated with two process‐based hydrological models: HBV and HyMOD. Particle filter algorithm (PF) is chosen for uncertainty quantification. All these combinations led to four different process‐aware DDMs: HBV‐PF‐RF, HBV‐PF‐SVM, HyMOD‐PF‐RF and HyMOD‐PF‐SVM. The assessment of these models on the Baitarani, Beas and Sunkoshi river basins exemplified an improvement in the accuracy of the daily streamflow simulations with a reduction in the prediction uncertainty across all the models for all the basins. For example, the nash‐sutcliffe efficiency improved by 54.69% and 10.61% in calibration and validation of the Baitarani river basin, respectively. Equivalently, average bandwidth improved by 79.37% and 71.59%, respectively. This signified the (a) potential of the DDMs in quantifying and reducing the prediction uncertainty of the hydrological model simulations, (b) transferability of the model with an appreciable performance irrespective of the choice of basins having varying topography and climatology and (c) ability to perform significantly irrespective of different process‐based and DDMs being involved, thereby ensuring generalizability. Thus, the framework is expected to assist in effective decision‐making, including various environmental management and disaster preparedness. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
7. Assessment of Hull and Propeller Degradation Due to Biofouling Using Tree-Based Models.
- Author
-
Themelis, Nikos, Nikolaidis, George, and Zagkas, Vasilios
- Subjects
SHIP hull fouling ,STATISTICAL smoothing ,DECISION trees ,FOULING ,SHIP maintenance - Abstract
A hull and propeller biofouling assessment framework is presented and demonstrated using a bulk carrier as a case study corresponding to an operational period of two and a half years. The aim is to support the decision-making process for optimizing maintenance related to hull and propeller cleaning actions. For the degradation assessment, an appropriate key performance indicator is defined comparing the expected shaft power required with the measured power under the same operational conditions. The power prediction models are data-driven based on machine learning algorithms. The process includes feature engineering, filtering, and data smoothing, while an evaluation of regression algorithms of the decision tree family is performed. The extra trees algorithm was selected, presenting a mean absolute percentage error of 1.1%. The analysis incorporates two prediction models corresponding to two different approaches. In the first, the model is employed as a reference performance baseline representing the clean vessel. When applied to a dataset reflecting advanced stages of biofouling, an average power increase of 11.3% is predicted. In the second approach, the model entails a temporal feature enabling the examination of scenarios at different points in time. Considering synthetic data corresponding to 300 days since hull cleaning, it was derived that the fouled vessel required an average 20.5% increase in power. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
8. Digital Twin Framework for Aircraft Lifecycle Management Based on Data-Driven Models.
- Author
-
Kabashkin, Igor
- Subjects
- *
DIGITAL twins , *TECHNOLOGICAL innovations , *DATA analytics , *DIGITAL technology , *MODEL airplanes , *DIGITAL communications - Abstract
This paper presents a comprehensive framework for implementing digital twins in aircraft lifecycle management, with a focus on using data-driven models to enhance decision-making and operational efficiency. The proposed framework integrates cutting-edge technologies such as IoT sensors, big data analytics, machine learning, 6G communication, and cloud computing to create a robust digital twin ecosystem. This paper explores the key components of the framework, including lifecycle phases, new technologies, and models for digital twins. It discusses the challenges of creating accurate digital twins during aircraft operation and maintenance and proposes solutions using emerging technologies. The framework incorporates physics-based, data-driven, and hybrid models to simulate and predict aircraft behavior. Supporting components like data management, federated learning, and analytics tools enable seamless integration and operation. This paper also examines decision-making models, a knowledge-driven approach, limitations of current implementations, and future research directions. This holistic framework aims to transform fragmented aircraft data into comprehensive, real-time digital representations that can enhance safety, efficiency, and sustainability throughout the aircraft lifecycle. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
9. A review of current research on occupant-centric control for improving comfort and energy efficiency.
- Author
-
Yuan, Yue, Song, Chengcheng, Gao, Liying, Zeng, Kejun, and Chen, Yixing
- Abstract
Occupant-centric control (OCC) is intelligent control of building systems based on the real comfort needs of occupants. This paper provides a comprehensive review of how real-world data on energy-related occupant behavior (OB) can be integrated and applied in OCC systems. The aim is to accurately portray the real occupant needs and improve energy efficiency without sacrificing occupant comfort. This paper first introduces two types of OB: detailed occupancy states and energy-interaction behaviors, including methods to monitor, establish, and predict these OB. Then, OCC is divided into real-time control and model-based predictive control, and each of these four scenarios is discussed. It extensively reviews OCC methods for different equipment in four cases, covering control strategies, control scales, comfort enhancement scenarios, and energy-saving potential for each category. It is summarized that despite extensive research on OB, there are still significant challenges in integrating this research into OCC. A major issue is the lack of a bridge connecting monitoring acquired information and controls. In addition, the article reviews the current state of OCC platform development. The future direction should be combined with advanced Internet of Things (IoT) technologies, WiFi, and other communication technologies to obtain information about people's behavior and real needs in order to create truly energy efficient and comfortable smart environments. The article also discusses how enhancing the real-time feedback capability of the OCC system can help improve the overall control system capability and the importance of testing through experimentation. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
10. Urban Building Energy Modeling: A Comparative Study of Process-Driven and Data-Driven Models.
- Author
-
Montazeri, Ahad, Usta, Yasemin, and Mutani, Guglielmina
- Subjects
MACHINE learning ,ENERGY consumption of buildings ,SUSTAINABLE urban development ,RANDOM forest algorithms ,CONSUMPTION (Economics) - Abstract
This study investigates the predictive capabilities of process-driven (PD) energy modeling and Machine Learning techniques, specifically Light Gradient Boosting Machine (LGBM) and Random Forest (RF) algorithms, in analyzing building energy consumption patterns. Leveraging a comprehensive dataset encompassing diverse building characteristics, energy-related variables, and operational configurations, the comparative performances of these methodologies is explored. Results reveal that while all approaches demonstrate promising predictive accuracies, LGBM exhibits a slight advantage over RF and the process-driven model. Moreover, the process-driven model showcases efficacy in colder seasons and for buildings of extreme ages, while encountering limitations in accurately modeling energy consumption for structures constructed during 1970s to 1990s. Conversely, Machine Learning models demonstrate consistent performance (with relative errors of 5-10%) across varied building ages, underscoring their adaptability and potential for capturing nuanced energy dynamics. However, a notable constraint lies in the availability of sufficient data for training Machine Learning models, posing challenges for model testing. These findings contribute to advancing our understanding of energy modeling methodologies at urban scale and offer insights for optimizing building energy efficiency strategies for a sustainable development of urban environments. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
11. Machine-learning based prediction of hydrogen/methane mixture solubility in brine
- Author
-
Farag M. A. Altalbawy, Mustafa Jassim Al-saray, Krunal Vaghela, Nodira Nazarova, Raja Praveen K. N., Bharti Kumari, Kamaljeet Kaur, Salima B. Alsaadi, Sally Salih Jumaa, Ahmed Muzahem Al-Ani, Mohammed Al-Farouni, and Ahmad Khalid
- Subjects
Machine learning ,Data-driven models ,Relevancy factor ,Outlier detection ,Medicine ,Science - Abstract
Abstract With regard to underground hydrogen storage projects, presuming that the hydrogen storage site has served as a repository for methane, the coexistence of a blend of methane and hydrogen is anticipated during the incipient stage of hydrogen storage. Therefore, the solubility of hydrogen/methane mixtures in brine becomes imperative. On the contrary, laboratory tasks of such measurements are hard because of its extreme corrosion ability and flammability, hence modeling methodologies are highly preferred. Therefore, in this study, we seek to create accurate data-driven intelligent models based upon laboratory data using hybrid models of adaptive neuro-fuzzy inference system (ANFIS) and least squares support vector machine (LSSVM) optimized with either particle swarm optimization (PSO), genetic algorithm (GA) and coupled simulated annealing (CSA) to predict hydrogen/methane mixture solubility in brine as a function of pressure, temperature, hydrogen mole fraction in hydrogen/methane mixture and brine salt concentration. The results indicate that almost all the gathered experimental data are technically suitable for the model development. The sensitivity study shows that pressure and hydrogen mole fraction in the mixture are strongly related with the solubility data with direct and indirect effects, respectively. The analyses of evaluation indexes and graphical methods indicates that the developed LSSVM-GA and LSSVM-CSA models are the most accurate as they exhibit the lowest AARE% and MSE values and the highest R-squared values. These findings show that machine learning methods could be a useful tool for predicting hydrogen solubility in brine encountered in underground hydrogen storage projects, aiding in the advancement of intelligent, affordable, and secure hydrogen storage technologies.
- Published
- 2024
- Full Text
- View/download PDF
12. Predictive machine learning in optimizing the performance of electric vehicle batteries: Techniques, challenges, and solutions.
- Author
-
Naresh, Vankamamidi S., Ratnakara Rao, Guduru V. N. S. R., and Prabhakar, D. V. N.
- Subjects
- *
REMAINING useful life , *MACHINE learning , *DEEP learning , *ENERGY consumption , *RESEARCH personnel , *ELECTRIC vehicle batteries - Abstract
This research paper explores the importance of optimizing the performance of electric vehicle (EV) batteries to align with the rapid growth in EV usage. It uses predictive machine learning (ML) techniques to achieve this optimization. The paper covers various ML methods like supervised, unsupervised, and deep learning (DL) and ways to measure their effectiveness. Significant battery performance factors, such as state of charge (SoC), state of health (SoH), state of function (SoF), and remaining useful life (RUL), are discussed, along with methods to collect and prepare data for accurate predictions. The paper introduces an operation research model for optimizing the performance of EV Batteries. It also looks at challenges unique to battery systems and ways to overcome them. The study showcases ML models' ability to predict battery behavior for real‐time monitoring, efficient energy use, and proactive maintenance. The paper categorizes different applications and case studies, providing valuable insights and forward‐looking perspectives for researchers, practitioners, and policymakers involved in improving EV battery performance through predictive ML. This article is categorized under:Technologies > ClassificationFundamental Concepts of Data and Knowledge > Explainable AITechnologies > Machine Learning [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
13. Guidance on the construction and selection of relatively simple to complex data-driven models for multi-task streamflow forecasting.
- Author
-
Tran, Trung Duc and Kim, Jongho
- Subjects
- *
ARTIFICIAL neural networks , *BOX-Jenkins forecasting , *PARTICLE swarm optimization , *LEAD time (Supply chain management) , *TRANSFORMER models - Abstract
With the goal of forecasting streamflow time series with sufficient lead time, we evaluate the efficiency and accuracy of data-based models ranging from relatively simple to complex. Based on this, we systematically explain the model construction and selection process according to lead time, type and amount of data, and optimization method. This analysis involved optimizing the inputs and hyperparameters of four unique data-driven models: Autoregressive Integrated Moving Average (ARIMA), Artificial Neural Network (ANN), Long Short-Term Memory (LSTM), and Transformer (TRANS), which were applied to the Soyang watershed, South Korea. The type and amount of model inputs are determined through a fine-tuning process that samples based on a correlation threshold, correlation to predictand, and autocorrelation to historical data and evaluates the simulated objective function. Hyperparameters are simultaneously optimized using three conventional optimization methods: Bayesian optimization (BO), particle swarm optimization (PSO), and gray wolf optimization (GWO). The experimental results provide insight into the role of input predictors, data preparations (e.g., wavelet transform), hyperparameter optimization, and model structures. From this, we can provide guidelines for model selection. Relatively simple models can be used when the dataset is small or there are few input variables, when only the near future is predicted, or when the selection of optimization methods is limited. However, a more complex model should be selected if the type and amount of data are sufficient, various optimization methods can be applied, or it is necessary to secure more lead time. More parameters, more complex model structures, and more training materials make this possible. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
14. A reduced order model formulation for left atrium flow: an atrial fibrillation case.
- Author
-
Balzotti, Caterina, Siena, Pierfrancesco, Girfoglio, Michele, Stabile, Giovanni, Dueñas-Pamplona, Jorge, Sierra-Pallares, José, Amat-Santos, Ignacio, and Rozza, Gianluigi
- Subjects
- *
CARDIAC output , *BLOOD flow , *ATRIAL fibrillation , *LEFT heart atrium , *BLOOD testing , *BLOOD viscosity - Abstract
A data-driven reduced order model (ROM) based on a proper orthogonal decomposition-radial basis function (POD-RBF) approach is adopted in this paper for the analysis of blood flow dynamics in a patient-specific case of atrial fibrillation (AF). The full order model (FOM) is represented by incompressible Navier–Stokes equations, discretized with a finite volume (FV) approach. Both the Newtonian and the Casson's constitutive laws are employed. The aim is to build a computational tool able to efficiently and accurately reconstruct the patterns of relevant hemodynamics indices related to the stasis of the blood in a physical parametrization framework including the cardiac output in the Newtonian case and also the plasma viscosity and the hematocrit in the non-Newtonian one. Many FOM-ROM comparisons are shown to analyze the performance of our approach as regards errors and computational speed-up. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
15. Applying Machine Learning Methods to Improve Rainfall–Runoff Modeling in Subtropical River Basins.
- Author
-
Yu, Haoyuan and Yang, Qichun
- Subjects
MACHINE learning ,KRIGING ,SUPPORT vector machines ,MACHINE performance ,STREAMFLOW ,WATERSHEDS - Abstract
Machine learning models' performance in simulating monthly rainfall–runoff in subtropical regions has not been sufficiently investigated. In this study, we evaluate the performance of six widely used machine learning models, including Long Short-Term Memory Networks (LSTMs), Support Vector Machines (SVMs), Gaussian Process Regression (GPR), LASSO Regression (LR), Extreme Gradient Boosting (XGB), and the Light Gradient Boosting Machine (LGBM), against a rainfall–runoff model (WAPABA model) in simulating monthly streamflow across three subtropical sub-basins of the Pearl River Basin (PRB). The results indicate that LSTM generally demonstrates superior capability in simulating monthly streamflow than the other five machine learning models. Using the streamflow of the previous month as an input variable improves the performance of all the machine learning models. When compared with the WAPABA model, LSTM demonstrates better performance in two of the three sub-basins. For simulations in wet seasons, LSTM shows slightly better performance than the WAPABA model. Overall, this study confirms the suitability of machine learning methods in rainfall–runoff modeling at the monthly scale in subtropical basins and proposes an effective strategy for improving their performance. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
16. Recent Advances in Laser Surface Hardening: Techniques, Modeling Approaches, and Industrial Applications.
- Author
-
Łach, Łukasz
- Subjects
SURFACE hardening ,METALLIC surfaces ,FINITE element method ,ARTIFICIAL intelligence ,WEAR resistance - Abstract
The article provides a comprehensive review of the latest developments in the field of laser surface hardening (LSH) and its modeling techniques. LSH is a crucial process for enhancing the surface properties of metals, particularly their hardness and wear resistance, without compromising their bulk properties. This review highlights the fundamental principles of LSH, the types of lasers used, and the key parameters influencing the hardening process. It delves into various modeling approaches, including finite element method (FEM) simulations, analytical models, and empirical models (using statistical methods), emphasizing the integration of advanced computational techniques such as machine learning and artificial intelligence to improve the accuracy and efficiency of LSH simulations. The review also explores practical applications across different industries, showcasing how LSH models have been used to solve real-world challenges in the automotive, aerospace, and tool manufacturing sectors. Finally, it addresses current limitations and outlines future research directions, suggesting potential areas for further advancements in the modeling and application of LSH processes. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
17. Uncovering Historical Reservoir Operation Rules and Patterns: Insights From 452 Large Reservoirs in the Contiguous United States.
- Author
-
Li, Donghui, Chen, Yanan, Lyu, Lingqi, and Cai, Ximing
- Subjects
HEDGING (Finance) ,CONUS ,BRICKS ,STORAGE - Abstract
Reservoir operations are influenced by hydroclimatic variability, reservoir characteristics (i.e., size and purpose), policy regulation, as well as operators' experiences and justification. Data‐driven reservoir operation models based on long‐term historical records shed light on understanding reservoir operation rules and patterns. This study applies generic data‐driven reservoir operation models (GDROMs) developed for 452 data‐rich reservoirs with diversified operation purposes across the CONUS to explore typical operation rules and patterns. We find that the operating policies of any of these reservoirs can be modeled with a small number (1–8) of typical operation modules. The derived modules applied to different conditions of the 452 reservoirs can be categorized into five basic types, that is, constant release, inflow‐driven piecewise constant release, inflow‐driven linear release, storage‐driven piecewise constant release, and storage‐driven nonlinear (or piecewise linear) release. Additionally, a joint‐driven release module, constructed from these five basic types, has been identified. The analysis further shows the module application transition patterns featuring operation dynamics for reservoirs of different operation purposes, sizes, and locations. The typical module types can be used as "Lego" bricks to build operation models, especially for data‐scarce reservoirs. These module types and their application and transition conditions can inform Standard Operation Policy (SOP) and Hedging Policy (HP) with specific inflow, storage, and/or both conditions. Key Points: Five basic types of operation modules are categorized for 452 reservoirsSeasonal patterns for module application transition are identified for reservoirs with different sizes, operation purposes, and locationsThe basic types of modules and operation patterns inform reservoir operation modeling [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
18. Blood Glucose Prediction from Nutrition Analytics in Type 1 Diabetes: A Review.
- Author
-
Lubasinski, Nicole, Thabit, Hood, Nutter, Paul W., and Harper, Simon
- Abstract
Introduction: Type 1 Diabetes (T1D) affects over 9 million worldwide and necessitates meticulous self-management for blood glucose (BG) control. Utilizing BG prediction technology allows for increased BG control and a reduction in the diabetes burden caused by self-management requirements. This paper reviews BG prediction models in T1D, which include nutritional components. Method: A systematic search, utilizing the PRISMA guidelines, identified articles focusing on BG prediction algorithms for T1D that incorporate nutritional variables. Eligible studies were screened and analyzed for model type, inclusion of additional aspects in the model, prediction horizon, patient population, inputs, and accuracy. Results: The study categorizes 138 blood glucose prediction models into data-driven (54%), physiological (14%), and hybrid (33%) types. Prediction horizons of ≤30 min are used in 36% of models, 31–60 min in 34%, 61–90 min in 11%, 91–120 min in 10%, and >120 min in 9%. Neural networks are the most used data-driven technique (47%), and simple carbohydrate intake is commonly included in models (data-driven: 72%, physiological: 52%, hybrid: 67%). Real or free-living data are predominantly used (83%). Conclusion: The primary goal of blood glucose prediction in T1D is to enable informed decisions and maintain safe BG levels, considering the impact of all nutrients for meal planning and clinical relevance. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
19. Prediction of volume of shallow landslides due to rainfall using data-driven models.
- Author
-
Tuganishuri, Jeremie, Yune, Chan-Young, Adhikari, Manik Das, Lee, Seung Woo, Kim, Gihong, and Yum, Sang-Guk
- Subjects
LANDSLIDES ,RAINFALL ,ARTIFICIAL neural networks ,NATURAL disaster warning systems ,NATURAL disasters ,LANDSLIDE prediction ,FLOOD warning systems ,SUPPORT vector machines - Abstract
Landslides due to rainfall are among most destructive natural disasters that cause property damages, huge financial losses and human deaths in different parts of the World. To plan for mitigation and resilience, the prediction of the volume of rainfall-induced landslides is essential to understand the relationship between the volume of soil materials debris and their associated predictors. Objectives of this research are to construct a model by utilizing advanced data-driven algorithms (i.e., ordinary least square or Linear regression (OLS), random forest (RF), support vector machine (SVM), extreme gradient boosting (EGB), generalized linear model (GLM), decision tree (DT), and deep neural network (DNN), K-nearest neighbor (KNN) and Ridge regression (RR)) for the prediction of the volume of landslides due to rainfall considering geological, geomorphological, and environmental conditions. Models were tested on the Korean landslide dataset to observe the best-performing model, and among tested algorithms, the extreme gradient boosting ranked high with the coefficient of determination (R
2 = 0.85) and mean absolute error (MAE = 150.421 m3 ). The volume of landslides was strongly influenced by slope length, drainage status, slope angle, aspect, and age of trees. The anticipated volume of landslide can be important for land use allocation and efficient landslide risk management. [ABSTRACT FROM AUTHOR]- Published
- 2024
- Full Text
- View/download PDF
20. A DESCENT ALGORITHM FOR THE OPTIMAL CONTROL OF ReLU NEURAL NETWORK INFORMED PDEs BASED ON APPROXIMATE DIRECTIONAL DERIVATIVES.
- Author
-
GUOZHI DONG, HINTERMÜLLER, MICHAEL, and PAPAFITSOROS, KOSTAS
- Subjects
- *
PARTIAL differential equations , *DIRECTIONAL derivatives , *ALGORITHMS - Abstract
We propose and analyze a numerical algorithm for solving a class of optimal control problems for learning-informed semilinear partial differential equations (PDEs). Such PDEs contain constituents that are in principle unknown and are approximated by nonsmooth ReLU neural networks. We first show that direct smoothing of the ReLU network with the aim of using classical numerical solvers can have disadvantages, such as potentially introducing multiple solutions for the corresponding PDE. This motivates us to devise a numerical algorithm that treats directly the nonsmooth optimal control problem, by employing a descent algorithm inspired by a bundle-free method. Several numerical examples are provided and the efficiency of the algorithm is shown. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
21. Empowering smart cities with digital twins of buildings: Applications and implementation considerations of data-driven energy modelling in building management.
- Author
-
Elnour, Mariam, Ahmad, Ahmad M., Abdelkarim, Shimaa, Fadli, Fodil, and Naji, Khalid
- Abstract
Smart buildings and cities are rapidly emerging as solutions to address the challenges of efficiency, urbanisation, and sustainability in the sector. The study proposes deploying data-driven digital twins for smart buildings by utilising the available building's technology and IT infrastructure to complement and augment existing functions. The digital twin will consist of a core data-driven energy model and a 2D visual representation of the building's systems, with the potential for future evolution into a 3D model. This study aims to present a preliminary investigation into the idea of data-driven digital twins in building management towards enhancing the operations of smart buildings and empowering the concept of smart cities. It is demonstrated on a building on the campus of Qatar University. With an emphasis on the air conditioning systems of the building, considering their substantial contribution to overall energy consumption, the study maintains an open approach to also encompass other energy systems within the buildings, and presents a comparative evaluation between simulation-based and data-driven modelling on the case study, as well as an exploration of various machine learning algorithms that can be used. Furthermore, exploring essential smart applications of the building's data-driven digital twin. Practical Application: The study provides a comprehensive exploration of the practical aspects of deploying data-driven digital twins for smart buildings, addressing challenges related to data collection, model development, integration with building infrastructure, and potential limitations. The paper aims to advance the field of facility management and promote smart and sustainable practices in building operations. By contributing to the existing knowledge in facility services and management, our study offers practical guidance towards optimising building performance, reducing energy consumption, and fostering sustainable urban development. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
22. Designing an explainable bio-inspired model for suspended sediment load estimation: eXtreme Gradient Boosting coupled with Marine Predators Algorithm
- Author
-
Roozbeh Moazenzadeh, Okan Mert Katipoğlu, Ahmadreza Shateri, Hamid Nasiri, and Mohammed Abdallah
- Subjects
Suspended sediment load ,Machine learning ,Hybrid algorithm ,Data-driven models ,Sediment concentration ,SHAP ,Engineering (General). Civil engineering (General) ,TA1-2040 - Abstract
This study aimed to develop an accurate and reliable model for predicting suspended sediment load (SL) in river systems, which is crucial for water resource management and environmental protection. While Xtreme Gradient Boosting (XGB), a powerful ensemble machine learning (ML) model, has been employed in previous studies, the novelty of this research lies in the introduction of a hybrid approach that synergistically combines XGB with the bio-inspired Marine Predators Algorithm (XGB-MPA) to estimate SL in the Yeşilirmak River (Turkey). To this end, streamflow (Q) and sediment concentration (SC) values as well as their lag times (1 to 3 month lag times) were fed as input variables – under 9 scenarios – into ML models. A time series of datasets from March 1973 to December 2011 and January 2012 to March 2023 were used for training and testing of ML models, respectively. The superiority of the proposed model (XGB-MPA) compared to two other hybrid models, including XGB-PSO (Particle Swarm Optimization) and XGB-GWO (Grey Wolf Optimization) was also investigated. According to the results, the simultaneous application of Q and SC lag time values as inputs has led to the best SL estimates by XGB-MPA, with XGB-MPA9 (RMSE = 103.7 ton/day; NSE = 0.96) exhibiting the lowest error rates. In addition, XGB-MPA has performed better than XGB in all scenarios, with the lowest and highest reduction in RMSE being 19.3% (scenario 5) and 97.4% (scenario 1), respectively. When comparing the performance of hybrid models, the proposed XGB-MPA model has performed best with MAE, RMSE and NSE of 40.94, 103.7 and 0.96, respectively, in comparison with 816.02, 1063.74 and −2.94 for XGB-PSO and 693.16, 981.68 and −2.37 for XGB-GWO. Further research can include the use of time series of efficient variables extracted from satellite images (e.g. land cover, river morphology, etc.) as model inputs.Highlights Improvement of SL estimation by coupling MPA with XGB modelUsing delayed combinations of streamflow and sediment concentration as model inputsSuperiority of MPA compare to PSO and GWO in SL estimationGreater variations in SHAP caused by sediment concentration compared to streamflowFurther studies required on the effects of hydrological and topographical features
- Published
- 2024
- Full Text
- View/download PDF
23. A self-learning framework combining association rules and mathematical models to solve production scheduling programs
- Author
-
Mateo Del Gallo, Sara Antomarioni, Giovanni Mazzuto, Giulio Marcucci, and Filippo Emanuele Ciarapica
- Subjects
Production scheduling and control ,association rules ,data-driven models ,big data analytics ,optimization techniques ,Technology ,Manufactures ,TS1-2301 ,Business ,HF5001-6182 - Abstract
Data-driven production scheduling and control systems are essential for manufacturing organisations to quickly adjust to the demand for a wide range of bespoke products, often within short lead times. This paper presents a self-learning framework that combines association rules and optimization techniques to create data-driven production scheduling. A new approach to predicting interruptions in the production process through association rules was implemented, using a mathematical model to sequence production activities in real or near real-time. The framework was tested in a case study of a ceramics manufacturer, updating confidence values by comparing planned values to actual values recorded during production control. It also sets a production corrective factor based on confidence value and success rate to avoid product shortages. The results were generated in just 1.25 seconds, resulting in a makespan reduction of 9% and 6% compared to two heuristics based on First-In-First-Out and Short Processing Time strategies.
- Published
- 2024
- Full Text
- View/download PDF
24. Nonlinear parametric models of viscoelastic fluid flows
- Author
-
C. M. Oishi, A. A. Kaptanoglu, J. Nathan Kutz, and S. L. Brunton
- Subjects
viscoelastic fluids ,computational fluid dynamics ,data-driven models ,sparse identification of nonlinear dynamics ,reduced-order models ,machine learning ,Science - Abstract
Reduced-order models (ROMs) have been widely adopted in fluid mechanics, particularly in the context of Newtonian fluid flows. These models offer the ability to predict complex dynamics, such as instabilities and oscillations, at a considerably reduced computational cost. In contrast, the reduced-order modelling of non-Newtonian viscoelastic fluid flows remains relatively unexplored. This work leverages the sparse identification of nonlinear dynamics (SINDy) algorithm to develop interpretable ROMs for viscoelastic flows. In particular, we explore a benchmark oscillatory viscoelastic flow on the four-roll mill geometry using the classical Oldroyd-B fluid. This flow exemplifies many canonical challenges associated with non-Newtonian flows, including transitions, asymmetries, instabilities, and bifurcations arising from the interplay of viscous and elastic forces, all of which require expensive computations in order to resolve the fast timescales and long transients characteristic of such flows. First, we demonstrate the effectiveness of our data-driven surrogate model to predict the transient evolution and accurately reconstruct the spatial flow field for fixed flow parameters. We then develop a fully parametric, nonlinear model capable of capturing the dynamic variations as a function of the Weissenberg number. While the training data are predominantly concentrated on a limit cycle regime for moderate [Formula: see text], we show that the parametrized model can be used to extrapolate, accurately predicting the dominant dynamics in the case of high Weissenberg numbers. The proposed methodology represents an initial step in applying machine learning and reduced-order modelling techniques to viscoelastic flows.
- Published
- 2024
- Full Text
- View/download PDF
25. AI- and ML-based Models for Predicting Remaining Useful Life (RUL) of Nanocomposites and Reinforced Laminated Structures
- Author
-
Goyal, Samarthya, Mondal, Suman, Mohanty, Sutanuka, Katari, Vinay, Sharma, Henu, Sahu, Kisor K., Kumar, Ashwani, editor, Kumar Singla, Yogesh, editor, and Maughan, Michael R., editor
- Published
- 2024
- Full Text
- View/download PDF
26. Smart Data-Driven Building Management Framework and Demonstration
- Author
-
Zhang, Jing, Ma, Tianyou, Xu, Kan, Chen, Zhe, Xiao, Fu, Ho, Jeremy, Leung, Calvin, Yeung, Sammy, Goos, Gerhard, Founding Editor, Hartmanis, Juris, Founding Editor, Bertino, Elisa, Editorial Board Member, Gao, Wen, Editorial Board Member, Steffen, Bernhard, Editorial Board Member, Yung, Moti, Editorial Board Member, Jørgensen, Bo Nørregaard, editor, da Silva, Luiz Carlos Pereira, editor, and Ma, Zheng, editor
- Published
- 2024
- Full Text
- View/download PDF
27. Modeling Nonlinear Structures Using Physics-Guided, Machine-Learnt Models
- Author
-
Szydlowski, Michal J., Schwingshackl, Christoph, Renson, Ludovic, Zimmerman, Kristin B., Series Editor, Brake, Matthew R.W., editor, Renson, Ludovic, editor, Kuether, Robert J., editor, and Tiso, Paolo, editor
- Published
- 2024
- Full Text
- View/download PDF
28. Assessment and deployment of a LSTM-based virtual sensor in an industrial process control loop
- Author
-
González-Herbón, Raúl, González-Mateos, Guzmán, Rodríguez-Ossorio, José R., Prada, Miguel A., Morán, Antonio, Alonso, Serafín, Fuertes, Juan J., and Domínguez, Manuel
- Published
- 2024
- Full Text
- View/download PDF
29. Modeling compressive strength and environmental impact points of fly ash-admixed concrete using data-driven approaches
- Author
-
Singh, Sandeep, Meena, Y. R., Rapeti, Srinivasa Rao, Kedia, Navin, Issa, Salman Khalaf, and Abbas, Haider M.
- Published
- 2024
- Full Text
- View/download PDF
30. A review of the state of the art in solar photovoltaic output power forecasting using data-driven models
- Author
-
Gupta, Ankur Kumar and Singh, Rishi Kumar
- Published
- 2024
- Full Text
- View/download PDF
31. Effects of compositional uncertainties in cracked NH3/biosyngas fuel blends on the combustion characteristics and performance of a combined-cycle gas turbine: A numerical thermokinetic study.
- Author
-
Soyler, Israfil, Zhang, Kai, Jiang, Xi, and Karimi, Nader
- Subjects
- *
GAS turbines , *COMBUSTION , *FLAME , *DIESEL motor combustion , *ADIABATIC temperature , *FLAME temperature , *POLYNOMIAL chaos - Abstract
Blending of partially cracked ammonia with biosyngas is an attractive strategy for improving NH 3 combustion. In practice, products of biomass gasification and those of thermo-catalytic cracking of NH 3 are subject to some compositional uncertainties. Despite their practical importance, so far, the effects of such uncertainties on combustion systems remained largely unexplored. Hence, this paper quantifies the effects of small compositional uncertainties of reactants upon combustion of partially cracked NH 3 /syngas/air mixtures. An uncertainty quantification method, based on polynomial chaos expansion and a data-driven model, is utilised to investigate the effects of uncertainty in fuel composition on the laminar flame speed (S L) and adiabatic flame temperature (T ad) at different inlet pressures (P i). The analysis is then extended to the power output of a combined-cycle gas turbine fuelled by the reactants. It is found that 1.5% fuel compositional uncertainty can cause 12–21% of S L uncertainty depending on the inlet pressure. Furthermore, the effect of compositional uncertainty on T ad increases at higher ratios of H 2 to NH 3. Sensitivity analysis reveals that the uncertainty of CO contribution to S L uncertainty is higher than that of NH 3 , while the trend is reversed for the T ad uncertainty. In addition, the power output from the combined-cycle gas turbine system varies between 4 and 6% with 1.5% of fuel compositional uncertainty. This become more noticeable at elevated P i [5–10 atm], particularly when the fuel mixture contains high H 2 which is the main contributor to T ad variability. • PCE-UQ model to analyse compositional variability effects on flame properties (S L and T ad). • Additional UQ analysis examines the uncertainty impacts on a CCGT system. • H 2 variability is the strongest contributor to the uncertainty in the CCGT power output. • Richer fuel mixtures are most affected (6%) by minor fuel compositional variability. • Uncertainty impact is higher for high H 2 fuel mixtures combustion at elevated pressures. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
32. Analysis of Data Generation and Preparation for Porosity Prediction in Cold Spray using Machine Learning.
- Author
-
Eberle, Martin, Pinches, Samuel, Osborne, Max, Qin, Kai, and Ang, Andrew
- Subjects
- *
MACHINE learning , *FEATURE selection , *POROSITY , *DATA analysis , *DATA scrubbing , *COATING processes - Abstract
Cold spray is an additive manufacturing and coating process in which powder particles are accelerated to supersonic speeds without melting them and then deposit on a surface to form a layer of a coating. Process parameters and materials affect the characteristics of manufactured parts and therefore must be chosen with care. Machine learning (ML) techniques have been specifically applied in additive manufacturing for tasks such as predicting and characterizing porosity. Machine learning algorithms can learn how a variation in the input spray parameters affects annotated output data, such as experimentally measured part properties. In this work, a dataset was developed from experiments reported in published academic papers, to train ML algorithms for the porosity prediction of cold spray manufactured parts. Data cleaning steps, such as null value replacement and categorical feature handling, were applied to prepare the dataset for the training of different ML models. The dataset was split into training and testing portions, and floating feature selection and hyperparameter optimization were performed using parts of the training set. A final evaluation of all trained models, using the test portion of the dataset, showed that a prediction accuracy with an average deviation of 0-2% porosity of the predicted values compared to the true values can be achieved. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
33. Application of hybridized ANN–GARCH, ANN–SETAR, MARS–SPSO, and CANFIS–SPSO meta-models for improving accuracy of monthly streamflow prediction.
- Author
-
Mikaeili, Omidreza and Shourian, Mojtaba
- Subjects
- *
WATER management , *HYDROLOGIC cycle , *STREAMFLOW , *EXTREME value theory - Abstract
Among the components of the hydrological cycle, stream flow has a major role in integrated water resources management. Establishing an accurate and reliable forecasting method for prediction of stream flow is very useful. Nowadays, data-driven methods are variously applied for river flow prediction. By hybridizing, one can take advantage of the cons of different methods for the proposed purpose. In the present research, we have combined SETAR and GARCH methods with ANN and also coupled MARS and CANFIS with SPSO to predict the monthly flow of the Maroon River in south west of Iran. Thus, four hybridized data-driven models of ANN–GARCH, ANN–SETAR, MARS–SPSO, and CANFIS–SPSO are developed and compared to see which method has the best performance. Although all the models yielded good results but it was seen that the ANN–SETAR model found more accurate answers in prediction of the stream flow with an average 5% higher accuracy in the results. The IQR of ANN–SETAR model is similar to observed value that this showed the efficiency of the ANN–SETAR for dependable simulation of extreme values of river flow compared to other models. So, it was concluded that the ANN–SETAR model is better than the other methods for forecasting the monthly streamflow. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
34. Sim2Real Neural Controllers for Physics-Based Robotic Deployment of Deformable Linear Objects.
- Author
-
Tong, Dezhong, Choi, Andrew, Qin, Longhui, Huang, Weicheng, Joo, Jungseock, and Jawed, Mohammad Khalid
- Subjects
- *
ARTIFICIAL neural networks , *ROBOTICS , *GRAVITATIONAL energy , *FRICTION materials , *MACHINE learning , *HEURISTIC - Abstract
Deformable linear objects (DLOs), such as rods, cables, and ropes, play important roles in daily life. However, manipulation of DLOs is challenging as large geometrically nonlinear deformations may occur during the manipulation process. This problem is made even more difficult as the different deformation modes (e.g., stretching, bending, and twisting) may result in elastic instabilities during manipulation. In this paper, we formulate a physics-guided data-driven method to solve a challenging manipulation task—accurately deploying a DLO (an elastic rod) onto a rigid substrate along various prescribed patterns. Our framework combines machine learning, scaling analysis, and physical simulations to develop a physics-based neural controller for deployment. We explore the complex interplay between the gravitational and elastic energies of the manipulated DLO and obtain a control method for DLO deployment that is robust against friction and material properties. Out of the numerous geometrical and material properties of the rod and substrate, we show that only three non-dimensional parameters are needed to describe the deployment process with physical analysis. Therefore, the essence of the controlling law for the manipulation task can be constructed with a low-dimensional model, drastically increasing the computation speed. The effectiveness of our optimal control scheme is shown through a comprehensive robotic case study comparing against a heuristic control method for deploying rods for a wide variety of patterns. In addition to this, we also showcase the practicality of our control scheme by having a robot accomplish challenging high-level tasks such as mimicking human handwriting, cable placement, and tying knots. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
35. Learning dynamical models of single and collective cell migration: a review.
- Author
-
Brückner, David B and Broedersz, Chase P
- Subjects
- *
CELL migration , *CELL motility , *EQUATIONS of motion , *IMAGE analysis , *STOCHASTIC models - Abstract
Single and collective cell migration are fundamental processes critical for physiological phenomena ranging from embryonic development and immune response to wound healing and cancer metastasis. To understand cell migration from a physical perspective, a broad variety of models for the underlying physical mechanisms that govern cell motility have been developed. A key challenge in the development of such models is how to connect them to experimental observations, which often exhibit complex stochastic behaviours. In this review, we discuss recent advances in data-driven theoretical approaches that directly connect with experimental data to infer dynamical models of stochastic cell migration. Leveraging advances in nanofabrication, image analysis, and tracking technology, experimental studies now provide unprecedented large datasets on cellular dynamics. In parallel, theoretical efforts have been directed towards integrating such datasets into physical models from the single cell to the tissue scale with the aim of conceptualising the emergent behaviour of cells. We first review how this inference problem has been addressed in both freely migrating and confined cells. Next, we discuss why these dynamics typically take the form of underdamped stochastic equations of motion, and how such equations can be inferred from data. We then review applications of data-driven inference and machine learning approaches to heterogeneity in cell behaviour, subcellular degrees of freedom, and to the collective dynamics of multicellular systems. Across these applications, we emphasise how data-driven methods can be integrated with physical active matter models of migrating cells, and help reveal how underlying molecular mechanisms control cell behaviour. Together, these data-driven approaches are a promising avenue for building physical models of cell migration directly from experimental data, and for providing conceptual links between different length-scales of description. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
36. NOx emission reduction of coal‐fired power plants through data‐driven model and particle swarm optimization.
- Author
-
Abebe, Misganaw, Seo, Jin, Kang, Young‐Jin, Choi, Hyunho, and Noh, Yoojeong
- Subjects
COAL-fired power plants ,PARTICLE swarm optimization ,GREENHOUSE gas mitigation ,CARBON emissions ,PROCESS optimization - Abstract
Several studies have aimed to predict and control carbon emissions from coal‐fired power plants. However, the highly complex combustion mechanisms in coal‐fired power plant boilers pose a significant challenge in direct modeling and optimization. To tackle this challenge, this study introduced a data‐driven approach along with model‐based process optimization to mitigate NOx emissions from coal‐fired power plants. The process involved collecting a 5‐month operational dataset containing 67 controllable parameters from a 500 MW coal‐fired power plant. Steady‐state data was isolated from the load output using a moving average method, followed by the application of an isolation forest algorithm to detect and remove anomalies. Correlation analysis was then used to evaluate parameter relationships and eliminate redundant ones. Subsequently, a NOx prediction model was developed, combining an extra tree regressor data‐driven prediction model with particle swarm optimization to optimize the most influential controllable parameters for reducing NOx emissions. Testing the proposed model across four different target loads consistently resulted in a reduction of over 20% in NOx emissions by optimizing boiler combustion parameters, representing a significant achievement in optimizing coal‐fired combustion. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
37. Data-driven modelling with coarse-grid network models.
- Author
-
Lie, Knut-Andreas and Krogstad, Stein
- Subjects
- *
CALIBRATION , *TOPOLOGY , *FLUIDS , *MAPS - Abstract
We propose to use a conventional simulator, formulated on the topology of a coarse volumetric 3D grid, as a data-driven network model that seeks to reproduce observed and predict future well responses. The conceptual difference from standard history matching is that the tunable network parameters are calibrated freely without regard to the physical interpretation of their calibrated values. The simplest version uses a minimal rectilinear mesh covering the assumed map outline and base/top surface of the reservoir. The resulting CGNet models fit immediately in any standard simulator and are very fast to evaluate because of the low cell count. We show that surprisingly accurate network models can be developed using grids with a few tens or hundreds of cells. Compared with similar interwell network models (e.g., Ren et al., 2019, 10.2118/193855-MS), a typical CGNet model has fewer computational cells but a richer connection graph and more tunable parameters. In our experience, CGNet models therefore calibrate better and are simpler to set up to reflect known fluid contacts, etc. For cases with poor vertical connection or internal fluid contacts, it is advantageous if the model has several horizontal layers in the network topology. We also show that starting with a good ballpark estimate of the reservoir volume is a precursor to a good calibration. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
38. Comprehensive performance analysis of training functions in flow prediction model using artificial neural network.
- Author
-
Shikhar, K. C., Bhattarai, Khem Prasad, Tang De Shan, Mishra, Saurabh, Joshi, Ishwar, and Singh, Anurag Kumar
- Subjects
- *
ARTIFICIAL neural networks , *PREDICTION models , *STANDARD deviations , *FLOOD forecasting , *HYDROLOGICAL stations , *WATERSHEDS - Abstract
Higher Himalayan catchments are often poorly monitored for hydrological activities involving flood flow prediction for the safety of riverside communities and the successful operation of hydropower projects. This study aimed to estimate the comparative performance of artificial neural network (ANN) based flow prediction models using 10 years of daily river flow data of Kaligandaki catchment at Kotagaun, Nepal, which is a snow-fed catchment in the Himalayan region. The flow prediction models were trained and tested at a hydrological station using the previous 3 days' river flow data to predict the 1-day ahead flow data. Eight different training functions were employed in an ANN model for comprehensive statistical assessment of accuracy and precision of each training function. The most significant and validated result obtained in this study is the comprehensive comparison of various training functions' performance, and identification of the most efficient training function for the study case. Among the training functions investigated, the Levenberg-Marquardt backpropagation function exhibits the best performance for the model having Nash-Sutcliffe efficiency, root mean square error and mean absolute error values of 0.866, 209.578 and 75.422, respectively. This study provides a fundamental basis for accurate flow prediction of topographically challenged catchments where hydrological monitoring and data collection may be limited. In particular, this model will help to improve early warning system, hydrological planning, and the safety of riverside communities in the Himalayan region. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
39. Data‐driven stochastic model for quantifying the interplay between amyloid‐beta and calcium levels in Alzheimer's disease.
- Author
-
Shaheen, Hina, Melnik, Roderick, and Singh, Sundeep
- Subjects
- *
ALZHEIMER'S disease , *HOMEOSTASIS , *STOCHASTIC models , *AMYLOID plaque , *CALCIUM - Abstract
The abnormal aggregation of extracellular amyloid‐β(Aβ)$$ \left(A\beta \right) $$ in senile plaques resulting in calcium Ca+2$$ \left({Ca}^{+2}\right) $$ dyshomeostasis is one of the primary symptoms of Alzheimer's disease (AD). Significant research efforts have been devoted in the past to better understand the underlying molecular mechanisms driving Aβ$$ A\beta $$ deposition and Ca+2$$ {Ca}^{+2} $$ dysregulation. Importantly, synaptic impairments, neuronal loss, and cognitive failure in AD patients are all related to the buildup of intraneuronal Aβ$$ A\beta $$ accumulation. Moreover, increasing evidence show a feed‐forward loop between Aβ$$ A\beta $$ and Ca+2$$ {Ca}^{+2} $$ levels, that is, Aβ$$ A\beta $$ disrupts neuronal Ca+2$$ {Ca}^{+2} $$ levels, which in turn affects the formation of Aβ$$ A\beta $$. To better understand this interaction, we report a novel stochastic model where we analyze the positive feedback loop between Aβ$$ A\beta $$ and Ca+2$$ {Ca}^{+2} $$ using ADNI data. A good therapeutic treatment plan for AD requires precise predictions. Stochastic models offer an appropriate framework for modeling AD since AD studies are observational in nature and involve regular patient visits. The etiology of AD may be described as a multi‐state disease process using the approximate Bayesian computation method. So, utilizing ADNI data from 2$$ 2 $$‐year visits for AD patients, we employ this method to investigate the interplay between Aβ$$ A\beta $$ and Ca+2$$ {Ca}^{+2} $$ levels at various disease development phases. Incorporating the ADNI data in our physics‐based Bayesian model, we discovered that a sufficiently large disruption in either Aβ$$ A\beta $$ metabolism or intracellular Ca+2$$ {Ca}^{+2} $$ homeostasis causes the relative growth rate in both Ca+2$$ {Ca}^{+2} $$ and Aβ$$ A\beta $$, which corresponds to the development of AD. The imbalance of Ca+2$$ {Ca}^{+2} $$ ions causes Aβ$$ A\beta $$ disorders by directly or indirectly affecting a variety of cellular and subcellular processes, and the altered homeostasis may worsen the abnormalities of Ca+2$$ {Ca}^{+2} $$ ion transportation and deposition. This suggests that altering the Ca+2$$ {Ca}^{+2} $$ balance or the balance between Aβ$$ A\beta $$ and Ca+2$$ {Ca}^{+2} $$ by chelating them may be able to reduce disorders associated with AD and open up new research possibilities for AD therapy. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
40. Enhancing accuracy and interpretability of multi-steps water demand prediction through prior knowledge integration in neural network architecture
- Author
-
Zhengheng Pu, Deke Han, Hexiang Yan, Tao Tao, and Kunlun Xin
- Subjects
Short-term water demand forecasting ,Long-short term memory neural network ,Convolutional Neural Network ,Multi-steps forecasting ,Data-driven models ,Water supply system managements ,Environmental technology. Sanitary engineering ,TD1-1066 - Abstract
In the field of water supply management, multi-steps water demand forecasting plays a crucial role. While there have been many studies related to multi-steps water demand forecasting based on deep learning, little attention has been paid to the interpretability of forecasting models. Aiming to improve both the forecasting accuracy and interpretability of the model, a novel urban water demand forecasting neural network (UWDFNet) was presented in this paper. Compared with traditional deep learning models, it innovatively considered domain-specific prior knowledge from water supply management and incorporated the correlation relationship between different input variables into the design of the neural network structure, and verified the consistency between the knowledge learned by the model and prior knowledge through interpretability analysis. Additionally, a systematic performance evaluation was conducted and proved that UWDFNet possesses better accuracy and stability compared to other baseline models(e.g., gated recurrent unit network (GRUN), GRUN with a corrected Network (GRUN+CORRNet), GRUN+PID, GRUN+Kmeans).
- Published
- 2024
- Full Text
- View/download PDF
41. Refining Seasonal Precipitation Forecast in Brazil Using Simple Data-Driven Techniques and Climate Indices
- Author
-
Francisca Lanai Ribeiro Torres, Cassia Akemi Castro Kuki, Michelle Simões Reboita, Luana Medeiros Marangon Lima, José Wanderley Marangon Lima, and Anderson Rodrigo de Queiroz
- Subjects
data-driven models ,SEAS5 ,seasonal precipitation forecast ,teleconnection patterns ,time series forecasting ,Meteorology. Climatology ,QC851-999 - Abstract
Abstract Seasonal precipitation forecasts are essential for water resource management, agricultural activities, and the operational planning of hydropower systems. Any methodological advancement that enhances the accuracy of precipitation predictions will yield considerable societal benefits. In this context, this study proposes and evaluates two approaches for refining seasonal precipitation forecasts in Brazil, using simple data-based models, such as multiple linear regression (MLR) and nonlinear support vector machine (SVM). These models employ climate indices related to different teleconnection patterns that affect seasonal precipitation in Brazil, the unified gauge-based analysis of global daily precipitation from the Climate Prediction Center (CPC), and the precipitation forecasts from the Seasonal Forecast System 5 (SEAS5) as input variables. Both MLR and SVM models were validated from Jan-2017 to Dec-2020 using precipitation from the CPC as ground truth. The results suggest that, compared to SEAS5, MLR and SVM models enhance predictive accuracy and reduce bias in precipitation forecasts for the Southeast, Midwest, and North regions of Brazil during the austral summer. However, the performance of the models was found to be on par with the original predictions of SEAS5 in the Northeast and South regions, sectors of Brazil where the climate is significantly influenced by the El Niño-Southern Oscillation.
- Published
- 2024
- Full Text
- View/download PDF
42. Nonequilibrium statistical mechanics and optimal prediction of partially-observed complex systems
- Author
-
Rupe, Adam, Vesselinov, Velimir V, and Crutchfield, James P
- Subjects
Physical Sciences ,nonequilibrium statistical mechanics ,partially-observed systems ,Koopman operator ,Mori-Zwanzig ,data-driven models ,Fluids & Plasmas ,Physical sciences - Abstract
Only a subset of degrees of freedom are typically accessible or measurable in real-world systems. As a consequence, the proper setting for empirical modeling is that of partially-observed systems. Notably, data-driven models consistently outperform physics-based models for systems with few observable degrees of freedom; e.g. hydrological systems. Here, we provide an operator-theoretic explanation for this empirical success. To predict a partially-observed system’s future behavior with physics-based models, the missing degrees of freedom must be explicitly accounted for using data assimilation and model parametrization. Data-driven models, in contrast, employ delay-coordinate embeddings and their evolution under the Koopman operator to implicitly model the effects of the missing degrees of freedom. We describe in detail the statistical physics of partial observations underlying data-driven models using novel maximum entropy and maximum caliber measures. The resulting nonequilibrium Wiener projections applied to the Mori-Zwanzig formalism reveal how data-driven models may converge to the true dynamics of the observable degrees of freedom. Additionally, this framework shows how data-driven models infer the effects of unobserved degrees of freedom implicitly, in much the same way that physics models infer the effects explicitly. This provides a unified implicit-explicit modeling framework for predicting partially-observed systems, with hybrid physics-informed machine learning methods combining both implicit and explicit aspects.
- Published
- 2022
43. A Novel Unsupervised Anomaly Detection Framework for Early Fault Detection in Complex Industrial Settings
- Author
-
Eduardo Antonio Hinojosa-Palafox, Oscar Mario Rodriguez-Elias, Jesus Horacio Pacheco-Ramirez, Jose Antonio Hoyo-Montano, Madain Perez-Patricio, and Daniel Fernando Espejel-Blanco
- Subjects
Fault detection ,anomaly detection ,industrial analytics ,machine learning ,unsupervised learning ,data-driven models ,Electrical engineering. Electronics. Nuclear engineering ,TK1-9971 - Abstract
The increasing complexity and automation inherent in the contemporary Industry 4.0 paradigm necessitate robust and proactive fault detection methodologies to ensure both operational efficiency and safety. Existing unsupervised anomaly detection techniques, however, often encounter challenges when confronted with the high dimensionality, inherent noise, and complex interdependencies characteristic of industrial data. This paper proposes a novel unsupervised anomaly detection framework explicitly designed for early fault detection within such complex industrial environments. The proposed data-driven methodology systematically identifies the most effective unsupervised model for anomaly prediction from a candidate set of learning algorithms. This approach is particularly advantageous as it obviates the need for labeled historical fault data, a resource often limited in real-world operational settings. The 2015 PHM Data Challenge dataset, specifically selected for its inclusion of systems exhibiting incomplete fault logs, is used to validate the efficacy of the proposed framework. Findings underscore the significant potential of data-driven methodologies to enhance fault detection capabilities, thereby enabling timely intervention and contributing to the improvement of both the reliability and safety of industrial systems.
- Published
- 2024
- Full Text
- View/download PDF
44. Data-Driven Models for Yacht Hull Resistance Optimization: Exploring Geometric Parameters Beyond the Boundaries of the Delft Systematic Yacht Hull Series
- Author
-
Jake M. Walker, Andrea Coraddu, and Luca Oneto
- Subjects
Computational fluid dynamics ,data-driven models ,DelftBlue ,Delft Systematic Yacht Hull Series ,extrapolation ,hull parametrization ,Electrical engineering. Electronics. Nuclear engineering ,TK1-9971 - Abstract
Optimizing vessel hull resistance is pivotal for enhancing maritime performance and minimizing environmental impacts. Traditional methods combine expert intuition with Data-Driven Models (DDMs), relying on parametrization to predict and optimize hull geometries using Experimental Fluid Dynamics (EFD) or Computational Fluid Dynamics (CFD) data. However, these conventional approaches are hampered by several limitations: they require significant human input, are computationally intensive and costly, and lack flexibility in adapting to new families of geometries or parameters beyond predefined ranges. Addressing these challenges, our research introduces a novel method that significantly reduces the need for human intervention, computational resources, and costs, while also improving the model’s adaptability. By proposing a new a parametrization technique that accurately encompasses the Delft Systematic Yacht Hull Series (DSYHS), we demonstrate that DDMs can be effectively trained directly on EFD datasets. This eliminates the dependency on extensive CFD simulations or the generation of new EFD data tailored to a specific investigation. Our approach matches the performance of leading-edge CFD models, even in extrapolating conditions, with physical plausibility and minimal human oversight. The validation of our method under various and increasingly complex extrapolating scenarios, employing statistical analyses on the DSYHS EFD dataset and comparisons with state-of-the-art CFD models, underscores the effectiveness of our proposal. Furthermore, we demonstrated that our model can successfully optimize hull resistance when navigating geometric parameters outside the confines of the DSYHS validating our results through leading-edge CFD simulations. This work addresses the limitations of existing methodologies by offering a novel approach more accurate, efficient, cost-effective, flexible, automated, and robust to extrapolation for hull resistance optimization.
- Published
- 2024
- Full Text
- View/download PDF
45. A critical review on groundwater level depletion monitoring based on GIS and data-driven models: Global perspectives and future challenges
- Author
-
Md. Moniruzzaman Monir, Subaran Chandra Sarker, and Abu Reza Md. Towfiqul Islam
- Subjects
Groundwater level ,Depletion ,Data-driven models ,GIS approaches ,Causes and consequences ,Water supply for domestic and industrial purposes ,TD201-500 - Abstract
The present study aims to thoroughly review GWL depletion monitoring studies completed between 2000 and 2023 based on data-driven models and GIS approaches from a global perspective. The review summarizes the details of the reviewed papers, including location, period, time scale, key objective, input parameter, applied model, performance metrics, research gaps, limitations, and depletion rate. The mean rate of GWL depletion varied worldwide from 2.9 ± 1.56 to 1100 ± 33.76 mm/yr using data-driven models and from 7.6 ± 2.98 to 2046 ± 45.27 mm/yr using GIS-based approaches. This study assesses the strength of relationships between various keywords and analyzed co-author networks using the Vos-viewer. It proposes a groundwater development strategy based on the evaluated papers to provide a long-term solution to the water scarcity problem. Overall, this review highlights the existing research gaps and suggests potential future research paths to boost the associated new knowledge and increase the accuracy of the GWL depletion monitoring approaches.
- Published
- 2024
- Full Text
- View/download PDF
46. Assessment of Hull and Propeller Degradation Due to Biofouling Using Tree-Based Models
- Author
-
Nikos Themelis, George Nikolaidis, and Vasilios Zagkas
- Subjects
ship energy efficiency ,hull and propeller biofouling ,data analysis ,data-driven models ,maintenance optimization ,Technology ,Engineering (General). Civil engineering (General) ,TA1-2040 ,Biology (General) ,QH301-705.5 ,Physics ,QC1-999 ,Chemistry ,QD1-999 - Abstract
A hull and propeller biofouling assessment framework is presented and demonstrated using a bulk carrier as a case study corresponding to an operational period of two and a half years. The aim is to support the decision-making process for optimizing maintenance related to hull and propeller cleaning actions. For the degradation assessment, an appropriate key performance indicator is defined comparing the expected shaft power required with the measured power under the same operational conditions. The power prediction models are data-driven based on machine learning algorithms. The process includes feature engineering, filtering, and data smoothing, while an evaluation of regression algorithms of the decision tree family is performed. The extra trees algorithm was selected, presenting a mean absolute percentage error of 1.1%. The analysis incorporates two prediction models corresponding to two different approaches. In the first, the model is employed as a reference performance baseline representing the clean vessel. When applied to a dataset reflecting advanced stages of biofouling, an average power increase of 11.3% is predicted. In the second approach, the model entails a temporal feature enabling the examination of scenarios at different points in time. Considering synthetic data corresponding to 300 days since hull cleaning, it was derived that the fouled vessel required an average 20.5% increase in power.
- Published
- 2024
- Full Text
- View/download PDF
47. Impact of physical and attention mechanisms on U-Net for SST forecasting
- Author
-
Wang, Yong, Zhang, Yiming, and Wang, Gaige
- Published
- 2024
- Full Text
- View/download PDF
48. Comparative Study-Based Data-Driven Models for Lithium-Ion Battery State-of-Charge Estimation.
- Author
-
Hussein, Hossam M., Esoofally, Mustafa, Donekal, Abhishek, Rafin, S M Sajjad Hossain, and Mohammed, Osama
- Subjects
LITHIUM-ion batteries ,ARTIFICIAL neural networks ,ELECTRONIC equipment ,SYSTEMS on a chip ,RANDOM forest algorithms ,IRON ,ELECTRIC vehicles - Abstract
Batteries have been considered a key element in several applications, ranging from grid-scale storage systems through electric vehicles to daily-use small-scale electronic devices. However, excessive charging and discharging will impair their capabilities and could cause their applications to fail catastrophically. Among several diagnostic indices, state-of-charge estimation is essential for evaluating a battery's capabilities. Various approaches have been introduced to reach this target, including white, gray, and black box or data-driven battery models. The main objective of this work is to provide an extensive comparison of currently highly utilized machine learning-based estimation techniques. The paper thoroughly investigates these models' architectures, computational burdens, advantages, drawbacks, and robustness validation. The evaluation's main criteria were based on measurements recorded under various operating conditions at the Energy Systems Research Laboratory (ESRL) at FIU for the eFlex 52.8 V/5.4 kWh lithium iron phosphate battery pack. The primary outcome of this research is that, while the random forest regression (RFR) model emerges as the most effective tool for SoC estimation in lithium-ion batteries, there is potential to enhance the performance of simpler models through strategic adjustments and optimizations. Additionally, the choice of model ultimately depends on the specific requirements of the task at hand, balancing the need for accuracy with the complexity and computational resources available and how it can be merged with other SoC estimation approaches to achieve high precision. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
49. Monthly Streamflow Prediction of the Source Region of the Yellow River Based on Long Short-Term Memory Considering Different Lagged Months.
- Author
-
Chu, Haibo, Wang, Zhuoqi, and Nie, Chong
- Subjects
WATER management ,STREAMFLOW ,WATER use ,BASE flow (Hydrology) ,FORECASTING ,WATER supply - Abstract
Accurate and reliable monthly streamflow prediction plays a crucial role in the scientific allocation and efficient utilization of water resources. In this paper, we proposed a prediction framework that integrates the input variable selection method and Long Short-Term Memory (LSTM). The input selection methods, including autocorrelation function (ACF), partial autocorrelation function (PACF), and time lag cross-correlation (TLCC), were used to analyze the lagged time between variables. Then, the performance of the LSTM model was compared with three other traditional methods. The framework was used to predict monthly streamflow at the Jimai, Maqu, and Tangnaihai stations in the source area of the Yellow River. The results indicated that grid search and cross-validation can improve the efficiency of determining model parameters. The models incorporating ACF, PACF, and TLCC with lagged time are evidently superior to the models using the current variable as the model inputs. Furthermore, the LSTM model, which considers the lagged time, demonstrated better performance in predicting monthly streamflow. The coefficient of determination (R
2 ) improved by an average of 17.46%, 33.94%, and 15.29% for each station, respectively. The integrated framework shows promise in enhancing the accuracy of monthly streamflow prediction, thereby aiding in strategic decision-making for water resources management. [ABSTRACT FROM AUTHOR]- Published
- 2024
- Full Text
- View/download PDF
50. Dynamic modeling of photoacoustic sensor data to classify human blood samples.
- Author
-
Pérez-Pacheco, Argelia, Ramírez-Chavarría, Roberto G., Quispe-Siccha, Rosa M., and Colín-García, Marco P.
- Abstract
The photoacoustic effect is an attractive tool for diagnosis in several biomedical applications. Analyzing photoacoustic signals, however, is challenging to provide qualitative results in an automated way. In this work, we introduce a dynamic modeling scheme of photoacoustic sensor data to classify blood samples according to their physiological status. Thirty-five whole human blood samples were studied with a state-space model estimated by a subspace method. Furthermore, the samples are classified using the model parameters and the linear discriminant analysis algorithm. The classification performance is compared with time- and frequency-domain features and an autoregressive-moving-average model. As a result, the proposed analysis can predict five blood classes: healthy women and men, microcytic and macrocytic anemia, and leukemia. Our findings indicate that the proposed method outperforms conventional signal processing techniques to analyze photoacoustic data for medical diagnosis. Hence, the method is a promising tool in point-of-care devices to detect hematological diseases in clinical scenarios. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
Catalog
Discovery Service for Jio Institute Digital Library
For full access to our library's resources, please sign in.