19 results on '"Michele Lungaroni"'
Search Results
2. Investigating the Physics of Tokamak Global Stability with Interpretable Machine Learning Tools
- Author
-
Andrea Murari, Emmanuele Peluso, Michele Lungaroni, Riccardo Rossi, Michela Gelfusa, and JET Contributors
- Subjects
disruptions ,prediction ,support vector machines ,classification and regression trees (CART) ,ensemble of classifiers ,symbolic regression ,Technology ,Engineering (General). Civil engineering (General) ,TA1-2040 ,Biology (General) ,QH301-705.5 ,Physics ,QC1-999 ,Chemistry ,QD1-999 - Abstract
The inadequacies of basic physics models for disruption prediction have induced the community to increasingly rely on data mining tools. In the last decade, it has been shown how machine learning predictors can achieve a much better performance than those obtained with manually identified thresholds or empirical descriptions of the plasma stability limits. The main criticisms of these techniques focus therefore on two different but interrelated issues: poor “physics fidelity” and limited interpretability. Insufficient “physics fidelity” refers to the fact that the mathematical models of most data mining tools do not reflect the physics of the underlying phenomena. Moreover, they implement a black box approach to learning, which results in very poor interpretability of their outputs. To overcome or at least mitigate these limitations, a general methodology has been devised and tested, with the objective of combining the predictive capability of machine learning tools with the expression of the operational boundary in terms of traditional equations more suited to understanding the underlying physics. The proposed approach relies on the application of machine learning classifiers (such as Support Vector Machines or Classification Trees) and Symbolic Regression via Genetic Programming directly to experimental databases. The results are very encouraging. The obtained equations of the boundary between the safe and disruptive regions of the operational space present almost the same performance as the machine learning classifiers, based on completely independent learning techniques. Moreover, these models possess significantly better predictive power than traditional representations, such as the Hugill or the beta limit. More importantly, they are realistic and intuitive mathematical formulas, which are well suited to supporting theoretical understanding and to benchmarking empirical models. They can also be deployed easily and efficiently in real-time feedback systems.
- Published
- 2020
- Full Text
- View/download PDF
3. Geodesic Distance on Gaussian Manifolds to Reduce the Statistical Errors in the Investigation of Complex Systems
- Author
-
Michele Lungaroni, Andrea Murari, Emmanuele Peluso, Pasqualino Gaudio, and Michela Gelfusa
- Subjects
Electronic computers. Computer science ,QA75.5-76.95 - Abstract
In the last years the reputation of medical, economic, and scientific expertise has been strongly damaged by a series of false predictions and contradictory studies. The lax application of statistical principles has certainly contributed to the uncertainty and loss of confidence in the sciences. Various assumptions, generally held as valid in statistical treatments, have proved their limits. In particular, since some time it has emerged quite clearly that even slightly departures from normality and homoscedasticity can affect significantly classic significance tests. Robust statistical methods have been developed, which can provide much more reliable estimates. On the other hand, they do not address an additional problem typical of the natural sciences, whose data are often the output of delicate measurements. The data can therefore not only be sampled from a nonnormal pdf but also be affected by significant levels of Gaussian additive noise of various amplitude. To tackle this additional source of uncertainty, in this paper it is shown how already developed robust statistical tools can be usefully complemented with the Geodesic Distance on Gaussian Manifolds. This metric is conceptually more appropriate and practically more effective, in handling noise of Gaussian distribution, than the traditional Euclidean distance. The results of a series of systematic numerical tests show the advantages of the proposed approach in all the main aspects of statistical inference, from measures of location and scale to size effects and hypothesis testing. Particularly relevant is the reduction even of 35% in Type II errors, proving the important improvement in power obtained by applying the methods proposed in the paper. It is worth emphasizing that the proposed approach provides a general framework, in which also noise of different statistical distributions can be dealt with.
- Published
- 2019
- Full Text
- View/download PDF
4. Quantifying Total Influence between Variables with Information Theoretic and Machine Learning Techniques
- Author
-
Andrea Murari, Riccardo Rossi, Michele Lungaroni, Pasquale Gaudio, and Michela Gelfusa
- Subjects
machine learning tools ,information theory ,information quality ratio ,total correlations ,encoders ,autoencoders ,Science ,Astrophysics ,QB460-466 ,Physics ,QC1-999 - Abstract
The increasingly sophisticated investigations of complex systems require more robust estimates of the correlations between the measured quantities. The traditional Pearson correlation coefficient is easy to calculate but sensitive only to linear correlations. The total influence between quantities is, therefore, often expressed in terms of the mutual information, which also takes into account the nonlinear effects but is not normalized. To compare data from different experiments, the information quality ratio is, therefore, in many cases, of easier interpretation. On the other hand, both mutual information and information quality ratio are always positive and, therefore, cannot provide information about the sign of the influence between quantities. Moreover, they require an accurate determination of the probability distribution functions of the variables involved. As the quality and amount of data available are not always sufficient to grant an accurate estimation of the probability distribution functions, it has been investigated whether neural computational tools can help and complement the aforementioned indicators. Specific encoders and autoencoders have been developed for the task of determining the total correlation between quantities related by a functional dependence, including information about the sign of their mutual influence. Both their accuracy and computational efficiencies have been addressed in detail, with extensive numerical tests using synthetic data. A careful analysis of the robustness against noise has also been performed. The neural computational tools typically outperform the traditional indicators in practically every respect.
- Published
- 2020
- Full Text
- View/download PDF
5. On the Use of Entropy to Improve Model Selection Criteria
- Author
-
Andrea Murari, Emmanuele Peluso, Francesco Cianfrani, Pasquale Gaudio, and Michele Lungaroni
- Subjects
Model Selection Criteria ,Bayesian Information Criterion (BIC) ,Akaike Information Criterion (AIC) ,Shannon Entropy ,Geodesic Distance ,Science ,Astrophysics ,QB460-466 ,Physics ,QC1-999 - Abstract
The most widely used forms of model selection criteria, the Bayesian Information Criterion (BIC) and the Akaike Information Criterion (AIC), are expressed in terms of synthetic indicators of the residual distribution: the variance and the mean-squared error of the residuals respectively. In many applications in science, the noise affecting the data can be expected to have a Gaussian distribution. Therefore, at the same level of variance and mean-squared error, models, whose residuals are more uniformly distributed, should be favoured. The degree of uniformity of the residuals can be quantified by the Shannon entropy. Including the Shannon entropy in the BIC and AIC expressions improves significantly these criteria. The better performances have been demonstrated empirically with a series of simulations for various classes of functions and for different levels and statistics of the noise. In presence of outliers, a better treatment of the errors, using the Geodesic Distance, has proved essential.
- Published
- 2019
- Full Text
- View/download PDF
6. On the Use of Transfer Entropy to Investigate the Time Horizon of Causal Influences between Signals
- Author
-
Andrea Murari, Michele Lungaroni, Emmanuele Peluso, Pasquale Gaudio, Ernesto Lerche, Luca Garzotti, Michela Gelfusa, and JET Contributors
- Subjects
transfer entropy ,mutual information ,Pearson correlation coefficient ,time series ,causality detection ,sawteeth ,pacing ,ELMs ,pellets ,Science ,Astrophysics ,QB460-466 ,Physics ,QC1-999 - Abstract
Understanding the details of the correlation between time series is an essential step on the route to assessing the causal relation between systems. Traditional statistical indicators, such as the Pearson correlation coefficient and the mutual information, have some significant limitations. More recently, transfer entropy has been proposed as a powerful tool to understand the flow of information between signals. In this paper, the comparative advantages of transfer entropy, for determining the time horizon of causal influence, are illustrated with the help of synthetic data. The technique has been specifically revised for the analysis of synchronization experiments. The investigation of experimental data from thermonuclear plasma diagnostics proves the potential and limitations of the developed approach.
- Published
- 2018
- Full Text
- View/download PDF
7. A practical utility-based but objective approach to model selection for regression in scientific applications.
- Author
-
Andrea Murari, Riccardo Rossi, Luca Spolladore, Michele Lungaroni, Pasquale Gaudio, and Michela Gelfusa
- Published
- 2023
- Full Text
- View/download PDF
8. Information theoretic and neural computational tools for meta-analysis of cumulative databases in the age of Big Physics experiments.
- Author
-
Andrea Murari, Michele Lungaroni, Luca Spolladore, Emmanuele Peluso, Riccardo Rossi, and Michela Gelfusa
- Published
- 2023
- Full Text
- View/download PDF
9. A systemic approach to classification for knowledge discovery with applications to the identification of boundary equations in complex systems.
- Author
-
Andrea Murari, Michela Gelfusa, Michele Lungaroni, Pasquale Gaudio, and Emmanuele Peluso
- Published
- 2022
- Full Text
- View/download PDF
10. Complex networks for the analysis of the synchronization of time series relevant for plasma fusion diagnostics.
- Author
-
Teddy Craciunescu, Andrea Murari, Emmanuele Peluso, Michela Gelfusa, Michele Lungaroni, and Pasqualino Gaudio
- Published
- 2017
- Full Text
- View/download PDF
11. Deriving Realistic Mathematical Models from Support Vector Machines for Scientific Applications.
- Author
-
Andrea Murari, Emmanuele Peluso, Saeed Talebzadeh, Pasqualino Gaudio, Michele Lungaroni, Ondrej Mikulin, Jesús Vega, and Michela Gelfusa
- Published
- 2017
- Full Text
- View/download PDF
12. A Metric to Improve the Robustness of Conformal Predictors in the Presence of Error Bars.
- Author
-
Andrea Murari, Saeed Talebzadeh, Jesús Vega, Emmanuele Peluso, Michela Gelfusa, Michele Lungaroni, and Pasqualino Gaudio
- Published
- 2016
- Full Text
- View/download PDF
13. How to Handle Error Bars in Symbolic Regression for Data Mining in Scientific Applications.
- Author
-
Andrea Murari, Emmanuele Peluso, Michela Gelfusa, Michele Lungaroni, and Pasqualino Gaudio
- Published
- 2015
- Full Text
- View/download PDF
14. A Practical Utility-Based but Objective Approach to Model Selection for Scientific Applications in the Age of Big Data
- Author
-
Andrea Murari, Riccardo Rossi, Luca Spolladore, Michele Lungaroni, Pasquale Gaudio, and Michela Gelfusa
- Abstract
In many fields of science, various types of models are available to describe phenomena, observations and the results of experiments. In the last decades, given the enormous advances of information gathering technologies, also machine learning techniques have been systematically deployed to extract models from the large available databases. However, regardless of their origins, no universal criterion has been found so far to select the most appropriate model given the data. A unique solution is probably a chimera, particularly in applications involving complex systems. Consequently, in this work a utility-based approach is advocated. However, the solutions proposed are not purely subjective but all based on “objective” criteria, rooted in the properties of the data, to preserve generality and to allow comparative assessments of the results. Several methods have been developed and tested, to improve the discrimination capability of basic Bayesian and information theoretic criteria. Both the quality of the fits and the evaluation of model complexity are aspects addressed by the advances proposed. The competitive advantages of the individual alternatives, for both cross sectional data and time series, are clearly identified, together with their most appropriate fields of application. Particular attention has been paid to ensure that the proposed improvements of the criteria are easy to implement in practice, in both confirmatory and exploratory settings. Extensive numerical tests have been performed to support the conceptual and theoretical considerations.
- Published
- 2023
- Full Text
- View/download PDF
15. On the potential of ruled-based machine learning for disruption prediction on JET
- Author
-
Stefan Matejcik, Francesco Romanelli, Augusto Pereira González, Jesús Vega, Bohdan Bieg, Emmanuele Peluso, Vladislav Plyusnin, José Vicente, Alberto Loarte, Michele Lungaroni, Bor Kos, Andrea Murari, Axel Jardin, Rajnikant Makwana, CHIARA MARCHETTO, Choong-Seock Chang, Manuel Garcia-munoz, Department of Physics, and Materials Physics
- Subjects
Boosting (machine learning) ,Computer science ,education ,Machine learning ,computer.software_genre ,114 Physical sciences ,01 natural sciences ,Boosting ,010305 fluids & plasmas ,Noise-based ensembles ,Bagging ,0103 physical sciences ,Classification and regression trees ,General Materials Science ,010306 general physics ,Civil and Structural Engineering ,business.industry ,Mechanical Engineering ,Settore ING-IND/18 - Fisica dei Reattori Nucleari ,Random forests ,Random forest ,Disruptions ,Nuclear Energy and Engineering ,Machine learning predictors ,Artificial intelligence ,business ,computer - Abstract
In the last years, it has become apparent that detecting disruptions with sufficient anticipation time is an essential but not exclusive task of predictors. It is also important that the prediction is accompanied by appropriate qualifications of its reliability and it is formulated in mathematical terms appropriate for the task at hand (mitigation, avoidance, classification etc.). In this paper, a wide series of rule-based predictors, of the Classification and Regression Trees (CART) family, have been compared to assess their relative merits. An original refinement of the training, called noise-based ensembles, has allowed not only to obtain significantly better performance but also to increase the interpretability of the results. The final predictors can indeed be represented by a tree or a series of specific and clear rules. Such performance has been proved by analysing large databases of shots on JET with both the carbon wall and the ITER Like Wall. In terms of performance, the developed tools are therefore very competitive with other machine learning techniques, with the specificity of formulating the final models in terms of trees and simple rules.
- Published
- 2018
- Full Text
- View/download PDF
16. Determining the prediction limits of models and classifiers with applications for disruption prediction in JET
- Author
-
Alexander Lukin, Pasqualino Gaudio, Stefan Matejcik, Soare Sorin, Francesco Romanelli, Emilio Blanco, Bohdan Bieg, Emmanuele Peluso, Vladislav Plyusnin, José Vicente, Alberto Loarte, Michele Lungaroni, Andrea Murari, Rajnikant Makwana, CHIARA MARCHETTO, Marco Wischmeier, Choong-Seock Chang, Aneta Gójska, and Manuel Garcia-munoz
- Subjects
Nuclear and High Energy Physics ,Jet (fluid) ,Exploit ,ELMs ,Mode (statistics) ,Feedback loop ,disruptions ,Condensed Matter Physics ,01 natural sciences ,Signal on ,Settore FIS/04 - Fisica Nucleare e Subnucleare ,010305 fluids & plasmas ,Reliability engineering ,conditional entropy ,prediction factor ,predictability ,0103 physical sciences ,Numerical tests ,Predictability ,010306 general physics ,sawteeth - Abstract
Understanding the many aspects of tokamak physics requires the development of quite sophisticated models. Moreover, in the operation of the devices, prediction of the future evolution of discharges can be of crucial importance, particularly in the case of the prediction of disruptions, which can cause serious damage to various parts of the machine. The determination of the limits of predictability is therefore an important issue for modelling, classifying and forecasting. In all these cases, once a certain level of performance has been reached, the question typically arises as to whether all the information available in the data has been exploited, or whether there are still margins for improvement of the tools being developed. In this paper, a theoretical information approach is proposed to address this issue. The excellent properties of the developed indicator, called the prediction factor (PF), have been proved with the help of a series of numerical tests. Its application to some typical behaviour relating to macroscopic instabilities in tokamaks has shown very positive results. The prediction factor has also been used to assess the performance of disruption predictors running in real time in the JET system, including the one systematically deployed in the feedback loop for mitigation purposes. The main conclusion is that the most advanced predictors basically exploit all the information contained in the locked mode signal on which they are based. Therefore, qualitative improvements in disruption prediction performance in JET would need the processing of additional signals, probably profiles.
- Published
- 2017
- Full Text
- View/download PDF
17. How to assess the efficiency of synchronization experiments in tokamaks
- Author
-
Alexander Lukin, Pasqualino Gaudio, Stefan Matejcik, Soare Sorin, Francesco Romanelli, Emilio Blanco, Teddy CRACIUNESCU, Bohdan Bieg, Luca Garzotti, Emmanuele Peluso, Vladislav Plyusnin, José Vicente, Alberto Loarte, Michele Lungaroni, Andrea Murari, Rajnikant Makwana, CHIARA MARCHETTO, Marco Wischmeier, Choong-Seock Chang, Aneta Gójska, Manuel Garcia-munoz, and Frigione, D.
- Subjects
recurrence plot ,Nuclear and High Energy Physics ,Tokamak ,Computer science ,Combined use ,Settore FIS/01 - Fisica Sperimentale ,transfer entropy ,Perturbation (astronomy) ,Time resolution ,recurrence plots ,Granger causality ,ELM pacing ,Condensed Matter Physics ,01 natural sciences ,010305 fluids & plasmas ,law.invention ,law ,Control theory ,0103 physical sciences ,Transfer entropy ,010306 general physics - Abstract
Control of instabilities such as ELMs and sawteeth is considered an important ingredient in the development of reactor-relevant scenarios. Various forms of ELM pacing have been tried in the past to influence their behavior using external perturbations. One of the main problems with these synchronization experiments resides in the fact that ELMs are periodic or quasi-periodic in nature. Therefore, after any pulsed perturbation, if one waits long enough, an ELM is always bound to occur. To evaluate the effectiveness of ELM pacing techniques, it is crucial to determine an appropriate interval over which they can have a real influence and an effective triggering capability. In this paper, three independent statistical methods are described to address this issue: Granger causality, transfer entropy and recurrence plots. The obtained results for JET with the ITER-like wall (ILW) indicate that the proposed techniques agree very well and provide much better estimates than the traditional heuristic criteria reported in the literature. Moreover, their combined use allows for the improvement of the time resolution of the assessment and determination of the efficiency of the pellet triggering in different phases of the same discharge. Therefore, the developed methods can be used to provide a quantitative and statistically robust estimate of the triggering efficiency of ELM pacing under realistic experimental conditions.
- Published
- 2016
- Full Text
- View/download PDF
18. On determining the prediction limits of mathematical models for time series
- Author
-
Alexander Lukin, Pasqualino Gaudio, Stefan Matejcik, Soare Sorin, Francesco Romanelli, Emilio Blanco, Bohdan Bieg, Emmanuele Peluso, Vladislav Plyusnin, José Vicente, Alberto Loarte, Michele Lungaroni, Andrea Murari, Rajnikant Makwana, CHIARA MARCHETTO, Marco Wischmeier, Choong-Seock Chang, Aneta Gójska, and Manuel Garcia-munoz
- Subjects
Calibration and fitting methods ,Series (mathematics) ,Mathematical model ,Computer science ,Cluster finding ,01 natural sciences ,Industrial engineering ,Settore FIS/07 - Fisica Applicata(Beni Culturali, Ambientali, Biol.e Medicin) ,010305 fluids & plasmas ,Scientific analysis ,Pattern recognition ,Data processing methods ,0103 physical sciences ,Analysis and statistical methods ,Predictability ,010306 general physics ,Instrumentation ,Mathematical Physics - Abstract
Prediction is one of the main objectives of scientific analysis and it refers to both modelling and forecasting. The determination of the limits of predictability is an important issue of both theoretical and practical relevance. In the case of modelling time series, reached a certain level in performance in either modelling or prediction, it is often important to assess whether all the information available in the data has been exploited or whether there are still margins for improvement of the tools being developed. In this paper, an information theoretic approach is proposed to address this issue and quantify the quality of the models and/or predictions. The excellent properties of the proposed indicator have been proved with the help of a systematic series of numerical tests and a concrete example of extreme relevance for nuclear fusion.
- Published
- 2016
- Full Text
- View/download PDF
19. Application of transfer entropy to causality detection and synchronization experiments in tokamaks
- Author
-
Alexander Lukin, Pasqualino Gaudio, Stefan Matejcik, Soare Sorin, Francesco Romanelli, FABIO PISANO, Emilio Blanco, Bohdan Bieg, Luca Garzotti, Emmanuele Peluso, Michela Gelfusa, Vladislav Plyusnin, José Vicente, Alberto Loarte, Michele Lungaroni, Andrea Murari, Rajnikant Makwana, CHIARA MARCHETTO, Marco Wischmeier, Choong-Seock Chang, Aneta Gójska, and Manuel Garcia-munoz
- Subjects
Nuclear and High Energy Physics ,Tokamak ,Series (mathematics) ,Computer science ,synchronization experiments ,transfer entropy ,food and beverages ,disruption precursors ,Condensed Matter Physics ,ELM pacing ,01 natural sciences ,010305 fluids & plasmas ,Task (project management) ,law.invention ,Causality (physics) ,causality detection ,Theoretical physics ,Nonlinear system ,law ,0103 physical sciences ,Synchronization (computer science) ,Transfer entropy ,Statistical physics ,010306 general physics - Abstract
Determination of causal-effect relationships can be a difficult task even in the analysis of time series. This is particularly true in the case of complex, nonlinear systems affected by significant levels of noise. Causality can be modelled as a flow of information between systems, allowing to better predict the behaviour of a phenomenon on the basis of the knowledge of the one causing it. Therefore, information theoretic tools, such as the transfer entropy, have been used in various disciplines to quantify the causal relationship between events. In this paper, Transfer Entropy is applied to determining the information relationship between various phenomena in Tokamaks. The proposed approach provides unique insight about information causality in difficult situations, such as the link between sawteeth and ELMs and ELM pacing experiments. The application to the determination of disruption causes, and therefore to the classification of disruption types, looks also very promising. The obtained results indicate that the proposed method can provide a quantitative and statistically sound criterion to address the causal-effect relationships in various difficult and ambiguous situations if the data is of sufficient quality.
- Published
- 2015
- Full Text
- View/download PDF
Catalog
Discovery Service for Jio Institute Digital Library
For full access to our library's resources, please sign in.