8 results on '"Michele Lungaroni"'
Search Results
2. Investigating the Physics of Tokamak Global Stability with Interpretable Machine Learning Tools
- Author
-
Andrea Murari, Emmanuele Peluso, Michele Lungaroni, Riccardo Rossi, Michela Gelfusa, and JET Contributors
- Subjects
disruptions ,prediction ,support vector machines ,classification and regression trees (CART) ,ensemble of classifiers ,symbolic regression ,Technology ,Engineering (General). Civil engineering (General) ,TA1-2040 ,Biology (General) ,QH301-705.5 ,Physics ,QC1-999 ,Chemistry ,QD1-999 - Abstract
The inadequacies of basic physics models for disruption prediction have induced the community to increasingly rely on data mining tools. In the last decade, it has been shown how machine learning predictors can achieve a much better performance than those obtained with manually identified thresholds or empirical descriptions of the plasma stability limits. The main criticisms of these techniques focus therefore on two different but interrelated issues: poor “physics fidelity” and limited interpretability. Insufficient “physics fidelity” refers to the fact that the mathematical models of most data mining tools do not reflect the physics of the underlying phenomena. Moreover, they implement a black box approach to learning, which results in very poor interpretability of their outputs. To overcome or at least mitigate these limitations, a general methodology has been devised and tested, with the objective of combining the predictive capability of machine learning tools with the expression of the operational boundary in terms of traditional equations more suited to understanding the underlying physics. The proposed approach relies on the application of machine learning classifiers (such as Support Vector Machines or Classification Trees) and Symbolic Regression via Genetic Programming directly to experimental databases. The results are very encouraging. The obtained equations of the boundary between the safe and disruptive regions of the operational space present almost the same performance as the machine learning classifiers, based on completely independent learning techniques. Moreover, these models possess significantly better predictive power than traditional representations, such as the Hugill or the beta limit. More importantly, they are realistic and intuitive mathematical formulas, which are well suited to supporting theoretical understanding and to benchmarking empirical models. They can also be deployed easily and efficiently in real-time feedback systems.
- Published
- 2020
- Full Text
- View/download PDF
3. Geodesic Distance on Gaussian Manifolds to Reduce the Statistical Errors in the Investigation of Complex Systems
- Author
-
Michele Lungaroni, Andrea Murari, Emmanuele Peluso, Pasqualino Gaudio, and Michela Gelfusa
- Subjects
Electronic computers. Computer science ,QA75.5-76.95 - Abstract
In the last years the reputation of medical, economic, and scientific expertise has been strongly damaged by a series of false predictions and contradictory studies. The lax application of statistical principles has certainly contributed to the uncertainty and loss of confidence in the sciences. Various assumptions, generally held as valid in statistical treatments, have proved their limits. In particular, since some time it has emerged quite clearly that even slightly departures from normality and homoscedasticity can affect significantly classic significance tests. Robust statistical methods have been developed, which can provide much more reliable estimates. On the other hand, they do not address an additional problem typical of the natural sciences, whose data are often the output of delicate measurements. The data can therefore not only be sampled from a nonnormal pdf but also be affected by significant levels of Gaussian additive noise of various amplitude. To tackle this additional source of uncertainty, in this paper it is shown how already developed robust statistical tools can be usefully complemented with the Geodesic Distance on Gaussian Manifolds. This metric is conceptually more appropriate and practically more effective, in handling noise of Gaussian distribution, than the traditional Euclidean distance. The results of a series of systematic numerical tests show the advantages of the proposed approach in all the main aspects of statistical inference, from measures of location and scale to size effects and hypothesis testing. Particularly relevant is the reduction even of 35% in Type II errors, proving the important improvement in power obtained by applying the methods proposed in the paper. It is worth emphasizing that the proposed approach provides a general framework, in which also noise of different statistical distributions can be dealt with.
- Published
- 2019
- Full Text
- View/download PDF
4. Quantifying Total Influence between Variables with Information Theoretic and Machine Learning Techniques
- Author
-
Andrea Murari, Riccardo Rossi, Michele Lungaroni, Pasquale Gaudio, and Michela Gelfusa
- Subjects
machine learning tools ,information theory ,information quality ratio ,total correlations ,encoders ,autoencoders ,Science ,Astrophysics ,QB460-466 ,Physics ,QC1-999 - Abstract
The increasingly sophisticated investigations of complex systems require more robust estimates of the correlations between the measured quantities. The traditional Pearson correlation coefficient is easy to calculate but sensitive only to linear correlations. The total influence between quantities is, therefore, often expressed in terms of the mutual information, which also takes into account the nonlinear effects but is not normalized. To compare data from different experiments, the information quality ratio is, therefore, in many cases, of easier interpretation. On the other hand, both mutual information and information quality ratio are always positive and, therefore, cannot provide information about the sign of the influence between quantities. Moreover, they require an accurate determination of the probability distribution functions of the variables involved. As the quality and amount of data available are not always sufficient to grant an accurate estimation of the probability distribution functions, it has been investigated whether neural computational tools can help and complement the aforementioned indicators. Specific encoders and autoencoders have been developed for the task of determining the total correlation between quantities related by a functional dependence, including information about the sign of their mutual influence. Both their accuracy and computational efficiencies have been addressed in detail, with extensive numerical tests using synthetic data. A careful analysis of the robustness against noise has also been performed. The neural computational tools typically outperform the traditional indicators in practically every respect.
- Published
- 2020
- Full Text
- View/download PDF
5. On the Use of Entropy to Improve Model Selection Criteria
- Author
-
Andrea Murari, Emmanuele Peluso, Francesco Cianfrani, Pasquale Gaudio, and Michele Lungaroni
- Subjects
Model Selection Criteria ,Bayesian Information Criterion (BIC) ,Akaike Information Criterion (AIC) ,Shannon Entropy ,Geodesic Distance ,Science ,Astrophysics ,QB460-466 ,Physics ,QC1-999 - Abstract
The most widely used forms of model selection criteria, the Bayesian Information Criterion (BIC) and the Akaike Information Criterion (AIC), are expressed in terms of synthetic indicators of the residual distribution: the variance and the mean-squared error of the residuals respectively. In many applications in science, the noise affecting the data can be expected to have a Gaussian distribution. Therefore, at the same level of variance and mean-squared error, models, whose residuals are more uniformly distributed, should be favoured. The degree of uniformity of the residuals can be quantified by the Shannon entropy. Including the Shannon entropy in the BIC and AIC expressions improves significantly these criteria. The better performances have been demonstrated empirically with a series of simulations for various classes of functions and for different levels and statistics of the noise. In presence of outliers, a better treatment of the errors, using the Geodesic Distance, has proved essential.
- Published
- 2019
- Full Text
- View/download PDF
6. On the Use of Transfer Entropy to Investigate the Time Horizon of Causal Influences between Signals
- Author
-
Andrea Murari, Michele Lungaroni, Emmanuele Peluso, Pasquale Gaudio, Ernesto Lerche, Luca Garzotti, Michela Gelfusa, and JET Contributors
- Subjects
transfer entropy ,mutual information ,Pearson correlation coefficient ,time series ,causality detection ,sawteeth ,pacing ,ELMs ,pellets ,Science ,Astrophysics ,QB460-466 ,Physics ,QC1-999 - Abstract
Understanding the details of the correlation between time series is an essential step on the route to assessing the causal relation between systems. Traditional statistical indicators, such as the Pearson correlation coefficient and the mutual information, have some significant limitations. More recently, transfer entropy has been proposed as a powerful tool to understand the flow of information between signals. In this paper, the comparative advantages of transfer entropy, for determining the time horizon of causal influence, are illustrated with the help of synthetic data. The technique has been specifically revised for the analysis of synchronization experiments. The investigation of experimental data from thermonuclear plasma diagnostics proves the potential and limitations of the developed approach.
- Published
- 2018
- Full Text
- View/download PDF
7. On the potential of ruled-based machine learning for disruption prediction on JET
- Author
-
Stefan Matejcik, Francesco Romanelli, Augusto Pereira González, Jesús Vega, Bohdan Bieg, Emmanuele Peluso, Vladislav Plyusnin, José Vicente, Alberto Loarte, Michele Lungaroni, Bor Kos, Andrea Murari, Axel Jardin, Rajnikant Makwana, CHIARA MARCHETTO, Choong-Seock Chang, Manuel Garcia-munoz, Department of Physics, and Materials Physics
- Subjects
Boosting (machine learning) ,Computer science ,education ,Machine learning ,computer.software_genre ,114 Physical sciences ,01 natural sciences ,Boosting ,010305 fluids & plasmas ,Noise-based ensembles ,Bagging ,0103 physical sciences ,Classification and regression trees ,General Materials Science ,010306 general physics ,Civil and Structural Engineering ,business.industry ,Mechanical Engineering ,Settore ING-IND/18 - Fisica dei Reattori Nucleari ,Random forests ,Random forest ,Disruptions ,Nuclear Energy and Engineering ,Machine learning predictors ,Artificial intelligence ,business ,computer - Abstract
In the last years, it has become apparent that detecting disruptions with sufficient anticipation time is an essential but not exclusive task of predictors. It is also important that the prediction is accompanied by appropriate qualifications of its reliability and it is formulated in mathematical terms appropriate for the task at hand (mitigation, avoidance, classification etc.). In this paper, a wide series of rule-based predictors, of the Classification and Regression Trees (CART) family, have been compared to assess their relative merits. An original refinement of the training, called noise-based ensembles, has allowed not only to obtain significantly better performance but also to increase the interpretability of the results. The final predictors can indeed be represented by a tree or a series of specific and clear rules. Such performance has been proved by analysing large databases of shots on JET with both the carbon wall and the ITER Like Wall. In terms of performance, the developed tools are therefore very competitive with other machine learning techniques, with the specificity of formulating the final models in terms of trees and simple rules.
- Published
- 2018
- Full Text
- View/download PDF
8. Determining the prediction limits of models and classifiers with applications for disruption prediction in JET
- Author
-
Alexander Lukin, Pasqualino Gaudio, Stefan Matejcik, Soare Sorin, Francesco Romanelli, Emilio Blanco, Bohdan Bieg, Emmanuele Peluso, Vladislav Plyusnin, José Vicente, Alberto Loarte, Michele Lungaroni, Andrea Murari, Rajnikant Makwana, CHIARA MARCHETTO, Marco Wischmeier, Choong-Seock Chang, Aneta Gójska, and Manuel Garcia-munoz
- Subjects
Nuclear and High Energy Physics ,Jet (fluid) ,Exploit ,ELMs ,Mode (statistics) ,Feedback loop ,disruptions ,Condensed Matter Physics ,01 natural sciences ,Signal on ,Settore FIS/04 - Fisica Nucleare e Subnucleare ,010305 fluids & plasmas ,Reliability engineering ,conditional entropy ,prediction factor ,predictability ,0103 physical sciences ,Numerical tests ,Predictability ,010306 general physics ,sawteeth - Abstract
Understanding the many aspects of tokamak physics requires the development of quite sophisticated models. Moreover, in the operation of the devices, prediction of the future evolution of discharges can be of crucial importance, particularly in the case of the prediction of disruptions, which can cause serious damage to various parts of the machine. The determination of the limits of predictability is therefore an important issue for modelling, classifying and forecasting. In all these cases, once a certain level of performance has been reached, the question typically arises as to whether all the information available in the data has been exploited, or whether there are still margins for improvement of the tools being developed. In this paper, a theoretical information approach is proposed to address this issue. The excellent properties of the developed indicator, called the prediction factor (PF), have been proved with the help of a series of numerical tests. Its application to some typical behaviour relating to macroscopic instabilities in tokamaks has shown very positive results. The prediction factor has also been used to assess the performance of disruption predictors running in real time in the JET system, including the one systematically deployed in the feedback loop for mitigation purposes. The main conclusion is that the most advanced predictors basically exploit all the information contained in the locked mode signal on which they are based. Therefore, qualitative improvements in disruption prediction performance in JET would need the processing of additional signals, probably profiles.
- Published
- 2017
- Full Text
- View/download PDF
Catalog
Discovery Service for Jio Institute Digital Library
For full access to our library's resources, please sign in.