12 results on '"Montavon, Grégoire"'
Search Results
2. Layer-Wise Relevance Propagation for Neural Networks with Local Renormalization Layers
- Author
-
Binder, Alexander, Montavon, Grégoire, Lapuschkin, Sebastian, Müller, Klaus-Robert, Samek, Wojciech, Hutchison, David, Series editor, Kanade, Takeo, Series editor, Kittler, Josef, Series editor, Kleinberg, Jon M., Series editor, Mattern, Friedemann, Series editor, Mitchell, John C., Series editor, Naor, Moni, Series editor, Pandu Rangan, C., Series editor, Steffen, Bernhard, Series editor, Terzopoulos, Demetri, Series editor, Tygar, Doug, Series editor, Weikum, Gerhard, Series editor, Villa, Alessandro E.P., editor, Masulli, Paolo, editor, and Pons Rivero, Antonio Javier, editor
- Published
- 2016
- Full Text
- View/download PDF
3. Deep learning for surrogate modelling of 2D mantle convection
- Author
-
Agarwal, Siddhant, Tosi, Nicola, Kessel, P, Breuer, Doris, and Montavon, Grégoire
- Subjects
Earth and Planetary Astrophysics (astro-ph.EP) ,FOS: Computer and information sciences ,Computer Science - Machine Learning ,Neural Networks ,Mantle Convection ,Fluid Dynamics (physics.flu-dyn) ,FOS: Physical sciences ,Fluid Dynamics ,Physics - Fluid Dynamics ,Machine Learning (cs.LG) ,Geophysics (physics.geo-ph) ,Physics - Geophysics ,Machine Learning ,Surrogate Modelling ,Astrophysics - Earth and Planetary Astrophysics - Abstract
Traditionally, 1D models based on scaling laws have been used to parameterized convective heat transfer rocks in the interior of terrestrial planets like Earth, Mars, Mercury and Venus to tackle the computational bottleneck of high-fidelity forward runs in 2D or 3D. However, these are limited in the amount of physics they can model (e.g. depth dependent material properties) and predict only mean quantities such as the mean mantle temperature. We recently showed that feedforward neural networks (FNN) trained using a large number of 2D simulations can overcome this limitation and reliably predict the evolution of entire 1D laterally-averaged temperature profile in time for complex models. We now extend that approach to predict the full 2D temperature field, which contains more information in the form of convection structures such as hot plumes and cold downwellings. Using a dataset of 10,525 two-dimensional simulations of the thermal evolution of the mantle of a Mars-like planet, we show that deep learning techniques can produce reliable parameterized surrogates (i.e. surrogates that predict state variables such as temperature based only on parameters) of the underlying partial differential equations. We first use convolutional autoencoders to compress the temperature fields by a factor of 142 and then use FNN and long-short term memory networks (LSTM) to predict the compressed fields. On average, the FNN predictions are 99.30% and the LSTM predictions are 99.22% accurate with respect to unseen simulations. Proper orthogonal decomposition (POD) of the LSTM and FNN predictions shows that despite a lower mean absolute relative accuracy, LSTMs capture the flow dynamics better than FNNs. When summed, the POD coefficients from FNN predictions and from LSTM predictions amount to 96.51% and 97.66% relative to the coefficients of the original simulations, respectively.
- Published
- 2021
- Full Text
- View/download PDF
4. Learning high dimensional surrogates from mantle convection simulations
- Author
-
Agarwal, Siddhant, Tosi, Nicola, Kessel, P, Breuer, Doris, Padovan, Sebastiano, and Montavon, Grégoire
- Subjects
Machine Learning ,Neural Networks ,Mantle Convection ,Fluid Dynamics ,Surrogate Modelling - Published
- 2020
5. Using machine learning to predict 1D steady-state temperature profiles from compressible mantle convection simulations
- Author
-
Agarwal, Siddhant, Tosi, Nicola, Breuer, Doris, Kessel, Pan, and Montavon, Grégoire
- Subjects
Machine Learning ,Planetenphysik ,Neural Networks ,Mantle Convection ,Institut für Planetenforschung ,Fluid Dynamics ,Surrogate Modelling - Abstract
Thermal evolution simulations of planetary mantles in 2D and 3D are computationally intensive. A low-fidelity alternative is to use scaling laws based on boundary-layer theory to express Nusselt Number (Nu) as a function of Rayleigh Number (Ra). Such a Ra-Nu relation can be used to run `0D' parametrized evolution models by solving a simple energy balance equation. Yet scaling relations are available only for simple flows that cannot capture the full complexity of mantle dynamics. We propose leveraging Machine Learning to find a higher-dimensional mapping from five different parameters to the entire 1D temperature profile. The parameters are Ra, internal heating Ra, dissipation number and the maximum viscosity contrast between top and bottom due to temperature and pressure. We train a Neural Network (NN) to take these inputs and predict the resulting steady-state temperature profile. The training data comes from a subset of 20,000 compressible simulations on a 2D cylindrical grid. This results in predictions with an average error of 1.6% on the test set. The NN can potentially be used to build a 1D evolution model by stacking several steady-state temperature profiles together, with each prediction serving as an input at the next time-step.
- Published
- 2019
6. Using Deep Learning neural networks to predict the interior composition of exoplanets
- Author
-
Baumeister, Philipp, Padovan, Sebastiano, Tosi, Nicola, and Montavon, Grégoire
- Subjects
Extrasolare Planeten und Atmosphären ,machine learning ,Planetenphysik ,exoplanets ,Astrophysics::Earth and Planetary Astrophysics ,neural networks ,interior structure - Abstract
One of the main goals in exoplanetary science is the interior characterization of observed exoplanets. A common approach to characterize the interior of a known exoplanet is the use of numerical models to compute an interior structure which complies with the measured mass and radius of the planet (Sotin et al. 2017, Seager et al. 2007). With only these two observables, possible solutions tend to be highly degenerate, with multiple, qualitatively different interior compositions that can match the observations equally well. Other potential observables include the Love number k2 (bearing information on the mass concentration in the interior of the planet), and the elemental abundances of the host star, which may be representative of those of the planet. We explore the application of a deep learning neural network to the interior characterization of exoplanets. We employ a simple 1D structure model to construct a large training set of sub-Neptunian exoplanets up to 20 Earth-masses. A model planet consists of five layers: an iron-rich core, a lower and upper silicate mantle, a water ice layer, and a gaseous H/He envelope. The size of each layer is constrained by prescribed mass fractions. Using a feedforward neural network trained on a large dataset of such modelled planets, we show that we can reasonably well predict the original model input parameters (core, mantle, ice layer and atmosphere mass fractions) from just mass, radius and the fluid Love number k2.
- Published
- 2018
7. "What is relevant in a text document?": An interpretable machine learning approach.
- Author
-
Arras, Leila, Horn, Franziska, Montavon, Grégoire, Müller, Klaus-Robert, and Samek, Wojciech
- Subjects
MACHINE learning ,MATHEMATICAL convolutions ,ARTIFICIAL neural networks ,SUPPORT vector machines ,DATA mining - Abstract
Text documents can be described by a number of abstract concepts such as semantic category, writing style, or sentiment. Machine learning (ML) models have been trained to automatically map documents to these abstract concepts, allowing to annotate very large text collections, more than could be processed by a human in a lifetime. Besides predicting the text’s category very accurately, it is also highly desirable to understand how and why the categorization process takes place. In this paper, we demonstrate that such understanding can be achieved by tracing the classification decision back to individual words using layer-wise relevance propagation (LRP), a recently developed technique for explaining predictions of complex non-linear classifiers. We train two word-based ML models, a convolutional neural network (CNN) and a bag-of-words SVM classifier, on a topic categorization task and adapt the LRP method to decompose the predictions of these models onto words. Resulting scores indicate how much individual words contribute to the overall classification decision. This enables one to distill relevant information from text documents without an explicit semantic information extraction step. We further use the word-wise relevance scores for generating novel vector-based document representations which capture semantic information. Based on these document vectors, we introduce a measure of model explanatory power and show that, although the SVM and CNN models perform similarly in terms of classification accuracy, the latter exhibits a higher level of explainability which makes it more comprehensible for humans and potentially more useful for other applications. [ABSTRACT FROM AUTHOR]
- Published
- 2017
- Full Text
- View/download PDF
8. Learning domain invariant representations by joint Wasserstein distance minimization.
- Author
-
Andéol, Léo, Kawakami, Yusei, Wada, Yuichiro, Kanamori, Takafumi, Müller, Klaus-Robert, and Montavon, Grégoire
- Subjects
- *
MACHINE learning , *SET theory , *SUPERVISED learning - Abstract
Domain shifts in the training data are common in practical applications of machine learning; they occur for instance when the data is coming from different sources. Ideally, a ML model should work well independently of these shifts, for example, by learning a domain-invariant representation. However, common ML losses do not give strong guarantees on how consistently the ML model performs for different domains, in particular, whether the model performs well on a domain at the expense of its performance on another domain. In this paper, we build new theoretical foundations for this problem, by contributing a set of mathematical relations between classical losses for supervised ML and the Wasserstein distance in joint space (i.e. representation and output space). We show that classification or regression losses, when combined with a GAN-type discriminator between domains, form an upper-bound to the true Wasserstein distance between domains. This implies a more invariant representation and also more stable prediction performance across domains. Theoretical results are corroborated empirically on several image datasets. Our proposed approach systematically produces the highest minimum classification accuracy across domains, and the most invariant representation. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
9. Interpretable Deep Learning in Drug Discovery
- Author
-
Preuer, Kristina, Klambauer, Günter, Rippmann, Friedrich, Hochreiter, Sepp, Unterthiner, Thomas, Goos, Gerhard, Founding Editor, Hartmanis, Juris, Founding Editor, Bertino, Elisa, Editorial Board Member, Gao, Wen, Editorial Board Member, Steffen, Bernhard, Editorial Board Member, Woeginger, Gerhard, Editorial Board Member, Yung, Moti, Editorial Board Member, Samek, Wojciech, editor, Montavon, Grégoire, editor, Vedaldi, Andrea, editor, Hansen, Lars Kai, editor, and Müller, Klaus-Robert, editor
- Published
- 2019
- Full Text
- View/download PDF
10. NeuralHydrology – Interpreting LSTMs in Hydrology
- Author
-
Kratzert, Frederik, Herrnegger, Mathew, Klotz, Daniel, Hochreiter, Sepp, Klambauer, Günter, Goos, Gerhard, Founding Editor, Hartmanis, Juris, Founding Editor, Bertino, Elisa, Editorial Board Member, Gao, Wen, Editorial Board Member, Steffen, Bernhard, Editorial Board Member, Woeginger, Gerhard, Editorial Board Member, Yung, Moti, Editorial Board Member, Samek, Wojciech, editor, Montavon, Grégoire, editor, Vedaldi, Andrea, editor, Hansen, Lars Kai, editor, and Müller, Klaus-Robert, editor
- Published
- 2019
- Full Text
- View/download PDF
11. Understanding Neural Networks via Feature Visualization: A Survey
- Author
-
Nguyen, Anh, Yosinski, Jason, Clune, Jeff, Goos, Gerhard, Founding Editor, Hartmanis, Juris, Founding Editor, Bertino, Elisa, Editorial Board Member, Gao, Wen, Editorial Board Member, Steffen, Bernhard, Editorial Board Member, Woeginger, Gerhard, Editorial Board Member, Yung, Moti, Editorial Board Member, Samek, Wojciech, editor, Montavon, Grégoire, editor, Vedaldi, Andrea, editor, Hansen, Lars Kai, editor, and Müller, Klaus-Robert, editor
- Published
- 2019
- Full Text
- View/download PDF
12. Towards Explainable Artificial Intelligence
- Author
-
Samek, Wojciech, Müller, Klaus-Robert, Goos, Gerhard, Founding Editor, Hartmanis, Juris, Founding Editor, Bertino, Elisa, Editorial Board Member, Gao, Wen, Editorial Board Member, Steffen, Bernhard, Editorial Board Member, Woeginger, Gerhard, Editorial Board Member, Yung, Moti, Editorial Board Member, Samek, Wojciech, editor, Montavon, Grégoire, editor, Vedaldi, Andrea, editor, Hansen, Lars Kai, editor, and Müller, Klaus-Robert, editor
- Published
- 2019
- Full Text
- View/download PDF
Catalog
Discovery Service for Jio Institute Digital Library
For full access to our library's resources, please sign in.