Back to Search Start Over

Enabling feature-level interpretability in non-linear latent variable models : a synthesis of statistical and machine learning techniques

Authors :
Màˆrtens, Kaspar
Holmes, Chris
Yau, Christopher
Publication Year :
2019
Publisher :
University of Oxford, 2019.

Abstract

Gaining insights into complex high-dimensional data is challenging and typically requires the use of dimensionality reduction methods. These methods let us identify low-dimensional structures embedded within the data that may reveal patterns of interest. In probabilistic models, such low-dimensional structures are captured via latent variables. In biomedical applications, e.g. in computational biology, it is common to use *linear* dimensionality reduction approaches like probabilistic PCA due to their interpretability. In machine learning, however, recently there has been substantial interest in non-linear "black box" latent variable models because of their improved predictive capabilities. In this thesis, we build upon and propose extensions to two non-linear dimensionality reduction frameworks: the Gaussian Process Latent Variable Model (GPLVM) and the Variational Autoencoder (VAE). The former is based on Gaussian Processes, whereas the latter utilises neural networks. It would be desirable if we could combine some interpretable aspects of linear models with the flexibility of modern non-linear probabilistic methods. The goal of our work is to enable feature-level interpretability within non-linear latent variable models, while also incorporating covariate information. As opposed to post-hoc explainability, we aim to directly construct decomposable non-linear models. By combining ideas from classical statistics and embedding these within latent variable models, we investigate two different perspectives on how to aid feature-level transparency. First, we propose approaches to characterise and quantify the contribution of a particular latent variable (or a covariate) for every feature. We achieve this by building upon the notion of functional ANOVA decompositions from classical statistics, and embedding these decompositions within the GPLVM and the VAE frameworks. Second, we investigate another approach towards improving feature-level interpretability, by providing a mechanism to answer the question which features exhibit similar patterns in the VAE framework. We propose to do this by introducing probabilistic clustering structure as part of the decoder network. We demonstrate the utility of our proposed methods on various toy and high-dimensional genomics data sets.

Subjects

Subjects :
519.2
Statistical Machine Learning

Details

Language :
English
Database :
British Library EThOS
Publication Type :
Dissertation/ Thesis
Accession number :
edsble.823481
Document Type :
Electronic Thesis or Dissertation