Back to Search Start Over

Higher-Order Explanations of Graph Neural Networks via Relevant Walks

Authors :
Kristof T. Schütt
Jonas Lederer
Oliver Eberle
Shinichi Nakajima
Klaus-Robert Mueller
Grégoire Montavon
Thomas Schnake
Source :
IEEE Transactions on Pattern Analysis and Machine Intelligence. 44:7581-7596
Publication Year :
2022
Publisher :
Institute of Electrical and Electronics Engineers (IEEE), 2022.

Abstract

Graph Neural Networks (GNNs) are a popular approach for predicting graph structured data. As GNNs tightly entangle the input graph into the neural network structure, common explainable AI approaches are not applicable. To a large extent, GNNs have remained black-boxes for the user so far. In this paper, we show that GNNs can in fact be naturally explained using higher-order expansions, i.e. by identifying groups of edges that jointly contribute to the prediction. Practically, we find that such explanations can be extracted using a nested attribution scheme, where existing techniques such as layer-wise relevance propagation (LRP) can be applied at each step. The output is a collection of walks into the input graph that are relevant for the prediction. Our novel explanation method, which we denote by GNN-LRP, is applicable to a broad range of graph neural networks and lets us extract practically relevant insights on sentiment analysis of text data, structure-property relationships in quantum chemistry, and image classification.<br />Comment: 14 pages + 6 pages supplement

Details

ISSN :
19393539 and 01628828
Volume :
44
Database :
OpenAIRE
Journal :
IEEE Transactions on Pattern Analysis and Machine Intelligence
Accession number :
edsair.doi.dedup.....4455cd57f236f14fb69022db8a0bd52b