1. Interpreting Internal Activation Patterns in Deep Temporal Neural Networks by Finding Prototypes
- Author
-
Sohee Cho, Wonjoon Chang, Jaesik Choi, and Ginkyeng Lee
- Subjects
Artificial neural network ,business.industry ,Computer science ,Temporal Regions ,Maximum mean discrepancy ,Segmentation ,Sequential data ,Pattern recognition ,Artificial intelligence ,Selection method ,business ,Communication channel ,Interpretability - Abstract
Deep neural networks have demonstrated competitive performance in classification tasks for sequential data. However, it remains difficult to understand which temporal patterns the internal channels of deep neural networks capture for decision-making in sequential data. To address this issue, we propose a new framework with which to visualize temporal representations learned in deep neural networks without hand-crafted segmentation labels. Given input data, our framework extracts highly activated temporal regions that contribute to activating internal nodes and characterizes such regions by prototype selection method based on Maximum Mean Discrepancy. Representative temporal patterns referred to here as Prototypes of Temporally Activated Patterns (PTAP) provide core examples of subsequences in the sequential data for interpretability. We also analyze the role of each channel by Value-LRP plots using representative prototypes and the distribution of the input attribution. Input attribution plots give visual information to recognize the shapes focused on by the channel for decision-making.
- Published
- 2021
- Full Text
- View/download PDF