Back to Search Start Over

Inseq: An Interpretability Toolkit for Sequence Generation Models

Authors :
Sarti, Gabriele
Feldhus, Nils
Sickert, Ludwig
van der Wal, Oskar
Nissim, Malvina
Bisazza, Arianna
Source :
Proceedings of ACL: System Demonstrations (2023) 421-435
Publication Year :
2023

Abstract

Past work in natural language processing interpretability focused mainly on popular classification tasks while largely overlooking generation settings, partly due to a lack of dedicated tools. In this work, we introduce Inseq, a Python library to democratize access to interpretability analyses of sequence generation models. Inseq enables intuitive and optimized extraction of models' internal information and feature importance scores for popular decoder-only and encoder-decoder Transformers architectures. We showcase its potential by adopting it to highlight gender biases in machine translation models and locate factual knowledge inside GPT-2. Thanks to its extensible interface supporting cutting-edge techniques such as contrastive feature attribution, Inseq can drive future advances in explainable natural language generation, centralizing good practices and enabling fair and reproducible model evaluations.<br />Comment: ACL 2023 Demo Track. Library: https://github.com/inseq-team/inseq, Docs: https://inseq.readthedocs.io, v0.4

Details

Database :
arXiv
Journal :
Proceedings of ACL: System Demonstrations (2023) 421-435
Publication Type :
Report
Accession number :
edsarx.2302.13942
Document Type :
Working Paper
Full Text :
https://doi.org/10.18653/v1/2023.acl-demo.40