Abstract
Transformer-based language models significantly advanced the state-of-the-art in many linguistic tasks. As this revolution continues, the ability to explain model predictions has become a major area of interest for the NLP community. In this work, we present Gradient Self-Attention Maps (Grad-SAM) - a novel gradient-based method that analyzes self-attention units and identifies the input elements that explain the model's prediction the best. Extensive evaluations on various benchmarks show that Grad-SAM obtains significant improvements over state-of-the-art alternatives.
Original language | English |
---|---|
Title of host publication | CIKM 2021 - Proceedings of the 30th ACM International Conference on Information and Knowledge Management |
Publisher | Association for Computing Machinery |
Pages | 2882-2887 |
Number of pages | 6 |
ISBN (Electronic) | 9781450384469 |
DOIs | |
State | Published - 26 Oct 2021 |
Event | 30th ACM International Conference on Information and Knowledge Management, CIKM 2021 - Virtual, Online, Australia Duration: 1 Nov 2021 → 5 Nov 2021 |
Publication series
Name | International Conference on Information and Knowledge Management, Proceedings |
---|
Conference
Conference | 30th ACM International Conference on Information and Knowledge Management, CIKM 2021 |
---|---|
Country/Territory | Australia |
City | Virtual, Online |
Period | 1/11/21 → 5/11/21 |
Bibliographical note
Publisher Copyright:© 2021 ACM.
Keywords
- bert
- deep learning
- explainable & interpretable ai
- nlp
- self-attention
- transformers
- transparent machine learning