This first release introduces 2 simple ways to plot attention of a Transformer-based model.
- TokenToTokenAttentions allow to plot attention matrices of a specific layer given a HuggingFace `model` and a `tokenizer`.
- TokenToTokenNormalizedAttentions allow to plot attention matrices normalized across head axis given a HuggingFace `model` and a `tokenizer`.
Plots are rendered with `matplotlib`.