Causal Interpretation of Self-Attention in Pre-Trained Transformers

Neural Information Processing Systems 

We propose a causal interpretation of self-attention in the Transformer neural network architecture. We interpret self-attention as a mechanism that estimates a structural equation model for a given input sequence of symbols (tokens). The structural equation model can be interpreted, in turn, as a causal structure over the input symbols under the specific context of the input sequence.

Similar Docs  Excel Report  more

TitleSimilaritySource
None found