Yann LeCun Team Uses Dictionary Learning To Peek Into Transformers' Black Boxes
Transformer architectures have become the building blocks for many state-of-the-art natural language processing (NLP) models. While transformers are certainly powerful, researchers' understanding of how they actually work remains limited. This is problematic due to the lack of transparency and the possibility of biases being inherited via training data and algorithms, which could cause models to produce unfair or incorrect predictions. In the paper Transformer Visualization via Dictionary Learning: Contextualized Embedding as a Linear Superposition of Transformer Factors, a Yann LeCun team from Facebook AI Research, UC Berkeley and New York University leverages dictionary learning techniques to provide detailed visualizations of transformer representations and insights into the semantic structures -- such as word-level disambiguation, sentence-level pattern formation, and long-range dependencies -- that are captured by transformers. Previous attempts to visualize and analyze this "black box" issue in transformers include direct visualization and, more recently, "probing tasks" designed to interpret transformer models.
Apr-6-2021, 07:50:29 GMT
- AI-Alerts:
- 2021 > 2021-04 > AAAI AI-Alert for Apr 6, 2021 (1.00)
- Country:
- North America > United States > New York (0.25)
- Industry:
- Transportation > Air (0.61)
- Technology: