Kernel Deformed Exponential Families for Sparse Continuous Attention
Moreno, Alexander, Nagesh, Supriya, Wu, Zhenke, Dempsey, Walter, Rehg, James M.
Attention mechanisms take an expectation of a data representation with respect to probability weights. This creates summary statistics that focus on important features. Recently, Martins et al. (2020; 2021) proposed continuous attention mechanisms, focusing on unimodal attention densities from the exponential and deformed exponential families: the latter has sparse support. Farinhas et al. (2021) extended this to use Gaussian mixture attention densities, which are a flexible class with dense support. In this paper, we extend this to two general flexible classes: kernel exponential families (Canu & Smola, 2006) and our new sparse counterpart kernel deformed exponential families. Theoretically, we show new existence results for both kernel exponential and deformed exponential families, and that the deformed case has similar approximation capabilities to kernel exponential families. Experiments show that kernel deformed exponential families can attend to multiple compact regions of the data domain. Attention mechanisms take weighted averages of data representations (Bahdanau et al., 2015), where the weights are a function of input objects. These are then used as inputs for prediction. Discrete attention 1) cannot easily handle data where observations are irregularly spaced 2) attention maps may be scattered, lacking focus.
Nov-12-2021
- Country:
- North America > United States
- Georgia > Fulton County
- Atlanta (0.04)
- Michigan > Washtenaw County
- Ann Arbor (0.04)
- Georgia > Fulton County
- North America > United States
- Genre:
- Research Report (0.50)
- Industry:
- Health & Medicine (0.68)
- Technology: