totem
Generalizability Under Sensor Failure: Tokenization + Transformers Enable More Robust Latent Spaces
Chau, Geeling, An, Yujin, Iqbal, Ahamed Raffey, Chung, Soon-Jo, Yue, Yisong, Talukder, Sabera
A major goal in neuroscience is to discover neural data representations that generalize. This goal is challenged by variability along recording sessions (e.g. environment), subjects (e.g. varying neural structures), and sensors (e.g. sensor noise), among others. Recent work has begun to address generalization across sessions and subjects, but few study robustness to sensor failure which is highly prevalent in neuroscience experiments. In order to address these generalizability dimensions we first collect our own electroencephalography dataset with numerous sessions, subjects, and sensors, then study two time series models: EEGNet (Lawhern et al., 2018) and TOTEM (Talukder et al., 2024). EEGNet is a widely used convolutional neural network, while TOTEM is a discrete time series tokenizer and transformer model. We find that TOTEM outperforms or matches EEGNet across all generalizability cases. Finally through analysis of TOTEM's latent codebook we observe that tokenization enables generalization.
- Europe > Austria > Styria > Graz (0.05)
- North America > United States > Utah (0.05)
- North America > United States > California > Los Angeles County > Pasadena (0.04)
TOTEM: TOkenized Time Series EMbeddings for General Time Series Analysis
Talukder, Sabera, Yue, Yisong, Gkioxari, Georgia
The field of general time series analysis has recently begun to explore unified modeling, where a common architectural backbone can be retrained on a specific task for a specific dataset. In this work, we approach unification from a complementary vantage point: unification across tasks and domains. To this end, we explore the impact of discrete, learnt, time series data representations that enable generalist, cross-domain training. Our method, TOTEM, or TOkenized Time Series EMbeddings, proposes a simple tokenizer architecture that embeds time series data from varying domains using a discrete vectorized representation learned in a self-supervised manner. TOTEM works across multiple tasks and domains with minimal to no tuning. We study the efficacy of TOTEM with an extensive evaluation on 17 real world time series datasets across 3 tasks. We evaluate both the specialist (i.e., training a model on each domain) and generalist (i.e., training a single model on many domains) settings, and show that TOTEM matches or outperforms previous best methods on several popular benchmarks. The code can be found at: https://github.com/SaberaTalukder/TOTEM.
- North America > United States > California > Santa Clara County > Palo Alto (0.04)
- North America > United States > California > Los Angeles County > Pasadena (0.04)
Why so many video games include tabletop games, from Gwent to Orlog
All of these tabletop games provide players an opportunity for turn-based strategy, which is naturally a slower, more methodical style of gameplay than open-world combat. In "Horizon Forbidden West," each totem for Machine Strike is modeled off the robotic animals you fight against in the wilds of Horizon's world. To defeat a glinthawk in "Forbidden West," you need to aim and strike the sack of chillwater in its breastbone, all while dodging to avoid the creature's incoming attacks. To win against a glinthawk totem in Machine Strike, you need to methodically outmanuever your opponent in a step-by-step process across tiles on the game board. You're allowed the time to step back and think.