Goto

Collaborating Authors

 swintrack



SwinTrack: A Simple and Strong Baseline for Transformer Tracking

Neural Information Processing Systems

Recently Transformer has been largely explored in tracking and shown state-of-the-art (SOTA) performance. However, existing efforts mainly focus on fusing and enhancing features generated by convolutional neural networks (CNNs). The potential of Transformer in representation learning remains under-explored. In this paper, we aim to further unleash the power of Transformer by proposing a simple yet efficient fully-attentional tracker, dubbed SwinTrack, within classic Siamese framework.


SwinTrack: A Simple and Strong Baseline for Transformer Tracking Liting Lin 1,2 Heng Fan

Neural Information Processing Systems

Recently Transformer has been largely explored in tracking and shown state-of-the-art (SOT A) performance. However, existing efforts mainly focus on fusing and enhancing features generated by convolutional neural networks (CNNs). The potential of Transformer in representation learning remains under-explored.


SwinTrack: A Simple and Strong Baseline for Transformer Tracking

Neural Information Processing Systems

Recently Transformer has been largely explored in tracking and shown state-of-the-art (SOTA) performance. However, existing efforts mainly focus on fusing and enhancing features generated by convolutional neural networks (CNNs). The potential of Transformer in representation learning remains under-explored. In this paper, we aim to further unleash the power of Transformer by proposing a simple yet efficient fully-attentional tracker, dubbed SwinTrack, within classic Siamese framework. Besides, to further enhance robustness, we present a novel motion token that embeds historical target trajectory to improve tracking by providing temporal context.