Not All Images are Worth 16x16 Words: Dynamic Transformers for Efficient Image Recognition 1

Neural Information Processing Systems 

Vision Transformers (ViT) have achieved remarkable success in large-scale image recognition. They split every 2D image into a fixed number of patches, each of which is treated as a token. Generally, representing an image with more tokens would lead to higher prediction accuracy, while it also results in drastically increased computational cost. To achieve a decent trade-off between accuracy and speed, the number of tokens is empirically set to 16x16 or 14x14. In this paper, we argue that every image has its own characteristics, and ideally the token number should be conditioned on each individual input.

Similar Docs  Excel Report  more

TitleSimilaritySource
None found