Papers with Code - Papers with Code Newsletter #4
Welcome to the 4th issue of the Papers with Code newsletter. Self-attention continues to be adopted to build deep learning architectures that address computer vision problems like instance segmentation and object detection. One recent example is Vision Transformer (ViT) proposed by Dosovitskiy et al. Despite being promising for vision tasks, these large models can show computational inefficiencies and inferior performance (compared to established vision architectures). This leaves room for improvements.
Mar-28-2021, 02:25:18 GMT