SegFormer: Simple and Efficient Design for Semantic Segmentation with Transformers

Neural Information Processing Systems 

The proposed MLP decoder aggregates information from different layers, and thus combining both local attention and global attention to render powerful representations. We show that this simple and lightweight design is the key to efficient segmentation on Transformers.