RATTENTION: Towards the Minimal Sliding Window Size in Local-Global Attention Models

Open in new window