Attention Basin: Why Contextual Position Matters in Large Language Models