$π$-Attention: Periodic Sparse Transformers for Efficient Long-Context Modeling