Paged Attention Meets FlexAttention: Unlocking Long-Context Efficiency in Deployed Inference