Context Parallelism for Scalable Million-Token Inference