PagedEviction: Structured Block-wise KV Cache Pruning for Efficient Large Language Model Inference