Mask Tokens as Prophet: Fine-Grained Cache Eviction for Efficient dLLM Inference