Semi-Clairvoyant Scheduling of Speculative Decoding Requests to Minimize LLM Inference Latency