SPIRe: Boosting LLM Inference Throughput with Speculative Decoding