Scalable and Cost-Efficient ML Inference: Parallel Batch Processing with Serverless Functions