inferlet
Pie: A Programmable Serving System for Emerging LLM Applications
Gim, In, Ma, Zhiyao, Lee, Seung-seob, Zhong, Lin
Emerging large language model (LLM) applications involve diverse reasoning strategies and agentic workflows, straining the capabilities of existing serving systems built on a monolithic token generation loop. This paper introduces Pie, a programmable LLM serving system designed for flexibility and efficiency. Pie decomposes the traditional generation loop into fine-grained service handlers exposed via an API and delegates control of the generation process to user-provided programs, called inferlets. This enables applications to implement new KV cache strategies, bespoke generation logic, and seamlessly integrate computation and I/O-entirely within the application, without requiring modifications to the serving system. Pie executes inferlets using WebAssembly, benefiting from its lightweight sandboxing. Our evaluation shows Pie matches state-of-the-art performance on standard tasks (3-12% latency overhead) while significantly improving latency and throughput (1.3x-3.4x higher) on agentic workflows by enabling application-specific optimizations.
- Asia > South Korea > Seoul > Seoul (0.05)
- North America > United States > Virginia (0.04)
- North America > United States > New York > New York County > New York City (0.04)
- Workflow (0.70)
- Research Report (0.64)