Kubernetes and Stateful Inference: How llm-d Solves the Routing and Caching Challenge for LLM Worklo…
As LLM production workloads grow, it becomes clear: classic Kubernetes mechanisms do not understand the nature of inference. llm-d is an attempt to bridge this gap at the platform level. The main limitation becomes apparent when inference goes beyond a “stateless HTTP service.” Requests to LLMs have different costs: prompt length, generation phase, KV-cache hits. … Read more