vllm/vllm_v0.10.0/docs/deployment/integrations/kserve.md

288 B

KServe

vLLM can be deployed with KServe on Kubernetes for highly scalable distributed model serving.

Please see this guide for more details on using vLLM with KServe.