LiteLLM vs vLLM

FeatureLiteLLMvLLM
CategoryAI DevelopmentLocal AI Infrastructure
PricingFree (open-source) + EnterpriseFree (open-source)
GitHub Stars15,00045,000
PlatformsLinux, macOS, DockerLinux
Features
  • βœ“ 100+ providers
  • βœ“ Load balancing
  • βœ“ Budget tracking
  • βœ“ Caching
  • βœ“ OpenAI-compatible
  • βœ“ PagedAttention
  • βœ“ Continuous batching
  • βœ“ Tensor parallelism
  • βœ“ OpenAI-compatible API
  • βœ“ Multi-GPU
  • βœ“ Quantization
Tags
api-gatewaymulti-providerproxyopen-source
open-sourceinferenceservinggpuhigh-throughput