Repository brief

vllm-project/vllm

Read the upstream summary on the left, browse the cached forks below it, and load each fork comparison into the right-hand panel.

Cached analysis
cached 2026-03-30T10:40:13.421Z

vllm-project/vllm

vLLM is a very active, widely adopted open-source LLM inference and serving engine focused on high throughput and memory efficiency. It has a large ecosystem footprint, with 74,699 stars and 14,954 forks, and recent commits show ongoing work across bug fixes, security, ROCm, quantization, and model support.

GitHub
Stars74,699
Forks14,954
Default branchmain
Last pushed2026-03-30T10:26:47Z
Best maintainedSystemPanic/vllm-windows
Closest to upstreamEmbeddedLLM/vllm
Most feature-richSaid-Akbar/vllm-rocm
Most opinionatedSaid-Akbar/vllm-rocm
Forks

Choose a fork to inspect

20 cached fork briefs