Hacker News new | past | comments | ask | show | jobs | submit | from login
vLLM V1: A Major Upgrade to vLLM's Core Architecture (vllm.ai)
2 points by ozgune 44 days ago | past
vLLM V1: A Major Upgrade to vLLM's Core Architecture (vllm.ai)
5 points by xmo 48 days ago | past
VLLM 2024 Retrospective and 2025 Vision (vllm.ai)
1 point by shenli3514 56 days ago | past
Installing and Developing VLLM with Ease (vllm.ai)
1 point by brethil 62 days ago | past
vLLM v0.6.0: 2.7x Throughput Improvement and 5x Latency Reduction (vllm.ai)
3 points by xmo 6 months ago | past
VLLM automatic prefix / prompt caching (vllm.ai)
2 points by danielhanchen 6 months ago | past | 1 comment
VLLM hosts local LLMs easily (vllm.ai)
2 points by myprotegeai 7 months ago | past
Llama 3.1 Support in VLLM (vllm.ai)
2 points by e12e 7 months ago | past
vLLM (vllm.ai)
2 points by jonbaer 10 months ago | past
VLLM: Easy, Fast, and Cheap LLM Serving with PagedAttention (vllm.ai)
2 points by udev4096 on Jan 7, 2024 | past
Notes on VLLM v.s. DeepSpeed-FastGen (vllm.ai)
3 points by Palmik on Nov 15, 2023 | past
vLLM: Easy, Fast, and Cheap LLM Serving with PagedAttention (vllm.ai)
295 points by wskwon on June 20, 2023 | past | 42 comments

Join us for AI Startup School this June 16-17 in San Francisco!

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: