Hacker News
new
|
past
|
comments
|
ask
|
show
|
jobs
|
submit
|
from
login
vLLM V1: A Major Upgrade to vLLM's Core Architecture
(
vllm.ai
)
2 points
by
ozgune
44 days ago
|
past
vLLM V1: A Major Upgrade to vLLM's Core Architecture
(
vllm.ai
)
5 points
by
xmo
48 days ago
|
past
VLLM 2024 Retrospective and 2025 Vision
(
vllm.ai
)
1 point
by
shenli3514
56 days ago
|
past
Installing and Developing VLLM with Ease
(
vllm.ai
)
1 point
by
brethil
62 days ago
|
past
vLLM v0.6.0: 2.7x Throughput Improvement and 5x Latency Reduction
(
vllm.ai
)
3 points
by
xmo
6 months ago
|
past
VLLM automatic prefix / prompt caching
(
vllm.ai
)
2 points
by
danielhanchen
6 months ago
|
past
|
1 comment
VLLM hosts local LLMs easily
(
vllm.ai
)
2 points
by
myprotegeai
7 months ago
|
past
Llama 3.1 Support in VLLM
(
vllm.ai
)
2 points
by
e12e
7 months ago
|
past
vLLM
(
vllm.ai
)
2 points
by
jonbaer
10 months ago
|
past
VLLM: Easy, Fast, and Cheap LLM Serving with PagedAttention
(
vllm.ai
)
2 points
by
udev4096
on Jan 7, 2024
|
past
Notes on VLLM v.s. DeepSpeed-FastGen
(
vllm.ai
)
3 points
by
Palmik
on Nov 15, 2023
|
past
vLLM: Easy, Fast, and Cheap LLM Serving with PagedAttention
(
vllm.ai
)
295 points
by
wskwon
on June 20, 2023
|
past
|
42 comments
Join us for
AI Startup School
this June 16-17 in San Francisco!
Guidelines
|
FAQ
|
Lists
|
API
|
Security
|
Legal
|
Apply to YC
|
Contact
Search: