Hacker News new | past | comments | ask | show | jobs | submit login

Deepseek v3 needs 16 H100s or 8 H200 GPUs for inference.



Or a single 2 processor AMD EPYC rig, for less than $6k.

https://xcancel.com/carrigmat/status/1884244369907278106

The only reason you need all those GPUs is because they only have a fraction of the ram you can cram in a server.

With AMD focusing on ram channels and cores the above rig can do 6-8 tokens per second inference.

The GPUs will be faster, but the point is inference on the top deepseek model is possible for $6k with an AMD server rig. 8 H200's alone would cost $256,000 and gobble up way more power than the 400 watt envelope of that EPYC rig.


I thought the whole story about Deepseek was that Deepseek does not have H100s?




Join us for AI Startup School this June 16-17 in San Francisco!

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: