Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

You mentioned "on local agents". I've noticed this too. How do ChatGPT and the others get around this, and provide instant responses on long conversations?




Not getting around it, just benefiting from parallel compute / huge flops of GPUs. Fundamentally, it's just that prefill compute is itself highly parallel and HBM is just that much faster than LPDDR. Effectively H100s and B100s can chew through the prefill in under a second at ~50k token lengths, so the TTFT (Time to First Token) can feel amazingly fast.

They cache the intermediate data (KV cache).



Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: