Hacker News new | past | comments | ask | show | jobs | submit login
Ask HN: What's the current state of in-browser GPU accelerated LLM inference?
6 points by ramesh31 6 months ago | hide | past | favorite | 1 comment
Curious if anything is approaching usability at GPT3+ levels, even with a large binary download.



I don't think there has been much change since this. https://news.ycombinator.com/item?id=35583349

WebGPU is getting close to general support which will make things a bit faster but compute isn't as much of the issue as ram.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: