Hacker Newsnew | past | comments | ask | show | jobs | submitlogin
Self-hosting an AI (LLM) chatbot without going broke (kronis.dev)
4 points by KronisLV on June 6, 2023 | hide | past | favorite | 2 comments


With how many people utilize LLMs, I wanted to see about running my own. Unfortunately my RX 570 GPU has neither the memory, nor the support for running most of the models out there, nor do I have the money to rent a VPS with a GPU in the cloud.

Thankfully, I managed to find a way to get things running on your regular Windows/Linux consumer hardware, or in my case, also inside of Docker containers, thanks to a few different communities and projects out there. This does let me run the GGML models from Hugging Face locally or on x86 servers.

While none of this will be replacing GitHub Copilot or ChatGPT anytime soon, it's still nice that there are self-hostable options even for something like this out there, even without you needing specialized hardware.


Very informative - thanks for the write up!




Consider applying for YC's Fall 2025 batch! Applications are open till Aug 4

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: