Hi HN,
I spent my easter weekend stuck in the house with COVID and I decided to play with llama.cpp [1] and fauxpilot [2] to see if I could get LLM code assist working on pure CPU.
As a proof of concept I'd say I've proven that it's possible. However there's still a lot to do. The auto complete is quite slow at the moment. PRs welcome.
[1] https://github.com/ggerganov/llama.cpp
[2] https://github.com/fauxpilot/fauxpilot