Hacker News new | past | comments | ask | show | jobs | submit login

IME ollama ran mixtral on a 1070 fast enough.



Though it most probably does not run in on the 1070 but rather on the cpu. It cannot fit on a 1070, it is not about speed, a 1070 cannot run it period.


In llama.cpp You can offload some of the layers to gpu with -ngl X. Where x is the number of layers




Join us for AI Startup School this June 16-17 in San Francisco!

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: