it runs on the CPU
this news story is that they are now extending GPU support to llama.cpp
You can probably find a google colab link.