Amazingly easy to run LLMs with Ollama
There is a new tool called Ollama that makes it really easy to try out different LLMs on your local machine. Here are several youtube videos about this.
There is a new tool called Ollama that makes it really easy to try out different LLMs on your local machine. Here are several youtube videos about this.
curl https://ollama.ai/install.sh | sh
Then to run a model we can do things like:
ollama run llama2:70b
At https://ollama.ai/library you can see what models are available.
After you click on a model you can click "tags" to see all the
different versions.
Then you can click to copy the command to run that model.
I look for the largest one that fits in my 48 GB GPU.
At https://ollama.ai/library/llama2/tags there are different model names
for different quantization levels.
It is so easy to be interacting with a model on a local machine now.
It is wonderful!
Update Dec 22: I made a video showing some model runs on GPU and CPU modes.
Comments
Post a Comment