Amazingly easy to run LLMs with Ollama
There is a new tool called Ollama that makes it really easy to try out different LLMs on your local machine. Here are several youtube videos about this.
 
There is a new tool called Ollama that makes it really easy to try out different LLMs on your local machine. Here are several youtube videos about this.
 
curl https://ollama.ai/install.sh | sh Then to run a model we can do things like: ollama run llama2:70b At https://ollama.ai/library you can see what models are available.After you click on a model you can click "tags" to see all the different versions.Then you can click to copy the command to run that model. I look for the largest one that fits in my 48 GB GPU.  At https://ollama.ai/library/llama2/tags there are different model names for different quantization levels.  It is so easy to be interacting with a model on a local machine now.  It is wonderful! Update Dec 22:   I made a video showing some model runs on GPU and CPU modes. 
  
Comments
Post a Comment