Install
curl -fsSL https://ollama.com/install.sh | sh
Run a model
Pull a model, then run it interactively.
ollama pull llama4:scout
ollama run llama4:scout
API
Send chat requests to your local server.
curl http://localhost:11434/api/chat -d '{
"model": "llama4:scout",
"messages": [{ "role": "user", "content": "Why is the sky blue?" }]
}'
Cloud
For models that don't fit on your laptop, run on Ollama Cloud — same CLI, same API, just --cloud.
ollama run --cloud kimi-k2:1t