No description
api | ||
app | ||
cmd | ||
docs | ||
llama | ||
server | ||
templates | ||
web | ||
.dockerignore | ||
.gitignore | ||
.prettierrc.json | ||
Dockerfile | ||
go.mod | ||
go.sum | ||
LICENSE | ||
main.go | ||
Makefile | ||
models.json | ||
README.md |
Ollama
Run large language models with llama.cpp
.
Note: certain models that can be run with this project are intended for research and/or non-commercial use only.
Features
- Download and run popular large language models
- Switch between multiple models on the fly
- Hardware acceleration where available (Metal, CUDA)
- Fast inference server written in Go, powered by llama.cpp
- REST API to use with your application (python, typescript SDKs coming soon)
Install
- Download for macOS
- Download for Windows (coming soon)
- Docker:
docker run -p 8080:8080 ollama/ollama
You can also build the binary from source.
Quickstart
Run the model that started it all.
ollama run llama
Example models
💬 Chat
Have a conversation.
ollama run vicuna "Why is the sky blue?"
🗺️ Instructions
Ask questions. Get answers.
ollama run orca "Write an email to my boss."
👩💻 Code completion
Sometimes you just need a little help writing code.
ollama run replit "Give me react code to render a button"
📖 Storytelling
Venture into the unknown.
ollama run storyteller "Once upon a time"
Building
make
To run it start the server:
./ollama server &
Finally, run a model!
./ollama run ~/Downloads/vicuna-7b-v1.3.ggmlv3.q4_1.bin
API Reference
POST /completion
Complete a prompt
curl --unix-socket ~/.ollama/ollama.sock http://localhost/api/generate \
-X POST \
-d '{"model": "/path/to/model", "prompt": "Once upon a time", "stream": true}'