No description
2fb52261ad
* basic distribution w/ push/pull * add the parser * add create, pull, and push * changes to the parser, FROM line, and fix commands * mkdirp new manifest directories * make `blobs` directory if it does not exist * fix go warnings * add progressbar for model pulls * move model struct --------- Co-authored-by: Jeffrey Morgan <jmorganca@gmail.com> |
||
---|---|---|
api | ||
app | ||
cmd | ||
docs | ||
examples/python | ||
llama | ||
parser | ||
scripts | ||
server | ||
web | ||
.dockerignore | ||
.gitignore | ||
.prettierrc.json | ||
Dockerfile | ||
ggml-metal.metal | ||
go.mod | ||
go.sum | ||
LICENSE | ||
main.go | ||
models.json | ||
README.md |
Ollama
Run large language models with llama.cpp
.
Note: certain models that can be run with Ollama are intended for research and/or non-commercial use only.
Features
- Download and run popular large language models
- Switch between multiple models on the fly
- Hardware acceleration where available (Metal, CUDA)
- Fast inference server written in Go, powered by llama.cpp
- REST API to use with your application (python, typescript SDKs coming soon)
Install
- Download for macOS
- Download for Windows (coming soon)
You can also build the binary from source.
Quickstart
Run a fast and simple model.
ollama run orca
Example models
💬 Chat
Have a conversation.
ollama run vicuna "Why is the sky blue?"
🗺️ Instructions
Get a helping hand.
ollama run orca "Write an email to my boss."
🔎 Ask questions about documents
Send the contents of a document and ask questions about it.
ollama run nous-hermes "$(cat input.txt)", please summarize this story
📖 Storytelling
Venture into the unknown.
ollama run nous-hermes "Once upon a time"
Advanced usage
Run a local model
ollama run ~/Downloads/vicuna-7b-v1.3.ggmlv3.q4_1.bin
Building
go build .
To run it start the server:
./ollama server &
Finally, run a model!
./ollama run ~/Downloads/vicuna-7b-v1.3.ggmlv3.q4_1.bin
API Reference
POST /api/pull
Download a model
curl -X POST http://localhost:11343/api/pull -d '{"model": "orca"}'
POST /api/generate
Complete a prompt
curl -X POST http://localhost:11434/api/generate -d '{"model": "orca", "prompt": "hello!"}'