No description
Find a file
2023-06-27 13:46:46 -04:00
desktop desktop: load model from file 2023-06-27 12:35:55 -04:00
docs add development doc 2023-06-27 13:46:46 -04:00
models adding orca models 2023-06-27 11:41:54 -04:00
.gitignore add templates to prompt command 2023-06-26 13:41:16 -04:00
build.py move back to root 2023-06-27 12:12:44 -04:00
LICENSE proto -> ollama 2023-06-26 15:57:13 -04:00
model_prompts.json fix build 2023-06-27 12:32:21 -04:00
ollama.py fix build 2023-06-27 12:32:21 -04:00
README.md add development doc 2023-06-27 13:46:46 -04:00
requirements.txt move back to root 2023-06-27 12:12:44 -04:00
template.py fix build 2023-06-27 12:32:21 -04:00

Ollama

  • Run models easily
  • Download, manage and import models

Install

pip install ollama

Example quickstart

import ollama
ollama.generate("./llama-7b-ggml.bin", "hi")

Reference

ollama.generate(model, message)

Generate a completion

ollama.generate("./llama-7b-ggml.bin", "hi")

ollama.load(model)

Load a model for generation

ollama.load("model")

ollama.models()

List available local models

models = ollama.models()

ollama.serve()

Serve the ollama http server

Cooming Soon

ollama.pull(model)

Download a model

ollama.pull("huggingface.co/thebloke/llama-7b-ggml")

ollama.import(filename)

Import a model from a file

ollama.import("./path/to/model")

ollama.search("query")

Search for compatible models that Ollama can run

ollama.search("llama-7b")

Future CLI

In the future, there will be an easy CLI for running models

ollama run huggingface.co/thebloke/llama-7b-ggml
> Downloading [================>          ] 66.67% (2/3) 30.2MB/s

Documentation