No description
Find a file
2023-06-29 18:25:07 -04:00
desktop update README.md with instructions for poetry 2023-06-28 11:47:37 -04:00
docs update development.md 2023-06-28 12:41:30 -07:00
ollama consistency between generate and add naming 2023-06-29 18:22:54 -04:00
.gitignore add templates to prompt command 2023-06-26 13:41:16 -04:00
build.py desktop: fixes for initial publish 2023-06-27 14:34:56 -04:00
LICENSE proto -> ollama 2023-06-26 15:57:13 -04:00
models.json use correct name for orca-mini 2023-06-29 18:16:33 -04:00
poetry.lock spinner on generate 2023-06-29 16:42:43 -04:00
pyproject.toml update version 2023-06-29 16:58:18 -04:00
README.md simplify README.md 2023-06-29 18:25:07 -04:00
requirements.txt remove server extras for now 2023-06-28 17:59:23 -07:00

Ollama

Ollama is a tool for running any large language model on any machine. It's designed to be easy to use and fast, supporting the largest number of models possible by using the fastest loader available for your platform and model.

Note: this project is a work in progress.

Install

pip install ollama

Quickstart

To run a model, use ollama run:

ollama run orca-mini-3b

You can also run models from hugging face:

ollama run huggingface.co/TheBloke/orca_mini_3B-GGML

Or directly via downloaded model files:

ollama run ~/Downloads/orca-mini-13b.ggmlv3.q4_0.bin

Python SDK

Example

import ollama
ollama.generate("orca-mini-3b", "hi")

ollama.generate(model, message)

Generate a completion

ollama.generate("./llama-7b-ggml.bin", "hi")

ollama.models()

List available local models

models = ollama.models()

ollama.serve()

Serve the ollama http server

ollama.serve()

ollama.add(filepath)

Add a model by importing from a file

ollama.add("./path/to/model")

ollama.load(model)

Manually a model for generation

ollama.load("model")

ollama.unload(model)

Unload a model

ollama.unload("model")

ollama.pull(model)

Download a model

ollama.pull("huggingface.co/thebloke/llama-7b-ggml")

Coming Soon

ollama.search("query")

Search for compatible models that Ollama can run

ollama.search("llama-7b")

Documentation