ollama/README.md
2023-06-27 13:52:32 -04:00

86 lines
1.2 KiB
Markdown

# Ollama
- Run models easily
- Download, manage and import models
## Install
```
pip install ollama
```
## Example quickstart
```python
import ollama
ollama.generate("./llama-7b-ggml.bin", "hi")
```
## Reference
### `ollama.generate(model, message)`
Generate a completion
```python
ollama.generate("./llama-7b-ggml.bin", "hi")
```
### `ollama.load(model)`
Load a model for generation
```python
ollama.load("model")
```
### `ollama.models()`
List available local models
```
models = ollama.models()
```
### `ollama.serve()`
Serve the ollama http server
## Cooming Soon
### `ollama.pull(model)`
Download a model
```python
ollama.pull("huggingface.co/thebloke/llama-7b-ggml")
```
### `ollama.import(filename)`
Import a model from a file
```python
ollama.import("./path/to/model")
```
### `ollama.search("query")`
Search for compatible models that Ollama can run
```python
ollama.search("llama-7b")
```
## Future CLI
In the future, there will be an `ollama` CLI for running models on servers, in containers or for local development environments.
```
ollama generate huggingface.co/thebloke/llama-7b-ggml "hi"
> Downloading [================> ] 66.67% (2/3) 30.2MB/s
```
## Documentation
- [Development](docs/development.md)