small README.md tweaks

This commit is contained in:
Jeffrey Morgan 2023-06-27 12:51:36 -04:00
parent 9497afb893
commit 7e6fd7b457

View file

@ -18,7 +18,15 @@ ollama.generate("./llama-7b-ggml.bin", "hi")
## Reference ## Reference
### `ollama.load` ### `ollama.generate(model, message)`
Generate a completion
```python
ollama.generate("./llama-7b-ggml.bin", "hi")
```
### `ollama.load(model)`
Load a model for generation Load a model for generation
@ -26,15 +34,7 @@ Load a model for generation
ollama.load("model name") ollama.load("model name")
``` ```
### `ollama.generate("message")` ### `ollama.models()`
Generate a completion
```python
ollama.generate(model, "hi")
```
### `ollama.models`
List available local models List available local models
@ -42,13 +42,13 @@ List available local models
models = ollama.models() models = ollama.models()
``` ```
### `ollama.serve` ### `ollama.serve()`
Serve the ollama http server Serve the ollama http server
## Cooing Soon ## Cooming Soon
### `ollama.pull` ### `ollama.pull("model")`
Download a model Download a model
@ -56,7 +56,7 @@ Download a model
ollama.pull("huggingface.co/thebloke/llama-7b-ggml") ollama.pull("huggingface.co/thebloke/llama-7b-ggml")
``` ```
### `ollama.import` ### `ollama.import("file")`
Import a model from a file Import a model from a file
@ -64,7 +64,7 @@ Import a model from a file
ollama.import("./path/to/model") ollama.import("./path/to/model")
``` ```
### `ollama.search` ### `ollama.search("query")`
Search for compatible models that Ollama can run Search for compatible models that Ollama can run
@ -74,7 +74,7 @@ ollama.search("llama-7b")
## Future CLI ## Future CLI
In the future, there will be an easy CLI for testing out models In the future, there will be an easy CLI for running models
``` ```
ollama run huggingface.co/thebloke/llama-7b-ggml ollama run huggingface.co/thebloke/llama-7b-ggml