update README.md
This commit is contained in:
parent
11614b6d84
commit
20cdd9fee6
3 changed files with 11 additions and 17 deletions
|
@ -74,10 +74,10 @@ ollama.search("llama-7b")
|
||||||
|
|
||||||
## Future CLI
|
## Future CLI
|
||||||
|
|
||||||
In the future, there will be an easy CLI for running models
|
In the future, there will be an `ollama` CLI for running models on servers, in containers or for local development environments.
|
||||||
|
|
||||||
```
|
```
|
||||||
ollama run huggingface.co/thebloke/llama-7b-ggml
|
ollama generaate huggingface.co/thebloke/llama-7b-ggml
|
||||||
> Downloading [================> ] 66.67% (2/3) 30.2MB/s
|
> Downloading [================> ] 66.67% (2/3) 30.2MB/s
|
||||||
```
|
```
|
||||||
|
|
||||||
|
|
|
@ -1,16 +1,18 @@
|
||||||
# Desktop
|
# Desktop
|
||||||
|
|
||||||
The Ollama desktop experience
|
The Ollama desktop app
|
||||||
|
|
||||||
## Running
|
## Running
|
||||||
|
|
||||||
|
In the background run the `ollama.py` [development](../docs/development.md) server:
|
||||||
|
|
||||||
|
```
|
||||||
|
python ../ollama.py serve --port 5001
|
||||||
|
```
|
||||||
|
|
||||||
|
Then run the desktop app:
|
||||||
|
|
||||||
```
|
```
|
||||||
npm install
|
npm install
|
||||||
npm start
|
npm start
|
||||||
```
|
```
|
||||||
|
|
||||||
## Packaging
|
|
||||||
|
|
||||||
```
|
|
||||||
npm run package
|
|
||||||
```
|
|
||||||
|
|
|
@ -14,14 +14,6 @@ Put your model in `models/` and run:
|
||||||
python3 ollama.py serve
|
python3 ollama.py serve
|
||||||
```
|
```
|
||||||
|
|
||||||
To run the app:
|
|
||||||
|
|
||||||
```
|
|
||||||
cd desktop
|
|
||||||
npm install
|
|
||||||
npm start
|
|
||||||
```
|
|
||||||
|
|
||||||
## Building
|
## Building
|
||||||
|
|
||||||
If using Apple silicon, you need a Python version that supports arm64:
|
If using Apple silicon, you need a Python version that supports arm64:
|
||||||
|
|
Loading…
Reference in a new issue