Update README to use cli options for server

This commit is contained in:
Andrei Betlen 2023-05-07 05:10:52 -04:00
parent 627811ea83
commit 3adc8fb3ae

View file

@ -68,18 +68,9 @@ This allows you to use llama.cpp compatible models with any OpenAI compatible cl
To install the server package and get started: To install the server package and get started:
Linux/MacOS
```bash ```bash
pip install llama-cpp-python[server] pip install llama-cpp-python[server]
export MODEL=./models/7B/ggml-model.bin python3 -m llama_cpp.server --model models/7B/ggml-model.bin
python3 -m llama_cpp.server
```
Windows
```cmd
pip install llama-cpp-python[server]
SET MODEL=..\models\7B\ggml-model.bin
python3 -m llama_cpp.server
``` ```
Navigate to [http://localhost:8000/docs](http://localhost:8000/docs) to see the OpenAPI documentation. Navigate to [http://localhost:8000/docs](http://localhost:8000/docs) to see the OpenAPI documentation.