docs: Update readme examples to use newer Qwen2 model (#1544)
This commit is contained in:
parent
5beec1a1fd
commit
27d53589ff
1 changed files with 2 additions and 2 deletions
|
@ -327,7 +327,7 @@ You'll need to install the `huggingface-hub` package to use this feature (`pip i
|
||||||
|
|
||||||
```python
|
```python
|
||||||
llm = Llama.from_pretrained(
|
llm = Llama.from_pretrained(
|
||||||
repo_id="Qwen/Qwen1.5-0.5B-Chat-GGUF",
|
repo_id="Qwen/Qwen2-0.5B-Instruct-GGUF",
|
||||||
filename="*q8_0.gguf",
|
filename="*q8_0.gguf",
|
||||||
verbose=False
|
verbose=False
|
||||||
)
|
)
|
||||||
|
@ -688,7 +688,7 @@ For possible options, see [llama_cpp/llama_chat_format.py](llama_cpp/llama_chat_
|
||||||
If you have `huggingface-hub` installed, you can also use the `--hf_model_repo_id` flag to load a model from the Hugging Face Hub.
|
If you have `huggingface-hub` installed, you can also use the `--hf_model_repo_id` flag to load a model from the Hugging Face Hub.
|
||||||
|
|
||||||
```bash
|
```bash
|
||||||
python3 -m llama_cpp.server --hf_model_repo_id Qwen/Qwen1.5-0.5B-Chat-GGUF --model '*q8_0.gguf'
|
python3 -m llama_cpp.server --hf_model_repo_id Qwen/Qwen2-0.5B-Instruct-GGUF --model '*q8_0.gguf'
|
||||||
```
|
```
|
||||||
|
|
||||||
### Web Server Features
|
### Web Server Features
|
||||||
|
|
Loading…
Reference in a new issue