docs: Update readme examples to use newer Qwen2 model (#1544)
This commit is contained in:
parent
5beec1a1fd
commit
27d53589ff
1 changed files with 2 additions and 2 deletions
|
@ -327,7 +327,7 @@ You'll need to install the `huggingface-hub` package to use this feature (`pip i
|
|||
|
||||
```python
|
||||
llm = Llama.from_pretrained(
|
||||
repo_id="Qwen/Qwen1.5-0.5B-Chat-GGUF",
|
||||
repo_id="Qwen/Qwen2-0.5B-Instruct-GGUF",
|
||||
filename="*q8_0.gguf",
|
||||
verbose=False
|
||||
)
|
||||
|
@ -688,7 +688,7 @@ For possible options, see [llama_cpp/llama_chat_format.py](llama_cpp/llama_chat_
|
|||
If you have `huggingface-hub` installed, you can also use the `--hf_model_repo_id` flag to load a model from the Hugging Face Hub.
|
||||
|
||||
```bash
|
||||
python3 -m llama_cpp.server --hf_model_repo_id Qwen/Qwen1.5-0.5B-Chat-GGUF --model '*q8_0.gguf'
|
||||
python3 -m llama_cpp.server --hf_model_repo_id Qwen/Qwen2-0.5B-Instruct-GGUF --model '*q8_0.gguf'
|
||||
```
|
||||
|
||||
### Web Server Features
|
||||
|
|
Loading…
Reference in a new issue