Bump version
This commit is contained in:
parent
3fbcded7cd
commit
6b3aa7fc8f
2 changed files with 18 additions and 1 deletions
17
CHANGELOG.md
17
CHANGELOG.md
|
@ -7,6 +7,23 @@ and this project adheres to [Semantic Versioning](https://semver.org/spec/v2.0.0
|
||||||
|
|
||||||
## [Unreleased]
|
## [Unreleased]
|
||||||
|
|
||||||
|
## [0.2.12]
|
||||||
|
|
||||||
|
- Update llama.cpp to ggerganov/llama.cpp@50337961a678fce4081554b24e56e86b67660163
|
||||||
|
- Fix missing `n_seq_id` in `llama_batch` by @NickAlgra in #842
|
||||||
|
- Fix exception raised in `__del__` when freeing models by @cebtenzzre in #848
|
||||||
|
- Performance improvement for logit bias by @zolastro in #851
|
||||||
|
- Fix suffix check arbitrary code execution bug by @mtasic85 in #854
|
||||||
|
- Fix typo in `function_call` parameter in `llama_types.py` by @akatora28 in #849
|
||||||
|
- Fix streaming not returning `finish_reason` by @gmcgoldr in #798
|
||||||
|
- Fix `n_gpu_layers` check to allow values less than 1 for server by @hxy9243 in #826
|
||||||
|
- Supppress stdout and stderr when freeing model by @paschembri in #803
|
||||||
|
- Fix `llama2` chat format by @delock in #808
|
||||||
|
- Add validation for tensor_split size by @eric1932 #820
|
||||||
|
- Print stack trace on server error by @abetlen in d6a130a052db3a50975a719088a9226abfebb266
|
||||||
|
- Update docs for gguf by @johnccshen in #783
|
||||||
|
- Add `chatml` chat format by @abetlen in 305482bd4156c70802fc054044119054806f4126
|
||||||
|
|
||||||
## [0.2.11]
|
## [0.2.11]
|
||||||
|
|
||||||
- Fix bug in `llama_model_params` object has no attribute `logits_all` by @abetlen in d696251fbe40015e8616ea7a7d7ad5257fd1b896
|
- Fix bug in `llama_model_params` object has no attribute `logits_all` by @abetlen in d696251fbe40015e8616ea7a7d7ad5257fd1b896
|
||||||
|
|
|
@ -1,4 +1,4 @@
|
||||||
from .llama_cpp import *
|
from .llama_cpp import *
|
||||||
from .llama import *
|
from .llama import *
|
||||||
|
|
||||||
__version__ = "0.2.11"
|
__version__ = "0.2.12"
|
Loading…
Reference in a new issue