Bump version
This commit is contained in:
parent
b82b0e1014
commit
d605875772
2 changed files with 9 additions and 1 deletions
|
@ -7,6 +7,14 @@ and this project adheres to [Semantic Versioning](https://semver.org/spec/v2.0.0
|
||||||
|
|
||||||
## [Unreleased]
|
## [Unreleased]
|
||||||
|
|
||||||
|
## [0.2.40]
|
||||||
|
|
||||||
|
- feat: Update llama.cpp to ggerganov/llama.cpp@3bdc4cd0f595a6096cca4a64aa75ffa8a3503465
|
||||||
|
- feat: Generic chatml Function Calling using chat_format="chatml-function-calling"` by @abetlen in #957
|
||||||
|
- fix: Circular dependancy preventing early Llama object free by @notwa in #1176
|
||||||
|
- docs: Set the correct command for compiling with syscl support by @akarshanbiswas in #1172
|
||||||
|
- feat: use gpu backend for clip if available by @iamlemec in #1175
|
||||||
|
|
||||||
## [0.2.39]
|
## [0.2.39]
|
||||||
|
|
||||||
- feat: Update llama.cpp to ggerganov/llama.cpp@b08f22c882a1443e6b97081f3ce718a4d1a741f8
|
- feat: Update llama.cpp to ggerganov/llama.cpp@b08f22c882a1443e6b97081f3ce718a4d1a741f8
|
||||||
|
|
|
@ -1,4 +1,4 @@
|
||||||
from .llama_cpp import *
|
from .llama_cpp import *
|
||||||
from .llama import *
|
from .llama import *
|
||||||
|
|
||||||
__version__ = "0.2.39"
|
__version__ = "0.2.40"
|
Loading…
Reference in a new issue