Andrei Betlen
|
f2a54ecb4c
|
Update CHANGELOG
|
2023-06-09 11:01:42 -04:00 |
|
Andrei Betlen
|
556c7edf47
|
Truncate max_tokens if it exceeds context length
|
2023-06-09 10:57:36 -04:00 |
|
Andrei
|
fb2c5f7fd9
|
Merge pull request #350 from abetlen/migrate-to-scikit-build-core
Migrate to scikit-build-core
|
2023-06-09 03:00:01 -04:00 |
|
Andrei Betlen
|
b025a859ae
|
Add full path to shared library installation path
|
2023-06-08 22:11:01 -04:00 |
|
Andrei Betlen
|
146ca2c59f
|
Add missing httpx
|
2023-06-08 22:03:24 -04:00 |
|
Andrei Betlen
|
1d6bdf8db6
|
Update server dependencies
|
2023-06-08 21:59:58 -04:00 |
|
Andrei Betlen
|
43854e6a83
|
Update server dependencies
|
2023-06-08 21:55:42 -04:00 |
|
Andrei Betlen
|
c12138f7bd
|
Update changelog
|
2023-06-08 21:53:38 -04:00 |
|
Andrei Betlen
|
7345456779
|
Migrate to scikit-build-core
|
2023-06-08 21:49:42 -04:00 |
|
Andrei Betlen
|
202ed4464b
|
Update gitignore
|
2023-06-08 20:33:30 -04:00 |
|
Andrei Betlen
|
312f6eb144
|
Update llama.cpp
|
2023-06-08 16:46:55 -04:00 |
|
Andrei Betlen
|
0c42168508
|
Fix cache implementation breaking changes
|
2023-06-08 13:19:23 -04:00 |
|
Andrei Betlen
|
90874c01cd
|
Bump version
|
2023-06-08 03:26:49 -04:00 |
|
Andrei Betlen
|
e169a960f6
|
Ensure make clean removes all shared library files
|
2023-06-08 00:40:04 -04:00 |
|
Andrei Betlen
|
607d217caa
|
Allow both .so and .dylib extensions for macos
|
2023-06-08 00:27:19 -04:00 |
|
Andrei Betlen
|
c9e79c6681
|
Add metal build flags
|
2023-06-08 00:22:39 -04:00 |
|
Andrei Betlen
|
69355403c6
|
Update llama.cpp
|
2023-06-08 00:15:17 -04:00 |
|
Andrei Betlen
|
bf322861e8
|
Update llama.cpp
|
2023-06-07 03:43:33 -04:00 |
|
Andrei Betlen
|
fe4ae72ad2
|
Merge branch 'main' of github.com:abetlen/llama_cpp_python into main
|
2023-06-07 03:43:18 -04:00 |
|
Andrei Betlen
|
cf6a9d6d8e
|
Add framework and archive destinations to cmake for macos and windows.
|
2023-06-07 03:42:55 -04:00 |
|
Andrei
|
0f0b447fa4
|
Merge pull request #289 from Maximilian-Winter/main
Diskcache implementation for llama state.
|
2023-06-06 17:03:03 -04:00 |
|
Andrei Betlen
|
0e156ffd66
|
Fix changelog format
|
2023-06-06 17:01:10 -04:00 |
|
Andrei
|
d508573fb4
|
Merge pull request #328 from spirilis/mirostat
Added mirostat support for completions, chat completions API
|
2023-06-06 16:58:23 -04:00 |
|
Andrei Betlen
|
aad4b17f52
|
Update llama.cpp
|
2023-06-06 16:23:55 -04:00 |
|
Andrei Betlen
|
8b4968ea67
|
Fix resize issue. Closes #330
|
2023-06-06 11:37:57 -04:00 |
|
Eric B
|
9b1c9e902c
|
Added mirostat support for completions, chat completions API
|
2023-06-05 22:37:11 -04:00 |
|
Andrei
|
cc8858a413
|
Merge pull request #324 from abetlen/dependabot/pip/typing-extensions-4.6.3
Bump typing-extensions from 4.6.2 to 4.6.3
|
2023-06-05 19:47:30 -04:00 |
|
dependabot[bot]
|
5702d30a83
|
Bump typing-extensions from 4.6.2 to 4.6.3
Bumps [typing-extensions](https://github.com/python/typing_extensions) from 4.6.2 to 4.6.3.
- [Changelog](https://github.com/python/typing_extensions/blob/main/CHANGELOG.md)
- [Commits](https://github.com/python/typing_extensions/compare/4.6.2...4.6.3)
---
updated-dependencies:
- dependency-name: typing-extensions
dependency-type: direct:production
update-type: version-update:semver-patch
...
Signed-off-by: dependabot[bot] <support@github.com>
|
2023-06-05 23:46:45 +00:00 |
|
Andrei
|
2f42d6b513
|
Merge pull request #322 from abetlen/dependabot/pip/scikit-build-0.17.6
Bump scikit-build from 0.17.5 to 0.17.6
|
2023-06-05 19:45:16 -04:00 |
|
dependabot[bot]
|
fa7285c51f
|
Bump scikit-build from 0.17.5 to 0.17.6
Bumps [scikit-build](https://github.com/scikit-build/scikit-build) from 0.17.5 to 0.17.6.
- [Release notes](https://github.com/scikit-build/scikit-build/releases)
- [Changelog](https://github.com/scikit-build/scikit-build/blob/main/CHANGES.rst)
- [Commits](https://github.com/scikit-build/scikit-build/compare/0.17.5...0.17.6)
---
updated-dependencies:
- dependency-name: scikit-build
dependency-type: direct:development
update-type: version-update:semver-patch
...
Signed-off-by: dependabot[bot] <support@github.com>
|
2023-06-05 23:30:17 +00:00 |
|
Andrei
|
40276cb417
|
Merge pull request #325 from abetlen/dependabot/pip/fastapi-0.96.0
Bump fastapi from 0.95.1 to 0.96.0
|
2023-06-05 19:28:46 -04:00 |
|
dependabot[bot]
|
9e40061664
|
Bump fastapi from 0.95.1 to 0.96.0
Bumps [fastapi](https://github.com/tiangolo/fastapi) from 0.95.1 to 0.96.0.
- [Release notes](https://github.com/tiangolo/fastapi/releases)
- [Commits](https://github.com/tiangolo/fastapi/compare/0.95.1...0.96.0)
---
updated-dependencies:
- dependency-name: fastapi
dependency-type: direct:production
update-type: version-update:semver-minor
...
Signed-off-by: dependabot[bot] <support@github.com>
|
2023-06-05 23:04:20 +00:00 |
|
Andrei
|
5b968615bd
|
Merge pull request #326 from abetlen/dependabot/pip/sse-starlette-1.6.1
Bump sse-starlette from 1.5.0 to 1.6.1
|
2023-06-05 18:57:11 -04:00 |
|
Andrei Betlen
|
7b57420ea9
|
Update llama.cpp
|
2023-06-05 18:17:29 -04:00 |
|
dependabot[bot]
|
4bcaa5293c
|
Bump sse-starlette from 1.5.0 to 1.6.1
Bumps [sse-starlette](https://github.com/sysid/sse-starlette) from 1.5.0 to 1.6.1.
- [Release notes](https://github.com/sysid/sse-starlette/releases)
- [Changelog](https://github.com/sysid/sse-starlette/blob/master/CHANGELOG.md)
- [Commits](https://github.com/sysid/sse-starlette/compare/v1.5.0...v1.6.1)
---
updated-dependencies:
- dependency-name: sse-starlette
dependency-type: direct:production
update-type: version-update:semver-minor
...
Signed-off-by: dependabot[bot] <support@github.com>
|
2023-06-05 21:04:04 +00:00 |
|
Andrei Betlen
|
18c7b8520e
|
Bump version
|
2023-06-04 23:31:51 -04:00 |
|
Andrei Betlen
|
8ff83db3ee
|
Merge branch 'main' of github.com:abetlen/llama_cpp_python into main
|
2023-06-04 23:30:47 -04:00 |
|
Andrei Betlen
|
6d5b049801
|
Update llama.cpp
|
2023-06-04 23:30:42 -04:00 |
|
Andrei Betlen
|
76e364cdf2
|
Added 0.1.57 notes
|
2023-06-04 23:30:10 -04:00 |
|
Maximilian Winter
|
c3dfb5c24a
|
Merge branch 'abetlen:main' into main
|
2023-06-03 03:44:35 +02:00 |
|
Andrei
|
3977eea8e9
|
Merge pull request #310 from gjmulder/auto-docker
Auto docker v2 - dockerised Open Llama 3B image w/OpenBLAS enabled server
|
2023-06-02 13:02:48 -04:00 |
|
Gary Mulder
|
30d32e996b
|
More README.md corrections and cleanup
|
2023-06-02 11:08:59 +00:00 |
|
Gary Mulder
|
d4eef735d9
|
Fixed .gitignore to ignore any downloaded model .bin files. Cleaned up README.md again
|
2023-06-02 11:03:19 +00:00 |
|
Gary Mulder
|
f24e7a7e52
|
Updated instructions
|
2023-06-02 10:44:52 +00:00 |
|
Gary Mulder
|
cf4931a400
|
Working Open Llama 3B in a box
|
2023-06-02 08:48:54 +00:00 |
|
Andrei Betlen
|
71f4582d44
|
Bump version
|
2023-05-31 23:25:39 -04:00 |
|
Andrei Betlen
|
a89e3ca35f
|
Merge branch 'main' of github.com:abetlen/llama_cpp_python into main
|
2023-05-31 23:25:06 -04:00 |
|
Andrei Betlen
|
5377f9784a
|
Update llama.cpp
|
2023-05-31 23:24:52 -04:00 |
|
Maximilian-Winter
|
1848afebe0
|
Added dependencies.
|
2023-05-31 22:41:35 +02:00 |
|
Maximilian-Winter
|
29f9c9cca3
|
Added both LlamaChache classes Disk and RAM.
|
2023-05-31 22:33:56 +02:00 |
|