Commit graph

923 commits

Author SHA1 Message Date
dependabot[bot]
fa79484a29
Bump scikit-build from 0.13.0 to 0.17.5
Bumps [scikit-build](https://github.com/scikit-build/scikit-build) from 0.13.0 to 0.17.5.
- [Release notes](https://github.com/scikit-build/scikit-build/releases)
- [Changelog](https://github.com/scikit-build/scikit-build/blob/main/CHANGES.rst)
- [Commits](https://github.com/scikit-build/scikit-build/compare/0.13.0...0.17.5)

---
updated-dependencies:
- dependency-name: scikit-build
  dependency-type: direct:development
  update-type: version-update:semver-minor
...

Signed-off-by: dependabot[bot] <support@github.com>
2023-05-30 00:29:33 +00:00
Andrei
41e2184431
Merge pull request #292 from abetlen/dependabot/pip/mkdocstrings-0.22.0
Bump mkdocstrings from 0.21.2 to 0.22.0
2023-05-29 20:28:02 -04:00
dependabot[bot]
62ac7c3761
Bump mkdocstrings from 0.21.2 to 0.22.0
Bumps [mkdocstrings](https://github.com/mkdocstrings/mkdocstrings) from 0.21.2 to 0.22.0.
- [Release notes](https://github.com/mkdocstrings/mkdocstrings/releases)
- [Changelog](https://github.com/mkdocstrings/mkdocstrings/blob/master/CHANGELOG.md)
- [Commits](https://github.com/mkdocstrings/mkdocstrings/compare/0.21.2...0.22.0)

---
updated-dependencies:
- dependency-name: mkdocstrings
  dependency-type: direct:development
  update-type: version-update:semver-minor
...

Signed-off-by: dependabot[bot] <support@github.com>
2023-05-29 21:03:33 +00:00
Maximilian-Winter
719c3eae0a Diskcache implementation for llama state. 2023-05-28 15:56:38 +02:00
Andrei Betlen
80066f0b80 Use async routes 2023-05-27 09:12:58 -04:00
Andrei Betlen
c2b59a5f59 Import unnused import 2023-05-26 22:59:29 -04:00
Andrei Betlen
8f2b4456ad Format 2023-05-26 22:04:31 -04:00
Andrei Betlen
84e313bd6e Align dtype to match c structs 2023-05-26 22:02:16 -04:00
Andrei Betlen
b0b154cfa6 Add changelog message for numpy 2023-05-26 20:26:08 -04:00
Andrei Betlen
66bcb8d70d Merge branch 'main' into add-numpy-support 2023-05-26 20:25:03 -04:00
Andrei Betlen
8f35bddd7e Fix stop sequence performance bug. 2023-05-26 20:23:49 -04:00
Andrei Betlen
7fc7bc30e7 Remove usage of eval_tokens for cache check 2023-05-26 20:12:05 -04:00
Andrei Betlen
fe331ec589 Replace eval_logits and eval_tokens with numpy arrays 2023-05-26 20:03:31 -04:00
Andrei Betlen
00ea3af51b Add makefile 2023-05-26 17:56:20 -04:00
Andrei Betlen
447a3d249e Merge branch 'main' into setup 2023-05-26 17:53:58 -04:00
Andrei Betlen
efb763bcdc Merge branch 'main' into add-numpy-support 2023-05-26 17:37:33 -04:00
Andrei Betlen
030fafe901 Add project changelog 2023-05-26 17:32:34 -04:00
Andrei Betlen
878c7d363f Merge branch 'main' into add-numpy-support 2023-05-26 17:22:29 -04:00
Andrei Betlen
6075e17cb6 Bump version 2023-05-26 17:21:51 -04:00
Andrei Betlen
bd4b95da45 Reduce numpy version dependency 2023-05-26 16:38:21 -04:00
Andrei Betlen
8eb9769f78 Add support for numpy 2023-05-26 16:12:45 -04:00
Andrei
2adf6f3f9a
Merge pull request #265 from dmahurin/fix-from-bytes-byteorder
fix "from_bytes() missing required argument 'byteorder'"
2023-05-26 12:53:06 -04:00
Andrei
34ad71f448
Merge pull request #274 from dmahurin/fix-missing-antiprompt
low_level_api_chat_cpp.py: Fix missing antiprompt output in chat.
2023-05-26 12:52:34 -04:00
Andrei
d78453c045
Merge pull request #264 from dmahurin/fix-min-keep
fix "missing 1 required positional argument: 'min_keep'"
2023-05-26 12:52:05 -04:00
Andrei Betlen
4c1b7f7a76 Bugfix for logits_processor and stopping_criteria 2023-05-26 10:25:28 -04:00
Don Mahurin
0fa2ec4903 low_level_api_chat_cpp.py: Fix missing antiprompt output in chat. 2023-05-26 06:54:28 -07:00
Andrei Betlen
433a2e3e8a Add extra logits_processor and stopping_criteria 2023-05-26 03:13:24 -04:00
Andrei Betlen
30bf8ec557 Update llama.cpp 2023-05-26 03:03:11 -04:00
Andrei Betlen
f74b90ed67 Fix streaming hang on last token when cache is on. 2023-05-26 03:03:01 -04:00
Andrei Betlen
5be8354e11 Added tokenizer 2023-05-26 03:00:51 -04:00
Andrei Betlen
8fa2ef1959 Format 2023-05-26 03:00:35 -04:00
Andrei Betlen
6bd1075291 Merge branch 'Maximilian-Winter/main' into main 2023-05-26 02:56:11 -04:00
Andrei Betlen
ca01f98e09 Add LlamaTokenizer class 2023-05-25 14:11:33 -04:00
Andrei Betlen
1d247e0f35 Add StoppingCriteria and LogitsProcessor to generate to match huggingface API 2023-05-25 14:04:54 -04:00
Maximilian Winter
c6a9659972
Merge branch 'abetlen:main' into main 2023-05-25 17:09:19 +02:00
Andrei
de8d9a810b
Merge pull request #270 from gjmulder/auto-docker
"bot-in-a-box" - model d/l and automatic install into a OpenBLAS or CuBLAS Docker image
2023-05-25 09:30:13 -04:00
Gary Mulder
0e0c9bb978 Merge branch 'auto-docker' of github.com:gjmulder/llama-cpp-python-gary into auto-docker 2023-05-25 11:50:34 +00:00
Gary Mulder
0d2cc21202 Fixed repeated imports 2023-05-25 11:50:02 +00:00
Maximilian-Winter
c2585b6889 Fixed list elements typing 2023-05-25 10:54:08 +02:00
Maximilian-Winter
da463e6c8c Added types to logit processor list and stop criteria list 2023-05-25 09:07:16 +02:00
Maximilian-Winter
c05fcdf42f Fixed none value of logits processors. 2023-05-24 22:02:06 +02:00
Maximilian-Winter
5bb780d455 Implemented logit processors and stop criteria's 2023-05-24 21:55:44 +02:00
Andrei Betlen
fab064ded9 Remove unnecessary ffi calls 2023-05-23 17:56:21 -04:00
Gary Mulder
ec44bdad61
Update README.md 2023-05-23 20:50:39 +01:00
Gary Mulder
ed19071ef8 Renamed and moved old Dockerfiles 2023-05-23 19:38:37 +00:00
Gary Mulder
70f629a72f
Update README.md 2023-05-23 20:36:21 +01:00
Gary Mulder
eaff7a8678 Initial commit of auto docker 2023-05-23 19:26:40 +00:00
Don Mahurin
d6a7adb17a fix "missing 1 required positional argument: 'min_keep'" 2023-05-23 06:42:22 -07:00
Don Mahurin
327eedbfe1 fix "from_bytes() missing required argument 'byteorder'" 2023-05-23 00:20:34 -07:00
Andrei Betlen
e5d596e0e9 Bump version 2023-05-22 23:50:58 -04:00