Commit graph

648 commits

Author SHA1 Message Date
Andrei
c3ed1330d7
Merge pull request #177 from joelkurian/main
Updated installation instructions for BLAS backends
2023-05-10 05:27:12 -04:00
Andrei
3c96b43cf4
Merge pull request #178 from Stonelinks/document-presence-frequency-penalty
Document presence frequency penalty
2023-05-09 23:55:52 -04:00
Lucas Doyle
02e8a018ae llama_cpp server: document presence_penalty and frequency_penalty, mark as supported 2023-05-09 16:25:00 -07:00
Lucas Doyle
bebe7712f7 README: better setup instructions for developers for pip and poetry
Give folks options + explicit instructions for installing with poetry or pip.
2023-05-09 16:04:15 -07:00
Lucas Doyle
7e03fdd766 poetry: add poetry.toml, configure to install in a virtualenv
Its cleaner to keep the virtualenv in the project directory, and
poetry makes it easy to do that via this config.
2023-05-09 16:03:19 -07:00
Lucas Doyle
99c016c9b3 pyproject.toml: add skbuild to dev dependencies
The README says to run `python3 setup.py develop` to build the library, however on a blank install this fails because scikit-build is not installed. This adds it to the dev dependencies so that it is installed.
2023-05-09 16:03:13 -07:00
Joel Kurian
17dc51a7d2 Updated installation instructions for BLAS backends 2023-05-09 21:34:46 +05:30
Andrei Betlen
d957422bf4 Implement sampling as in llama.cpp main example 2023-05-08 21:21:25 -04:00
Andrei Betlen
93a9019bb1 Merge branch 'main' of github.com:abetlen/llama_cpp_python into Maximilian-Winter/main 2023-05-08 19:57:09 -04:00
Andrei Betlen
f315b82832 Revert changes to llama.cpp and setup.py 2023-05-08 19:53:21 -04:00
Andrei
7499fc1cbb
Merge pull request #126 from Stonelinks/deprecate-example-server
Deprecate example server
2023-05-08 19:29:04 -04:00
Andrei
1971514fa5
Merge pull request #173 from abetlen/dependabot/pip/mkdocs-material-9.1.11
Bump mkdocs-material from 9.1.9 to 9.1.11
2023-05-08 19:28:01 -04:00
Andrei Betlen
7af1f4c672 Merge branch 'main' of github.com:abetlen/llama_cpp_python into main 2023-05-08 18:49:38 -04:00
Andrei Betlen
c37883b477 Bump version 2023-05-08 18:49:37 -04:00
Andrei Betlen
82d138fe54 Fix: default repeat_penalty 2023-05-08 18:49:11 -04:00
dependabot[bot]
b1489befda
Bump mkdocs-material from 9.1.9 to 9.1.11
Bumps [mkdocs-material](https://github.com/squidfunk/mkdocs-material) from 9.1.9 to 9.1.11.
- [Release notes](https://github.com/squidfunk/mkdocs-material/releases)
- [Changelog](https://github.com/squidfunk/mkdocs-material/blob/master/CHANGELOG)
- [Commits](https://github.com/squidfunk/mkdocs-material/compare/9.1.9...9.1.11)

---
updated-dependencies:
- dependency-name: mkdocs-material
  dependency-type: direct:development
  update-type: version-update:semver-patch
...

Signed-off-by: dependabot[bot] <support@github.com>
2023-05-08 21:04:42 +00:00
Andrei
ed0f48b4bb
Merge pull request #153 from SagsMug/main
Update low_level_api examples
2023-05-08 14:58:47 -04:00
Andrei Betlen
a3cc7bf5b2 Bump version 2023-05-08 14:46:50 -04:00
Andrei Betlen
29f094bbcf Bugfix: not falling back to environment variables when default is value is set. 2023-05-08 14:46:25 -04:00
Andrei Betlen
6d69461ef5 Bump version 2023-05-08 14:21:47 -04:00
Andrei Betlen
0d6c60097a Show default value when --help is called 2023-05-08 14:21:15 -04:00
Andrei Betlen
022e9ebcb8 Use environment variable if parsed cli arg is None 2023-05-08 14:20:53 -04:00
Mug
eaf9f19aa9 Fix lora 2023-05-08 15:27:42 +02:00
Mug
2c0d9b182c Fix session loading and saving in low level example chat 2023-05-08 15:27:03 +02:00
Mug
ed66a469c9 Merge branch 'main' of https://github.com/abetlen/llama-cpp-python 2023-05-08 14:49:48 +02:00
Andrei Betlen
0d751a69a7 Set repeat_penalty to 0 by default 2023-05-08 01:50:43 -04:00
Andrei Betlen
65d9cc050c Add openai frequency and presence penalty parameters. Closes #169 2023-05-08 01:30:18 -04:00
Andrei Betlen
75d8619b1a Bump version 2023-05-07 20:19:34 -04:00
Andrei Betlen
a0b61ea2a7 Bugfix for models endpoint 2023-05-07 20:17:52 -04:00
Andrei Betlen
e72f58614b Change pointer to lower overhead byref 2023-05-07 20:01:34 -04:00
Andrei Betlen
14da46f16e Added cache size to settins object. 2023-05-07 19:33:17 -04:00
Andrei Betlen
0e94a70de1 Add in-memory longest prefix cache. Closes #158 2023-05-07 19:31:26 -04:00
Andrei Betlen
8dfde63255 Fix return type 2023-05-07 19:30:14 -04:00
Andrei Betlen
2753b85321 Format 2023-05-07 13:19:56 -04:00
Andrei Betlen
4f8cf52a38 Update README 2023-05-07 05:20:04 -04:00
Andrei Betlen
3adc8fb3ae Update README to use cli options for server 2023-05-07 05:10:52 -04:00
Andrei Betlen
627811ea83 Add verbose flag to server 2023-05-07 05:09:10 -04:00
Andrei Betlen
5f43c553d5 Bump version 2023-05-07 03:29:33 -04:00
Andrei Betlen
3fbda71790 Fix mlock_supported and mmap_supported return type 2023-05-07 03:04:22 -04:00
Andrei Betlen
5a3413eee3 Update cpu_count 2023-05-07 03:03:57 -04:00
Andrei Betlen
1a00e452ea Update settings fields and defaults 2023-05-07 02:52:20 -04:00
Andrei Betlen
86753976c4 Revert "llama_cpp server: delete some ignored / unused parameters"
This reverts commit b47b9549d5.
2023-05-07 02:02:34 -04:00
Andrei Betlen
c382d8f86a Revert "llama_cpp server: mark model as required"
This reverts commit e40fcb0575.
2023-05-07 02:00:22 -04:00
Andrei Betlen
d8fddcce73 Merge branch 'main' of github.com:abetlen/llama_cpp_python into better-server-params-and-fields 2023-05-07 01:54:00 -04:00
Andrei Betlen
397ae97f64 Update README 2023-05-07 01:41:19 -04:00
Andrei Betlen
1c188459c4 Merge branch 'main' of github.com:abetlen/llama_cpp_python into main 2023-05-07 00:41:10 -04:00
Andrei Betlen
7c3743fe5f Update llama.cpp 2023-05-07 00:12:47 -04:00
Andrei
c76e0913bb
Update issue templates 2023-05-06 22:18:31 -04:00
Andrei Betlen
bc853e3742 Fix type for eval_logits in LlamaState object 2023-05-06 21:32:50 -04:00
Andrei
ec617c04c1
Merge pull request #164 from abetlen/dependabot/pip/mkdocstrings-0.21.2
Bump mkdocstrings from 0.20.0 to 0.21.2
2023-05-06 21:30:09 -04:00