Mug
|
e3ea354547
|
Allow local llama library usage
|
2023-04-05 14:23:01 +02:00 |
|
Mug
|
e4c6f34d95
|
Merge branch 'main' of https://github.com/abetlen/llama-cpp-python
|
2023-04-05 14:18:27 +02:00 |
|
Andrei Betlen
|
6de2f24aca
|
Bump version
|
2023-04-05 06:53:43 -04:00 |
|
Andrei Betlen
|
e96a5c5722
|
Make Llama instance pickleable. Closes #27
|
2023-04-05 06:52:17 -04:00 |
|
Andrei Betlen
|
152e4695c3
|
Bump Version
|
2023-04-05 04:43:51 -04:00 |
|
Andrei Betlen
|
c177c807e5
|
Add supported python versions
|
2023-04-05 04:43:19 -04:00 |
|
Andrei Betlen
|
17fdd1547c
|
Update workflow name and add badge to README
|
2023-04-05 04:41:24 -04:00 |
|
Andrei Betlen
|
7643f6677d
|
Bugfix for Python3.7
|
2023-04-05 04:37:33 -04:00 |
|
Andrei Betlen
|
4d015c33bd
|
Fix syntax error
|
2023-04-05 04:35:15 -04:00 |
|
Andrei Betlen
|
47570df17b
|
Checkout submodules
|
2023-04-05 04:34:19 -04:00 |
|
Andrei Betlen
|
e3f999e732
|
Add missing scikit-build install
|
2023-04-05 04:31:38 -04:00 |
|
Andrei Betlen
|
43c20d3282
|
Add initial github action to run automated tests
|
2023-04-05 04:30:32 -04:00 |
|
Andrei Betlen
|
b1babcf56c
|
Add quantize example
|
2023-04-05 04:17:26 -04:00 |
|
Andrei Betlen
|
c8e13a78d0
|
Re-organize examples folder
|
2023-04-05 04:10:13 -04:00 |
|
Andrei Betlen
|
c16bda5fb9
|
Add performance tuning notebook
|
2023-04-05 04:09:19 -04:00 |
|
Andrei Betlen
|
cefc69ea43
|
Add runtime check to ensure embedding is enabled if trying to generate embeddings
|
2023-04-05 03:25:37 -04:00 |
|
Andrei Betlen
|
5c50af7462
|
Remove workaround
|
2023-04-05 03:25:09 -04:00 |
|
Andrei Betlen
|
c3972b61ae
|
Add basic tests. Closes #24
|
2023-04-05 03:23:15 -04:00 |
|
Andrei Betlen
|
51dbcf2693
|
Bugfix: wrong signature for quantize function
|
2023-04-04 22:36:59 -04:00 |
|
Andrei Betlen
|
8279fb7d92
|
Bump version
|
2023-04-04 17:17:11 -04:00 |
|
Andrei Betlen
|
c137789143
|
Add verbose flag. Closes #19
|
2023-04-04 13:09:24 -04:00 |
|
Andrei Betlen
|
5075c16fcc
|
Bugfix: n_batch should always be <= n_ctx
|
2023-04-04 13:08:21 -04:00 |
|
Mug
|
c862e8bac5
|
Fix repeating instructions and an antiprompt bug
|
2023-04-04 17:54:47 +02:00 |
|
Andrei Betlen
|
248b0566fa
|
Update README
|
2023-04-04 10:57:22 -04:00 |
|
Mug
|
9cde7973cc
|
Fix stripping instruction prompt
|
2023-04-04 16:20:27 +02:00 |
|
Mug
|
da5a6a7089
|
Added instruction mode, fixed infinite generation, and various other fixes
|
2023-04-04 16:18:26 +02:00 |
|
Mug
|
0b32bb3d43
|
Add instruction mode
|
2023-04-04 11:48:48 +02:00 |
|
Andrei Betlen
|
ffe34cf64d
|
Allow user to set llama config from env vars
|
2023-04-04 00:52:44 -04:00 |
|
Andrei Betlen
|
05eb2087d8
|
Small fixes for examples
|
2023-04-03 20:33:07 -04:00 |
|
Andrei Betlen
|
caf3c0362b
|
Add return type for default __call__ method
|
2023-04-03 20:26:08 -04:00 |
|
Andrei Betlen
|
4aa349d777
|
Add docstring for create_chat_completion
|
2023-04-03 20:24:20 -04:00 |
|
Andrei Betlen
|
4615f1e520
|
Add chat completion method to docs
|
2023-04-03 20:14:03 -04:00 |
|
Andrei Betlen
|
5cf29d0231
|
Bump version
|
2023-04-03 20:13:46 -04:00 |
|
Andrei Betlen
|
7fedf16531
|
Add support for chat completion
|
2023-04-03 20:12:44 -04:00 |
|
Andrei Betlen
|
3dec778c90
|
Update to more sensible return signature
|
2023-04-03 20:12:14 -04:00 |
|
Andrei Betlen
|
f7ab8d55b2
|
Update context size defaults Close #11
|
2023-04-03 20:11:13 -04:00 |
|
Andrei Betlen
|
c0a5c0171f
|
Add embed back into documentation
|
2023-04-03 18:53:00 -04:00 |
|
Andrei Betlen
|
adf656d542
|
Bump version
|
2023-04-03 18:46:49 -04:00 |
|
Andrei Betlen
|
ae004eb69e
|
Fix #16
|
2023-04-03 18:46:19 -04:00 |
|
Mug
|
f1615f05e6
|
Chat llama.cpp example implementation
|
2023-04-03 22:54:46 +02:00 |
|
Andrei Betlen
|
7d1977e8f0
|
Bump version
|
2023-04-03 14:49:36 -04:00 |
|
Andrei Betlen
|
4530197629
|
Update llama.cpp
|
2023-04-03 14:49:07 -04:00 |
|
Andrei
|
1d9a988644
|
Merge pull request #10 from MillionthOdin16/patch-1
Improve Shared Library Loading Mechanism
|
2023-04-03 14:47:11 -04:00 |
|
MillionthOdin16
|
a0758f0077
|
Update llama_cpp.py with PR requests
lib_base_name and load_shared_library
to
_lib_base_name and _load_shared_library
|
2023-04-03 13:06:50 -04:00 |
|
MillionthOdin16
|
a40476e299
|
Update llama_cpp.py
Make shared library code more robust with some platform specific functionality and more descriptive errors when failures occur
|
2023-04-02 21:50:13 -04:00 |
|
Andrei Betlen
|
b9a4513363
|
Update README
|
2023-04-02 21:03:39 -04:00 |
|
Andrei Betlen
|
7284adcaa8
|
Bump version
|
2023-04-02 13:36:07 -04:00 |
|
Andrei Betlen
|
1ed8cd023d
|
Update llama_cpp and add kv_cache api support
|
2023-04-02 13:33:49 -04:00 |
|
Andrei Betlen
|
74061b209d
|
Bump version
|
2023-04-02 03:59:47 -04:00 |
|
Andrei Betlen
|
4f509b963e
|
Bugfix: Stop sequences and missing max_tokens check
|
2023-04-02 03:59:19 -04:00 |
|