Andrei Betlen
|
cd3f1bb387
|
feat: Update llama.cpp
|
2024-06-04 00:35:47 -04:00 |
|
Andrei Betlen
|
91d05aba46
|
fix: adjust kv_override member names to match llama.cpp
|
2024-05-29 02:28:58 -04:00 |
|
Andrei Betlen
|
454c9bb1cb
|
feat: Update llama.cpp
|
2024-05-27 10:51:57 -04:00 |
|
Andrei Betlen
|
087cc0b036
|
feat: Update llama.cpp
|
2024-05-24 01:43:36 -04:00 |
|
Andrei Betlen
|
5a595f035a
|
feat: Update llama.cpp
|
2024-05-22 02:40:31 -04:00 |
|
Andrei Betlen
|
50f5c74ecf
|
Update llama.cpp
|
2024-05-14 09:30:04 -04:00 |
|
Andrei Betlen
|
2a39b99575
|
feat: Update llama.cpp
|
2024-05-08 08:42:22 -04:00 |
|
Andrei Betlen
|
228949c1f7
|
feat: Update llama.cpp
|
2024-05-08 02:22:15 -04:00 |
|
Andrei Betlen
|
3e2597eac8
|
feat: Update llama.cpp
|
2024-05-05 12:12:27 -04:00 |
|
Andrei Betlen
|
8c2b24d5aa
|
feat: Update llama.cpp
|
2024-04-30 09:27:55 -04:00 |
|
Andrei Betlen
|
97fb860eba
|
feat: Update llama.cpp
|
2024-04-29 23:34:55 -04:00 |
|
Andrei Betlen
|
c9b85bf098
|
feat: Update llama.cpp
|
2024-04-27 23:41:54 -04:00 |
|
Andrei Betlen
|
173ebc7878
|
fix: Remove duplicate pooling_type definition and add misisng n_vocab definition in bindings
|
2024-04-25 21:36:09 -04:00 |
|
Douglas Hanley
|
f6ed21f9a2
|
feat: Allow for possibly non-pooled embeddings (#1380)
* allow for possibly non-pooled embeddings
* add more to embeddings section in README.md
---------
Co-authored-by: Andrei <abetlen@gmail.com>
|
2024-04-25 21:32:44 -04:00 |
|
Andrei Betlen
|
7f52335c50
|
feat: Update llama.cpp
|
2024-04-25 21:21:29 -04:00 |
|
Andrei Betlen
|
2a9979fce1
|
feat: Update llama.cpp
|
2024-04-25 02:48:26 -04:00 |
|
Andrei Betlen
|
159cc4e5d9
|
feat: Update llama.cpp
|
2024-04-21 20:46:40 -04:00 |
|
Andrei Betlen
|
1347e1d050
|
feat: Add typechecking for ctypes structure attributes
|
2024-04-10 02:40:41 -04:00 |
|
Andrei Betlen
|
889d0e8981
|
feat: Update llama.cpp
|
2024-04-10 02:25:58 -04:00 |
|
Andrei Betlen
|
56071c956a
|
feat: Update llama.cpp
|
2024-04-09 09:53:49 -04:00 |
|
Limour
|
f165048a69
|
feat: add support for KV cache quantization options (#1307)
* add KV cache quantization options
https://github.com/abetlen/llama-cpp-python/discussions/1220
https://github.com/abetlen/llama-cpp-python/issues/1305
* Add ggml_type
* Use ggml_type instead of string for quantization
* Add server support
---------
Co-authored-by: Andrei Betlen <abetlen@gmail.com>
|
2024-04-01 10:19:28 -04:00 |
|
Andrei Betlen
|
125b2358c9
|
feat: Update llama.cpp
|
2024-03-28 12:06:46 -04:00 |
|
Andrei Betlen
|
901fe02461
|
feat: Update llama.cpp
|
2024-03-26 22:58:53 -04:00 |
|
Andrei Betlen
|
e325a831f0
|
feat: Update llama.cpp
|
2024-03-22 23:43:29 -04:00 |
|
Andrei Betlen
|
8d298b4750
|
feat: Update llama.cpp
|
2024-03-18 10:26:36 -04:00 |
|
Andrei Betlen
|
6eb25231e4
|
feat: Update llama.cpp
|
2024-03-15 12:58:45 -04:00 |
|
Andrei Betlen
|
d318cc8b83
|
fix: Set default pooling_type to mean, check for null pointer.
|
2024-03-14 09:17:41 -04:00 |
|
Andrei Betlen
|
dd0ee56217
|
feat: Update llama.cpp
|
2024-03-13 15:57:35 -04:00 |
|
Andrei Betlen
|
08e910f7a7
|
feat: Update llama.cpp
|
2024-03-10 23:45:05 -04:00 |
|
Andrei Betlen
|
40c6b54f68
|
feat: Update llama.cpp
|
2024-03-08 20:58:50 -05:00 |
|
Andrei Betlen
|
93dc56ace8
|
Update llama.cpp
|
2024-03-06 01:32:00 -05:00 |
|
Andrei Betlen
|
87a6e5797e
|
feat: Update llama.cpp
|
2024-03-03 11:27:04 -05:00 |
|
Andrei Betlen
|
0e70984fb6
|
feat: Update llama.cpp
|
2024-03-02 22:20:04 -05:00 |
|
Andrei Betlen
|
f062a7f51d
|
feat: Update llama.cpp
|
2024-03-01 12:57:16 -05:00 |
|
Andrei Betlen
|
8c71725d53
|
fix: Remove deprecated cfg sampling functions
|
2024-02-28 14:37:07 -05:00 |
|
Andrei Betlen
|
0d37ce52b1
|
feat: Update llama.cpp
|
2024-02-28 14:27:16 -05:00 |
|
Andrei Betlen
|
fea33c9b94
|
feat: Update llama.cpp
|
2024-02-27 12:22:17 -05:00 |
|
Andrei Betlen
|
9558ce7878
|
feat: Update llama.cpp
|
2024-02-26 11:40:58 -05:00 |
|
Andrei Betlen
|
cbbcd888af
|
feat: Update llama.cpp
|
2024-02-25 20:52:14 -05:00 |
|
Andrei Betlen
|
19234aa0db
|
fix: Restore type hints for low-level api
|
2024-02-25 16:54:37 -05:00 |
|
Andrei Betlen
|
2292af5796
|
feat: Update llama.cpp
|
2024-02-25 16:53:58 -05:00 |
|
Andrei Betlen
|
221edb9ef1
|
feat: Update llama.cpp
|
2024-02-24 23:47:29 -05:00 |
|
Andrei Betlen
|
a0ce429dc0
|
misc: use decorator to bind low level api functions, fixes docs
|
2024-02-23 03:39:38 -05:00 |
|
Andrei Betlen
|
e10af30cf1
|
fix: TypeAlias import error
|
2024-02-22 03:27:28 -05:00 |
|
Andrei Betlen
|
aefcb8f71a
|
misc: additional type annotations for low level api
|
2024-02-22 02:00:09 -05:00 |
|
Andrei Betlen
|
0653e15c20
|
feat: Update llama.cpp
|
2024-02-21 23:04:52 -05:00 |
|
Andrei
|
7f51b6071f
|
feat(low-level-api): Improve API static type-safety and performance (#1205)
|
2024-02-21 16:25:38 -05:00 |
|
Andrei Betlen
|
4edde21b3d
|
feat: Update llama.cpp
|
2024-02-21 11:05:58 -05:00 |
|
Andrei Betlen
|
6225f027e5
|
feat: Update llama.cpp
|
2024-02-19 04:11:34 -05:00 |
|
Andrei Betlen
|
748c0ce057
|
feat: Update llama.cpp
|
2024-02-18 21:30:36 -05:00 |
|