Mug
|
10c7571117
|
Fixed too many newlines, now onto args.
Still needs shipping work so you could do "python -m llama_cpp.examples." etc.
|
2023-04-06 15:33:22 +02:00 |
|
Mug
|
085cc92b1f
|
Better llama.cpp interoperability
Has some too many newline issues so WIP
|
2023-04-06 15:30:57 +02:00 |
|
Mug
|
283e59c5e9
|
Fix bug in init_break not being set when exited via antiprompt and others.
|
2023-04-05 14:47:24 +02:00 |
|
Mug
|
99ceecfccd
|
Move to new examples directory
|
2023-04-05 14:28:02 +02:00 |
|
Mug
|
e4c6f34d95
|
Merge branch 'main' of https://github.com/abetlen/llama-cpp-python
|
2023-04-05 14:18:27 +02:00 |
|
Andrei Betlen
|
b1babcf56c
|
Add quantize example
|
2023-04-05 04:17:26 -04:00 |
|
Andrei Betlen
|
c8e13a78d0
|
Re-organize examples folder
|
2023-04-05 04:10:13 -04:00 |
|
Andrei Betlen
|
c16bda5fb9
|
Add performance tuning notebook
|
2023-04-05 04:09:19 -04:00 |
|
Mug
|
c862e8bac5
|
Fix repeating instructions and an antiprompt bug
|
2023-04-04 17:54:47 +02:00 |
|
Mug
|
9cde7973cc
|
Fix stripping instruction prompt
|
2023-04-04 16:20:27 +02:00 |
|
Mug
|
da5a6a7089
|
Added instruction mode, fixed infinite generation, and various other fixes
|
2023-04-04 16:18:26 +02:00 |
|
Mug
|
0b32bb3d43
|
Add instruction mode
|
2023-04-04 11:48:48 +02:00 |
|
Andrei Betlen
|
ffe34cf64d
|
Allow user to set llama config from env vars
|
2023-04-04 00:52:44 -04:00 |
|
Andrei Betlen
|
05eb2087d8
|
Small fixes for examples
|
2023-04-03 20:33:07 -04:00 |
|
Andrei Betlen
|
7fedf16531
|
Add support for chat completion
|
2023-04-03 20:12:44 -04:00 |
|
Andrei Betlen
|
f7ab8d55b2
|
Update context size defaults Close #11
|
2023-04-03 20:11:13 -04:00 |
|
Mug
|
f1615f05e6
|
Chat llama.cpp example implementation
|
2023-04-03 22:54:46 +02:00 |
|
Andrei Betlen
|
caff127836
|
Remove commented out code
|
2023-04-01 15:13:01 -04:00 |
|
Andrei Betlen
|
f28bf3f13d
|
Bugfix: enable embeddings for fastapi server
|
2023-04-01 15:12:25 -04:00 |
|
Andrei Betlen
|
ed6f2a049e
|
Add streaming and embedding endpoints to fastapi example
|
2023-04-01 13:05:20 -04:00 |
|
Andrei Betlen
|
9fac0334b2
|
Update embedding example to new api
|
2023-04-01 13:02:51 -04:00 |
|
Andrei Betlen
|
5e011145c5
|
Update low level api example
|
2023-04-01 13:02:10 -04:00 |
|
Andrei Betlen
|
5f2e822b59
|
Rename inference example
|
2023-04-01 13:01:45 -04:00 |
|
Andrei Betlen
|
70b8a1ef75
|
Add support to get embeddings from high-level api. Closes #4
|
2023-03-28 04:59:54 -04:00 |
|
Andrei Betlen
|
3dbb3fd3f6
|
Add support for stream parameter. Closes #1
|
2023-03-28 04:03:57 -04:00 |
|
Andrei Betlen
|
dfe8608096
|
Update examples
|
2023-03-24 19:10:31 -04:00 |
|
Andrei Betlen
|
a61fd3b509
|
Add example based on stripped down version of main.cpp from llama.cpp
|
2023-03-24 18:57:25 -04:00 |
|
Andrei Betlen
|
2cc499512c
|
Black formatting
|
2023-03-24 14:35:41 -04:00 |
|
Andrei Betlen
|
d29b05bb67
|
Update example to match alpaca training prompt
|
2023-03-24 14:34:15 -04:00 |
|
Andrei Betlen
|
15e3dc7897
|
Add fastapi example
|
2023-03-24 01:41:24 -04:00 |
|
Andrei Betlen
|
9af16b63fd
|
Added low-level api inference example
|
2023-03-23 23:45:59 -04:00 |
|
Andrei Betlen
|
8680332203
|
Update examples
|
2023-03-23 23:12:42 -04:00 |
|
Andrei Betlen
|
90c78723de
|
Add basic langchain demo
|
2023-03-23 16:25:24 -04:00 |
|
Andrei Betlen
|
3d6eb32c76
|
Update basic example
|
2023-03-23 14:57:31 -04:00 |
|
Andrei Betlen
|
79b304c9d4
|
Initial commit
|
2023-03-23 05:33:06 -04:00 |
|