Mug
3bb45f1658
More reasonable defaults
2023-04-10 16:38:45 +02:00
Mug
0cccb41a8f
Added iterative search to prevent instructions from being echoed, add ignore eos, add no-mmap, fixed 1 character echo too much bug
2023-04-10 16:35:38 +02:00
Andrei Betlen
241d608bbb
Update workflow permissions
2023-04-10 02:35:00 -04:00
Andrei Betlen
3d56c3b706
Run tests for pr's to main
2023-04-10 02:19:22 -04:00
Andrei Betlen
bc02ce353b
Bump version
2023-04-10 02:12:19 -04:00
Andrei Betlen
1f67ad2a0b
Add use_mmap option
2023-04-10 02:11:35 -04:00
Andrei Betlen
d41cb0ecf7
Add create release step to workflow
2023-04-10 01:54:52 -04:00
Andrei Betlen
8594b8388e
Add build and release
2023-04-10 01:29:32 -04:00
Andrei Betlen
a984f55d79
Quickfix: forgot to clone submodules when building and publishing pypi package
2023-04-10 00:51:25 -04:00
Andrei Betlen
196650ccb2
Update model paths to be more clear they should point to file
2023-04-09 22:45:55 -04:00
Andrei Betlen
a79d3eb732
Fix workflow name
2023-04-09 22:38:19 -04:00
Andrei Betlen
fda975e5a9
Rename test publish
2023-04-09 22:34:17 -04:00
Andrei Betlen
baa394491c
Add PyPI publish workflow
2023-04-09 22:32:30 -04:00
Andrei Betlen
8c2bb3042f
Bump version
2023-04-09 22:12:23 -04:00
Andrei Betlen
c3c2623e8b
Update llama.cpp
2023-04-09 22:01:33 -04:00
Andrei Betlen
e636214b4e
Add test publish workflow
2023-04-08 19:57:37 -04:00
Andrei Betlen
314ce7d1cc
Fix cpu count default
2023-04-08 19:54:04 -04:00
Andrei Betlen
3fbc06361f
Formatting
2023-04-08 16:01:45 -04:00
Andrei Betlen
0067c1a588
Formatting
2023-04-08 16:01:18 -04:00
Andrei Betlen
0a5c551371
Bump version
2023-04-08 15:09:48 -04:00
Andrei Betlen
38f442deb0
Bugfix: Wrong size of embeddings. Closes #47
2023-04-08 15:05:33 -04:00
Andrei Betlen
6d1bda443e
Add clients example. Closes #46
2023-04-08 09:35:32 -04:00
Andrei Betlen
c940193e64
Bump version
2023-04-08 03:13:39 -04:00
Andrei Betlen
edaaa1bd63
Only build wheels on workflow dispatch
2023-04-08 03:11:25 -04:00
Andrei Betlen
ae3e9c3d6f
Update shared library extension for macos
2023-04-08 02:45:21 -04:00
Andrei Betlen
6a143ac0db
Merge branch 'main' of github.com:abetlen/llama_cpp_python into main
2023-04-08 02:40:42 -04:00
Andrei Betlen
e611cfc56d
Build shared library with make on unix platforms
2023-04-08 02:39:17 -04:00
Andrei Betlen
a3f713039f
Update llama.cpp
2023-04-08 02:38:42 -04:00
Andrei
41365b0456
Merge pull request #15 from SagsMug/main
...
llama.cpp chat example implementation
2023-04-07 20:43:33 -04:00
Mug
16fc5b5d23
More interoperability to the original llama.cpp, and arguments now work
2023-04-07 13:32:19 +02:00
Andrei Betlen
c3b1aa6ab7
Clone submodule
2023-04-07 03:19:07 -04:00
Andrei Betlen
d4912a80da
Install build dependencies
2023-04-07 03:18:56 -04:00
Andrei Betlen
d74800da52
Build wheels
2023-04-07 03:14:38 -04:00
Andrei Betlen
0fd32046cb
Bump version
2023-04-06 22:48:54 -04:00
Andrei Betlen
88c23d04a8
Fix windows dll location issue
2023-04-06 22:44:31 -04:00
Andrei Betlen
241722c981
Quote destination
2023-04-06 22:38:53 -04:00
Andrei Betlen
d75196d7a1
Install with pip during build step
...
Use setup.py install
Upgrade version of setuptools
Revert to develop
Use setup.py build and pip install
Just use pip install
Use correct name in pyproject.toml
Make pip install verbose
2023-04-06 22:21:45 -04:00
Andrei Betlen
dd1c298620
Fix typo
2023-04-06 21:28:03 -04:00
Andrei Betlen
baa825dacb
Add windows and mac runners
2023-04-06 21:27:01 -04:00
Andrei Betlen
da539cc2ee
Safer calculation of default n_threads
2023-04-06 21:22:19 -04:00
Andrei Betlen
9b7526895d
Bump version
2023-04-06 21:19:08 -04:00
Andrei Betlen
7851cc1e3c
Don't install pydantic by default
2023-04-06 21:10:34 -04:00
Andrei Betlen
09707f5b2a
Remove console script
2023-04-06 21:08:32 -04:00
Andrei Betlen
930db37dd2
Merge branch 'main' of github.com:abetlen/llama_cpp_python into main
2023-04-06 21:07:38 -04:00
Andrei Betlen
55279b679d
Handle prompt list
2023-04-06 21:07:35 -04:00
Andrei
c2e690b326
Merge pull request #29 from MillionthOdin16/main
...
Fixes and Tweaks to Defaults
2023-04-06 21:06:31 -04:00
Mug
10c7571117
Fixed too many newlines, now onto args.
...
Still needs shipping work so you could do "python -m llama_cpp.examples." etc.
2023-04-06 15:33:22 +02:00
Mug
085cc92b1f
Better llama.cpp interoperability
...
Has some too many newline issues so WIP
2023-04-06 15:30:57 +02:00
MillionthOdin16
2e91affea2
Ignore ./idea folder
2023-04-05 18:23:17 -04:00
MillionthOdin16
c283edd7f2
Set n_batch to default values and reduce thread count:
...
Change batch size to the llama.cpp default of 8. I've seen issues in llama.cpp where batch size affects quality of generations. (It shouldn't) But in case that's still an issue I changed to default.
Set auto-determined num of threads to 1/2 system count. ggml will sometimes lock cores at 100% while doing nothing. This is being addressed, but can cause bad experience for user if pegged at 100%
2023-04-05 18:17:29 -04:00