This website requires JavaScript.
Explore
Help
Sign in
baalajimaestro
/
llama.cpp
Watch
1
Star
0
Fork
You've already forked llama.cpp
0
Code
Issues
Pull requests
Projects
Releases
Packages
Wiki
Activity
e7962d2c73
llama.cpp
/
llama_cpp
/
server
History
Andrei Betlen
e7962d2c73
Fix: default max_tokens matches openai api (16 for completion, max length for chat completion)
2023-11-10 02:49:27 -05:00
..
__init__.py
llama_cpp server: app is now importable, still runnable as a module
2023-04-29 11:41:25 -07:00
__main__.py
Remove confusing helpstring from server cli args.
Closes
#719
2023-09-15 14:09:43 -04:00
app.py
Fix: default max_tokens matches openai api (16 for completion, max length for chat completion)
2023-11-10 02:49:27 -05:00