This website requires JavaScript.
Explore
Help
Sign in
baalajimaestro
/
llama.cpp
Watch
1
Star
0
Fork
You've already forked llama.cpp
0
Code
Issues
Pull requests
Projects
Releases
Packages
Wiki
Activity
3b92419132
llama.cpp
/
llama_cpp
/
server
History
Kyle Mistele
9c36688b33
fix(cli): allow passing n_ctx=0 to openAI API server args to use model n_ctx_train field per
#1015
(
#1093
)
2024-01-16 18:54:06 -05:00
..
__init__.py
llama_cpp server: app is now importable, still runnable as a module
2023-04-29 11:41:25 -07:00
__main__.py
[Feat] Multi model support (
#931
)
2023-12-22 05:51:25 -05:00
app.py
Support Accept text/event-stream in chat and completion endpoints,
resolves
#1083
(
#1088
)
2024-01-16 12:52:52 -05:00
cli.py
[Feat] Multi model support (
#931
)
2023-12-22 05:51:25 -05:00
errors.py
server: Support none defaulting to infinity for completions (
#111
)
2023-12-22 14:05:13 -05:00
model.py
Implement GGUF metadata KV overrides (
#1011
)
2024-01-15 12:29:29 -05:00
settings.py
fix(cli): allow passing n_ctx=0 to openAI API server args to use model n_ctx_train field per
#1015
(
#1093
)
2024-01-16 18:54:06 -05:00
types.py
server: Support none defaulting to infinity for completions (
#111
)
2023-12-22 14:05:13 -05:00