llama.cpp/llama_cpp/server
khimaros ea1f88dd29
fix: Use '\n' seperator for EventSourceResponse (#1188)
this fixes compatibility with some OpenAI clients, including BetterChatGPT (https://github.com/ztjhz/BetterChatGPT/issues/537).

Co-authored-by: Andrei <abetlen@gmail.com>
2024-02-15 15:20:13 -05:00
..
__init__.py llama_cpp server: app is now importable, still runnable as a module 2023-04-29 11:41:25 -07:00
__main__.py [Feat] Multi model support (#931) 2023-12-22 05:51:25 -05:00
app.py fix: Use '\n' seperator for EventSourceResponse (#1188) 2024-02-15 15:20:13 -05:00
cli.py Fix python3.8 support 2024-01-19 08:17:49 -05:00
errors.py server: Support none defaulting to infinity for completions (#111) 2023-12-22 14:05:13 -05:00
model.py fix: broken import 2024-02-08 01:13:28 -05:00
settings.py Add speculative decoding (#1120) 2024-01-31 14:08:14 -05:00
types.py server: Support none defaulting to infinity for completions (#111) 2023-12-22 14:05:13 -05:00