4b01a873ef
* Support defaulting to infinity or -1 for chat completions * Check if completion_tokens is none in error handler. * fix: max_tokens in create completion should match openai spec * Fix __call__ --------- Co-authored-by: Andrei Betlen <abetlen@gmail.com> |
||
---|---|---|
.. | ||
server | ||
__init__.py | ||
_utils.py | ||
llama.py | ||
llama_chat_format.py | ||
llama_cpp.py | ||
llama_grammar.py | ||
llama_types.py | ||
llava_cpp.py | ||
py.typed |