diff --git a/llama_cpp/server/types.py b/llama_cpp/server/types.py index ce9c87a..a20b394 100644 --- a/llama_cpp/server/types.py +++ b/llama_cpp/server/types.py @@ -18,8 +18,6 @@ max_tokens_field = Field( temperature_field = Field( default=0.8, - ge=0.0, - le=2.0, description="Adjust the randomness of the generated text.\n\n" + "Temperature is a hyperparameter that controls the randomness of the generated text. It affects the probability distribution of the model's output tokens. A higher temperature (e.g., 1.5) makes the output more random and creative, while a lower temperature (e.g., 0.5) makes the output more focused, deterministic, and conservative. The default value is 0.8, which provides a balance between randomness and determinism. At the extreme, a temperature of 0 will always pick the most likely next token, leading to identical outputs in each run.", )