From 0d751a69a78c0a2f7b83c894d6a98ceec8daa680 Mon Sep 17 00:00:00 2001 From: Andrei Betlen Date: Mon, 8 May 2023 01:50:43 -0400 Subject: [PATCH] Set repeat_penalty to 0 by default --- llama_cpp/server/app.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/llama_cpp/server/app.py b/llama_cpp/server/app.py index c9f2aef..b459b80 100644 --- a/llama_cpp/server/app.py +++ b/llama_cpp/server/app.py @@ -146,7 +146,7 @@ top_k_field = Field( ) repeat_penalty_field = Field( - default=1.0, + default=0.0, ge=0.0, description="A penalty applied to each token that is already generated. This helps prevent the model from repeating itself.\n\n" + "Repeat penalty is a hyperparameter used to penalize the repetition of token sequences during text generation. It helps prevent the model from generating repetitive or monotonous text. A higher value (e.g., 1.5) will penalize repetitions more strongly, while a lower value (e.g., 0.9) will be more lenient.",