2024-01-25 21:46:20 +00:00
|
|
|
diff --git a/examples/server/server.cpp b/examples/server/server.cpp
|
2024-02-12 16:10:16 +00:00
|
|
|
index d86d7e04..2694e92e 100644
|
2024-01-25 21:46:20 +00:00
|
|
|
--- a/examples/server/server.cpp
|
|
|
|
+++ b/examples/server/server.cpp
|
2024-02-12 16:10:16 +00:00
|
|
|
@@ -901,13 +901,15 @@ struct llama_server_context
|
|
|
|
slot.sent_count += result.text_to_send.size();
|
|
|
|
// add the token to slot queue and cache
|
|
|
|
}
|
|
|
|
- slot.add_token_string(result);
|
2024-01-25 21:46:20 +00:00
|
|
|
+
|
2024-02-12 16:10:16 +00:00
|
|
|
if (slot.params.stream)
|
|
|
|
{
|
|
|
|
send_partial_response(slot, result);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
+ slot.add_token_string(result);
|
2024-01-25 21:46:20 +00:00
|
|
|
+
|
2024-02-12 16:10:16 +00:00
|
|
|
if (incomplete)
|
|
|
|
{
|
|
|
|
slot.has_next_token = true;
|