From f27393ab7ed06c769aba414dcaf2d544ab0c4c35 Mon Sep 17 00:00:00 2001 From: Andrei Betlen Date: Wed, 14 Jun 2023 21:46:48 -0400 Subject: [PATCH] Add additional verbose logs for cache --- llama_cpp/server/app.py | 4 ++++ 1 file changed, 4 insertions(+) diff --git a/llama_cpp/server/app.py b/llama_cpp/server/app.py index 2191005..e248472 100644 --- a/llama_cpp/server/app.py +++ b/llama_cpp/server/app.py @@ -119,8 +119,12 @@ def create_app(settings: Optional[Settings] = None): ) if settings.cache: if settings.cache_type == "disk": + if settings.verbose: + print(f"Using disk cache with size {settings.cache_size}") cache = llama_cpp.LlamaDiskCache(capacity_bytes=settings.cache_size) else: + if settings.verbose: + print(f"Using ram cache with size {settings.cache_size}") cache = llama_cpp.LlamaRAMCache(capacity_bytes=settings.cache_size) cache = llama_cpp.LlamaCache(capacity_bytes=settings.cache_size)