From 723059959302d109f3468f2426b442af1469542e Mon Sep 17 00:00:00 2001 From: Andrei Betlen Date: Sun, 23 Apr 2023 14:53:17 -0400 Subject: [PATCH] Disable mmap when applying lora weights. Closes #107 --- llama_cpp/llama.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/llama_cpp/llama.py b/llama_cpp/llama.py index ea9f0ff..70dcea9 100644 --- a/llama_cpp/llama.py +++ b/llama_cpp/llama.py @@ -79,7 +79,7 @@ class Llama: self.params.f16_kv = f16_kv self.params.logits_all = logits_all self.params.vocab_only = vocab_only - self.params.use_mmap = use_mmap + self.params.use_mmap = use_mmap if lora_path is None else False self.params.use_mlock = use_mlock self.params.embedding = embedding