feat: Use all available CPUs for batch processing (#1345)
This commit is contained in:
parent
a420f9608b
commit
c96b2daebf
2 changed files with 2 additions and 4 deletions
|
@ -262,9 +262,7 @@ class Llama:
|
|||
|
||||
self.n_batch = min(n_ctx, n_batch) # ???
|
||||
self.n_threads = n_threads or max(multiprocessing.cpu_count() // 2, 1)
|
||||
self.n_threads_batch = n_threads_batch or max(
|
||||
multiprocessing.cpu_count() // 2, 1
|
||||
)
|
||||
self.n_threads_batch = n_threads_batch or multiprocessing.cpu_count()
|
||||
|
||||
# Context Params
|
||||
self.context_params = llama_cpp.llama_context_default_params()
|
||||
|
|
|
@ -70,7 +70,7 @@ class ModelSettings(BaseSettings):
|
|||
description="The number of threads to use.",
|
||||
)
|
||||
n_threads_batch: int = Field(
|
||||
default=max(multiprocessing.cpu_count() // 2, 1),
|
||||
default=max(multiprocessing.cpu_count(), 1),
|
||||
ge=0,
|
||||
description="The number of threads to use when batch processing.",
|
||||
)
|
||||
|
|
Loading…
Reference in a new issue