From 3faa07e99df9708605ea687e8cd90b1c40fa0276 Mon Sep 17 00:00:00 2001 From: Maximilian Kimmich <maximilian.kimmich@ims.uni-stuttgart.de> Date: Tue, 1 Oct 2024 18:19:37 +0200 Subject: [PATCH] Offload all layers to GPU for llama_cpp --- evoprompt/models.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/evoprompt/models.py b/evoprompt/models.py index 2fa83f6..41b88a8 100644 --- a/evoprompt/models.py +++ b/evoprompt/models.py @@ -209,7 +209,7 @@ class Llama(LLMModel): chat_handler: str | None = None, verbose: int | bool = False, llama_verbose: bool = False, - n_gpu_layers: int = 60, + n_gpu_layers: int = -1, n_threads: int = 8, n_ctx: int = 8192, **kwargs, -- GitLab