From bded3dce7c1ae114489c5d1832b5d72735aa9d9d Mon Sep 17 00:00:00 2001 From: Saifeddine ALOUI Date: Thu, 18 May 2023 22:23:24 +0200 Subject: [PATCH] enhanced context size --- backends/llama_cpp_official/__init__.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/backends/llama_cpp_official/__init__.py b/backends/llama_cpp_official/__init__.py index c481f66b..7100ebdb 100644 --- a/backends/llama_cpp_official/__init__.py +++ b/backends/llama_cpp_official/__init__.py @@ -38,7 +38,7 @@ class LLAMACPP(GPTBackend): if seed <=0: seed = random.randint(1, 2**31) - self.model = Llama(model_path=f"./models/llama_cpp_official/{self.config['model']}", n_gpu_layers=40, seed=seed) + self.model = Llama(model_path=f"./models/llama_cpp_official/{self.config['model']}", n_ctx=self.config["ctx_size"], n_gpu_layers=40, seed=seed) def tokenize(self, prompt):