From 6fefe9666022cb6fd8078defcbd61a6b8f278b52 Mon Sep 17 00:00:00 2001 From: KillianLucas <63927363+KillianLucas@users.noreply.github.com> Date: Thu, 24 Aug 2023 00:13:56 +0000 Subject: [PATCH] GPU Experiment --- interpreter/llama_2.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/interpreter/llama_2.py b/interpreter/llama_2.py index 520449c2bd..9ef4385c82 100644 --- a/interpreter/llama_2.py +++ b/interpreter/llama_2.py @@ -97,7 +97,7 @@ def format_choice(model): return None # Initialize and return Llama-2 - llama_2 = Llama(model_path=model_path, n_gpu_layers=-1) + llama_2 = Llama(model_path=model_path, n_gpu_layers=1) return llama_2