From faafe780d8fb9cd30a992adf45a68ed1b55adcac Mon Sep 17 00:00:00 2001 From: Jhen Date: Tue, 7 Nov 2023 10:03:13 +0800 Subject: [PATCH] talk-llama : fix n_gpu_layers usage again --- examples/talk-llama/talk-llama.cpp | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/examples/talk-llama/talk-llama.cpp b/examples/talk-llama/talk-llama.cpp index bc0119a7f89..8c41ef542dc 100644 --- a/examples/talk-llama/talk-llama.cpp +++ b/examples/talk-llama/talk-llama.cpp @@ -267,7 +267,7 @@ int main(int argc, char ** argv) { auto lmparams = llama_model_default_params(); if (!params.use_gpu) { - lcparams.lmparams = 0; + lmparams.n_gpu_layers = 0; } struct llama_model * model_llama = llama_load_model_from_file(params.model_llama.c_str(), lmparams);