Skip to content
This repository was archived by the owner on Jul 4, 2025. It is now read-only.

Commit b920e80

Browse files
authored
Merge pull request #78 from janhq/llama-cpp-upgrade
pump llama cpp latest version
2 parents 3bea1d6 + 7fde10b commit b920e80

File tree

3 files changed

+152
-206
lines changed

3 files changed

+152
-206
lines changed

controllers/llamaCPP.cc

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -56,8 +56,8 @@ void llamaCPP::chatCompletion(
5656
"Below is a conversation between an AI system named ASSISTANT and USER\n";
5757
if (jsonBody) {
5858
llama.params.n_predict = (*jsonBody)["max_tokens"].asInt();
59-
llama.params.top_p = (*jsonBody)["top_p"].asFloat();
60-
llama.params.temp = (*jsonBody)["temperature"].asFloat();
59+
llama.params.sampling_params.top_p = (*jsonBody)["top_p"].asFloat();
60+
llama.params.sampling_params.temp = (*jsonBody)["temperature"].asFloat();
6161

6262
const Json::Value &messages = (*jsonBody)["messages"];
6363
for (const auto &message : messages) {
@@ -133,7 +133,7 @@ void llamaCPP::chatCompletion(
133133

134134
std::vector<completion_token_output> probs_output = {};
135135

136-
if (this->llama.params.n_probs > 0) {
136+
if (this->llama.params.sampling_params.n_probs > 0) {
137137
const std::vector<llama_token> to_send_toks =
138138
llama_tokenize(llama.ctx, to_send, false);
139139
size_t probs_pos =

0 commit comments

Comments
 (0)