Skip to content
This repository was archived by the owner on Jul 4, 2025. It is now read-only.

Commit f9a37cf

Browse files
authored
Merge pull request #79 from janhq/77-feat-add-support-for-inference-parameters
add extra params
2 parents b920e80 + c1dc2d9 commit f9a37cf

File tree

1 file changed

+7
-0
lines changed

1 file changed

+7
-0
lines changed

controllers/llamaCPP.cc

Lines changed: 7 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -58,6 +58,10 @@ void llamaCPP::chatCompletion(
5858
llama.params.n_predict = (*jsonBody)["max_tokens"].asInt();
5959
llama.params.sampling_params.top_p = (*jsonBody)["top_p"].asFloat();
6060
llama.params.sampling_params.temp = (*jsonBody)["temperature"].asFloat();
61+
llama.params.sampling_params.frequency_penalty =
62+
(*jsonBody)["frequency_penalty"].asFloat();
63+
llama.params.sampling_params.presence_penalty =
64+
(*jsonBody)["presence_penalty"].asFloat();
6165

6266
const Json::Value &messages = (*jsonBody)["messages"];
6367
for (const auto &message : messages) {
@@ -74,6 +78,9 @@ void llamaCPP::chatCompletion(
7478

7579
this->llama.prompt = formatted_output;
7680
this->llama.params.antiprompt.clear();
81+
for (const auto &stop_word : (*jsonBody)["stop"]) {
82+
llama.params.antiprompt.push_back(stop_word.asString());
83+
}
7784
this->llama.params.antiprompt.push_back("user:");
7885
this->llama.params.antiprompt.push_back("### USER:");
7986
this->llama.loadPrompt();

0 commit comments

Comments
 (0)