Skip to content

Commit

Permalink
server: fix print usage LF in new --n-predict option
Browse files Browse the repository at this point in the history
  • Loading branch information
phymbert committed Feb 17, 2024
1 parent 8852de3 commit b45e111
Showing 1 changed file with 1 addition and 1 deletion.
2 changes: 1 addition & 1 deletion examples/server/server.cpp
Original file line number Diff line number Diff line change
Expand Up @@ -1926,14 +1926,14 @@ static void server_print_usage(const char *argv0, const gpt_params &params,
printf(" --mmproj MMPROJ_FILE path to a multimodal projector file for LLaVA.\n");
printf(" --log-disable disables logging to a file.\n");
printf("\n");
printf(" -n, --n-predict maximum tokens to predict (default: %d)\n", params.n_predict);
printf(" --override-kv KEY=TYPE:VALUE\n");
printf(" advanced option to override model metadata by key. may be specified multiple times.\n");
printf(" types: int, float, bool. example: --override-kv tokenizer.ggml.add_bos_token=bool:false\n");
printf(" -gan N, --grp-attn-n N set the group attention factor to extend context size through self-extend(default: 1=disabled), used together with group attention width `--grp-attn-w`");
printf(" -gaw N, --grp-attn-w N set the group attention width to extend context size through self-extend(default: 512), used together with group attention factor `--grp-attn-n`");
printf(" --chat-template FORMAT_NAME");
printf(" set chat template, possible values is: llama2, chatml (default %s)", sparams.chat_template.c_str());
printf(" -n, --n-predict maximum tokens to predict (default: %d)\n", params.n_predict);
printf("\n");
}

Expand Down

0 comments on commit b45e111

Please sign in to comment.