Skip to content

Commit 00c6390

Browse files
jartmofosyne
andauthored
main : don't print special tokens with --grammar (#6923)
* main : don't print special tokens with --grammar The CLI interface was recently changed to print special control tokens like the </s> stop message one. This token shouldn't be printed if the grammar flag was passed, unless the grammar specifies it, because that breaks shell-scriptability. * main: use seperate stream for control characters * main: use dprintf and add --ctrl-token-no-out and --ctrl-token-fd-out * main: dprintf isn't part of the IEEE POSIX standard. Just use write(). * main: remove --ctrl-token-fd-out in favor for fcntl() based detection * common.cpp: accidentally removed --interactive-first * main: only merge stdout and control token if not in conversation or grammar mode * main: rejig control token descriptor handling * main: must check pipe status on very top of program * main: renamed --no-special from --ctrl-token-no-out and other refactoring * main: refactor ctrl_token_no_out --> no_special * llama: rename llama_token_is_control_token() to llama_token_is_control() * main: remove special token file descriptor feature (#5) --------- Co-authored-by: Brian <mofosyne@gmail.com>
1 parent faa0e69 commit 00c6390

File tree

5 files changed

+30
-3
lines changed

5 files changed

+30
-3
lines changed

common/common.cpp

+5
Original file line numberDiff line numberDiff line change
@@ -904,6 +904,10 @@ bool gpt_params_find_arg(int argc, char ** argv, const std::string & arg, gpt_pa
904904
params.interactive_specials = true;
905905
return true;
906906
}
907+
if (arg == "--no-special") {
908+
params.no_special = true;
909+
return true;
910+
}
907911
if (arg == "--embedding") {
908912
params.embedding = true;
909913
return true;
@@ -1364,6 +1368,7 @@ void gpt_params_print_usage(int /*argc*/, char ** argv, const gpt_params & param
13641368
printf(" -i, --interactive run in interactive mode\n");
13651369
printf(" --interactive-specials allow special tokens in user text, in interactive mode\n");
13661370
printf(" --interactive-first run in interactive mode and wait for input right away\n");
1371+
printf(" --no-special control tokens output disabled\n");
13671372
printf(" -cnv, --conversation run in conversation mode (does not print special tokens and suffix/prefix)\n");
13681373
printf(" -ins, --instruct run in instruction mode (use with Alpaca models)\n");
13691374
printf(" -cml, --chatml run in chatml mode (use with ChatML-compatible models)\n");

common/common.h

+1
Original file line numberDiff line numberDiff line change
@@ -146,6 +146,7 @@ struct gpt_params {
146146
bool use_color = false; // use color to distinguish generations and inputs
147147
bool interactive = false; // interactive mode
148148
bool interactive_specials = false; // whether to allow special tokens from user, during interactive mode
149+
bool no_special = false; // disable control token output
149150
bool conversation = false; // conversation mode (does not print special tokens and suffix/prefix)
150151
bool chatml = false; // chatml mode (used for models trained on chatml syntax)
151152
bool prompt_cache_all = false; // save user input and generations to prompt cache

examples/main/main.cpp

+17-3
Original file line numberDiff line numberDiff line change
@@ -740,18 +740,32 @@ int main(int argc, char ** argv) {
740740
// display text
741741
if (input_echo && display) {
742742
for (auto id : embd) {
743-
const std::string token_str = llama_token_to_piece(ctx, id, !params.conversation);
744-
printf("%s", token_str.c_str());
743+
const std::string token_str = llama_token_to_piece(ctx, id);
744+
745+
// Console/Stream Output
746+
if (!llama_token_is_control(llama_get_model(ctx), id)) {
747+
// Stream Output Token To Standard Output
748+
fprintf(stdout, "%s", token_str.c_str());
749+
} else if (!params.no_special && !params.conversation) {
750+
// Stream Control Token To Standard Output Stream
751+
fprintf(stdout, "%s", token_str.c_str());
752+
}
745753

754+
// Record Displayed Tokens To Log
755+
// Note: Generated tokens are created one by one hence this check
746756
if (embd.size() > 1) {
757+
// Incoming Requested Tokens
747758
input_tokens.push_back(id);
748759
} else {
760+
// Outgoing Generated Tokens
749761
output_tokens.push_back(id);
750762
output_ss << token_str;
751763
}
764+
765+
fflush(stdout);
752766
}
753-
fflush(stdout);
754767
}
768+
755769
// reset color to default if there is no pending user input
756770
if (input_echo && (int) embd_inp.size() == n_consumed) {
757771
console::set_display(console::reset);

llama.cpp

+4
Original file line numberDiff line numberDiff line change
@@ -17861,6 +17861,10 @@ bool llama_token_is_eog(const struct llama_model * model, llama_token token) {
1786117861
);
1786217862
}
1786317863

17864+
bool llama_token_is_control(const struct llama_model * model, llama_token token) {
17865+
return llama_is_control_token(model->vocab, token);
17866+
}
17867+
1786417868
llama_token llama_token_bos(const struct llama_model * model) {
1786517869
return model->vocab.special_bos_id;
1786617870
}

llama.h

+3
Original file line numberDiff line numberDiff line change
@@ -823,6 +823,9 @@ extern "C" {
823823
// Check if the token is supposed to end generation (end-of-generation, eg. EOS, EOT, etc.)
824824
LLAMA_API bool llama_token_is_eog(const struct llama_model * model, llama_token token);
825825

826+
// Identify if Token Id is a control token or a render-able token
827+
LLAMA_API bool llama_token_is_control(const struct llama_model * model, llama_token token);
828+
826829
// Special tokens
827830
LLAMA_API llama_token llama_token_bos(const struct llama_model * model); // beginning-of-sentence
828831
LLAMA_API llama_token llama_token_eos(const struct llama_model * model); // end-of-sentence

0 commit comments

Comments
 (0)