Skip to content

Commit

Permalink
[Tokenizer] Fix TokenizerFast missing clean_up_tokenization_spaces (P…
Browse files Browse the repository at this point in the history
  • Loading branch information
dynamicheart authored and lvdongyi committed Oct 23, 2024
1 parent 35c2c21 commit 6ea6847
Showing 1 changed file with 1 addition and 1 deletion.
2 changes: 1 addition & 1 deletion paddlenlp/transformers/tokenizer_utils_base.py
Original file line number Diff line number Diff line change
Expand Up @@ -1389,7 +1389,7 @@ def __init__(self, **kwargs):
self.model_input_names = kwargs.pop("model_input_names", self.model_input_names)

# By default, cleaning tokenization spaces for both fast and slow tokenizers
self.clean_up_tokenization_spaces = kwargs.pop("clean_up_tokenization_spaces", True)
self.clean_up_tokenization_spaces = kwargs.pop("clean_up_tokenization_spaces", False)

# By default, do not split special tokens for both fast and slow tokenizers
self.split_special_tokens = kwargs.pop("split_special_tokens", False)
Expand Down

0 comments on commit 6ea6847

Please sign in to comment.