Skip to content

Commit e937066

Browse files
authoredNov 19, 2023
gguf-py : export chat templates (#4125)
* gguf-py : export chat templates * llama.cpp : escape new lines in gguf kv info prints * gguf-py : bump version * gguf-py : check chat_template type * gguf-py : initialize chat_template
1 parent 28a2e6e commit e937066

File tree

5 files changed

+34
-15
lines changed

5 files changed

+34
-15
lines changed
 

‎gguf-py/gguf/constants.py

+15-14
Original file line numberDiff line numberDiff line change
@@ -56,20 +56,21 @@ class Rope:
5656
SCALING_FINETUNED = "{arch}.rope.scaling.finetuned"
5757

5858
class Tokenizer:
59-
MODEL = "tokenizer.ggml.model"
60-
LIST = "tokenizer.ggml.tokens"
61-
TOKEN_TYPE = "tokenizer.ggml.token_type"
62-
SCORES = "tokenizer.ggml.scores"
63-
MERGES = "tokenizer.ggml.merges"
64-
BOS_ID = "tokenizer.ggml.bos_token_id"
65-
EOS_ID = "tokenizer.ggml.eos_token_id"
66-
UNK_ID = "tokenizer.ggml.unknown_token_id"
67-
SEP_ID = "tokenizer.ggml.seperator_token_id"
68-
PAD_ID = "tokenizer.ggml.padding_token_id"
69-
ADD_BOS = "tokenizer.ggml.add_bos_token"
70-
ADD_EOS = "tokenizer.ggml.add_eos_token"
71-
HF_JSON = "tokenizer.huggingface.json"
72-
RWKV = "tokenizer.rwkv.world"
59+
MODEL = "tokenizer.ggml.model"
60+
LIST = "tokenizer.ggml.tokens"
61+
TOKEN_TYPE = "tokenizer.ggml.token_type"
62+
SCORES = "tokenizer.ggml.scores"
63+
MERGES = "tokenizer.ggml.merges"
64+
BOS_ID = "tokenizer.ggml.bos_token_id"
65+
EOS_ID = "tokenizer.ggml.eos_token_id"
66+
UNK_ID = "tokenizer.ggml.unknown_token_id"
67+
SEP_ID = "tokenizer.ggml.seperator_token_id"
68+
PAD_ID = "tokenizer.ggml.padding_token_id"
69+
ADD_BOS = "tokenizer.ggml.add_bos_token"
70+
ADD_EOS = "tokenizer.ggml.add_eos_token"
71+
HF_JSON = "tokenizer.huggingface.json"
72+
RWKV = "tokenizer.rwkv.world"
73+
CHAT_TEMPLATE = "tokenizer.chat_template"
7374

7475

7576
#

‎gguf-py/gguf/gguf_writer.py

+3
Original file line numberDiff line numberDiff line change
@@ -399,6 +399,9 @@ def add_add_bos_token(self, value: bool) -> None:
399399
def add_add_eos_token(self, value: bool) -> None:
400400
self.add_bool(Keys.Tokenizer.ADD_EOS, value)
401401

402+
def add_chat_template(self, value: str) -> None:
403+
self.add_string(Keys.Tokenizer.CHAT_TEMPLATE, value)
404+
402405
def _pack(self, fmt: str, value: Any, skip_pack_prefix: bool = False) -> bytes:
403406
pack_prefix = ''
404407
if not skip_pack_prefix:

‎gguf-py/gguf/vocab.py

+14
Original file line numberDiff line numberDiff line change
@@ -13,6 +13,7 @@ class SpecialVocab:
1313
merges: list[str]
1414
add_special_token: dict[str, bool]
1515
special_token_ids: dict[str, int]
16+
chat_template: str | None
1617

1718
def __init__(
1819
self, path: str | os.PathLike[str], load_merges: bool = False,
@@ -24,6 +25,7 @@ def __init__(
2425
self.n_vocab = n_vocab
2526
self.load_merges = load_merges
2627
self.merges = []
28+
self.chat_template = None
2729
if special_token_types is not None:
2830
self.special_token_types = special_token_types
2931
else:
@@ -67,6 +69,10 @@ def add_to_gguf(self, gw: GGUFWriter, quiet: bool = False) -> None:
6769
if not quiet:
6870
print(f'gguf: Setting add_{typ}_token to {value}')
6971
add_handler(value)
72+
if self.chat_template is not None:
73+
if not quiet:
74+
print(f'gguf: Setting chat_template to {self.chat_template}')
75+
gw.add_chat_template(self.chat_template)
7076

7177
def _load(self, path: Path) -> None:
7278
self._try_load_from_tokenizer_json(path)
@@ -132,6 +138,14 @@ def _try_load_from_tokenizer_json(self, path: Path) -> bool:
132138
return True
133139
with open(tokenizer_config_file, encoding = 'utf-8') as f:
134140
tokenizer_config = json.load(f)
141+
chat_template = tokenizer_config.get('chat_template')
142+
if chat_template is None or isinstance(chat_template, str):
143+
self.chat_template = chat_template
144+
else:
145+
print(
146+
f'gguf: WARNING: Bad type for chat_template field in {tokenizer_config_file!r} - ignoring',
147+
file = sys.stderr
148+
)
135149
for typ in self.special_token_types:
136150
add_entry = tokenizer_config.get(f'add_{typ}_token')
137151
if isinstance(add_entry, bool):

‎gguf-py/pyproject.toml

+1-1
Original file line numberDiff line numberDiff line change
@@ -1,6 +1,6 @@
11
[tool.poetry]
22
name = "gguf"
3-
version = "0.5.3"
3+
version = "0.6.0"
44
description = "Read and write ML models in GGUF for GGML"
55
authors = ["GGML <ggml@ggml.ai>"]
66
packages = [

‎llama.cpp

+1
Original file line numberDiff line numberDiff line change
@@ -1871,6 +1871,7 @@ struct llama_model_loader {
18711871
if (value.size() > MAX_VALUE_LEN) {
18721872
value = format("%s...", value.substr(0, MAX_VALUE_LEN - 3).c_str());
18731873
}
1874+
replace_all(value, "\n", "\\n");
18741875

18751876
LLAMA_LOG_INFO("%s: - kv %3d: %42s %-16s = %s\n", __func__, i, name, type_name.c_str(), value.c_str());
18761877
}

0 commit comments

Comments
 (0)
Please sign in to comment.