@@ -1310,6 +1310,7 @@ def set_vocab(self):
1310
1310
special_vocab ._set_special_token ("prefix" , 1 )
1311
1311
special_vocab ._set_special_token ("suffix" , 3 )
1312
1312
special_vocab ._set_special_token ("middle" , 2 )
1313
+ special_vocab .chat_template = None # do not add it twice
1313
1314
special_vocab .add_to_gguf (self .gguf_writer )
1314
1315
1315
1316
def set_gguf_parameters (self ):
@@ -2466,13 +2467,7 @@ class GemmaModel(Model):
2466
2467
model_arch = gguf .MODEL_ARCH .GEMMA
2467
2468
2468
2469
def set_vocab (self ):
2469
- tokens , scores , toktypes = self ._create_vocab_sentencepiece ()
2470
-
2471
- self .gguf_writer .add_tokenizer_model ("llama" )
2472
- self .gguf_writer .add_tokenizer_pre ("default" )
2473
- self .gguf_writer .add_token_list (tokens )
2474
- self .gguf_writer .add_token_scores (scores )
2475
- self .gguf_writer .add_token_types (toktypes )
2470
+ self ._set_vocab_sentencepiece ()
2476
2471
2477
2472
# TODO: these special tokens should be exported only for the CodeGemma family
2478
2473
special_vocab = gguf .SpecialVocab (self .dir_model , load_merges = False ,
@@ -2482,6 +2477,7 @@ def set_vocab(self):
2482
2477
special_vocab ._set_special_token ("middle" , 68 )
2483
2478
special_vocab ._set_special_token ("fsep" , 70 )
2484
2479
special_vocab ._set_special_token ("eot" , 107 )
2480
+ special_vocab .chat_template = None # do not add it twice
2485
2481
special_vocab .add_to_gguf (self .gguf_writer )
2486
2482
2487
2483
self .gguf_writer .add_add_space_prefix (False )
0 commit comments