@@ -2416,13 +2416,15 @@ def set_vocab(self):
2416
2416
2417
2417
text = piece .encode ("utf-8" )
2418
2418
score = 0.0
2419
- if len (piece ) != 0 and token_id < 64789 :
2419
+ # Referencing the tokenizer Python implementation(https://huggingface.co/THUDM/chatglm3-6b/blob/main/tokenization_chatglm.py),
2420
+ # it is only valid if it is less than tokenizer.tokenizer.sp_model.vocab_size()
2421
+ if len (piece ) != 0 and token_id < tokenizer .tokenizer .sp_model .vocab_size ():
2420
2422
score = tokenizer .tokenizer .sp_model .get_score (token_id )
2421
2423
2422
2424
if len (piece ) == 0 :
2423
2425
text = f"[PAD{ token_id } ]" .encode ("utf-8" )
2424
2426
2425
- if token_id >= 64789 :
2427
+ if token_id >= tokenizer . tokenizer . sp_model . vocab_size () :
2426
2428
toktype = SentencePieceTokenTypes .UNKNOWN
2427
2429
tokens .append (text )
2428
2430
scores .append (score )
@@ -2452,7 +2454,7 @@ def set_vocab(self):
2452
2454
special_vocab .add_to_gguf (self .gguf_writer )
2453
2455
2454
2456
def set_gguf_parameters (self ):
2455
- self .gguf_writer .add_name ("ChatGLM-6b-chat" )
2457
+ self .gguf_writer .add_name (self . dir_model . name )
2456
2458
n_embed = self .hparams .get ("hidden_size" , self .hparams .get ("n_embed" ))
2457
2459
n_head = self .hparams .get ("n_head" , self .hparams .get ("num_attention_heads" ))
2458
2460
n_head_kv = self .hparams .get ("multi_query_group_num" , n_head )
@@ -2468,16 +2470,12 @@ def set_gguf_parameters(self):
2468
2470
self .gguf_writer .add_add_bos_token (False )
2469
2471
2470
2472
def modify_tensors (self , data_torch : Tensor , name : str , bid : int | None ) -> Iterable [tuple [str , Tensor ]]:
2471
- if name .endswith (".rotary_pos_emb.inv_freq" ):
2472
- return []
2473
-
2474
2473
del bid # unused
2475
2474
2476
- name = re .sub (r'transformer\.' , '' , name )
2477
-
2478
- if name == "word_embeddings.weight" :
2479
- assert self .tensor_names is not None
2475
+ if name .endswith (".rotary_pos_emb.inv_freq" ):
2476
+ return []
2480
2477
2478
+ name = name .removeprefix ("transformer." )
2481
2479
return [(self .map_tensor_name (name ), data_torch )]
2482
2480
2483
2481
0 commit comments