@@ -4897,7 +4897,7 @@ def _xlmroberta_set_vocab(self) -> None:
48974897 with open (tokenizer_config_path , "r" , encoding = "utf-8" ) as fp :
48984898 tokenizer_config_json = json .load (fp )
48994899
4900- add_prefix = tokenizer . add_prefix_space
4900+ add_prefix = getattr ( tokenizer , " add_prefix_space" , False )
49014901 remove_whitespaces = tokenizer .clean_up_tokenization_spaces
49024902 precompiled_charsmap = b64decode (tokenizer_json ["normalizer" ]["precompiled_charsmap" ])
49034903
@@ -6508,10 +6508,10 @@ def _map_block_tensor(self, layer: int, rest: str, data_torch: Tensor, name: str
65086508 # layer norms
65096509 if rest .startswith ('norm1.' ):
65106510 suffix = parts [- 1 ]
6511- return [(f'v.blk.{ layer } .ln_1 .{ suffix } ' , data_torch )]
6511+ return [(f'v.blk.{ layer } .ln1 .{ suffix } ' , data_torch )]
65126512 if rest .startswith ('norm2.' ):
65136513 suffix = parts [- 1 ]
6514- return [(f'v.blk.{ layer } .ln_2 .{ suffix } ' , data_torch )]
6514+ return [(f'v.blk.{ layer } .ln2 .{ suffix } ' , data_torch )]
65156515 if rest .startswith ('attn.inner_attn_ln.' ):
65166516 suffix = parts [- 1 ]
65176517 return [(f'v.blk.{ layer } .attn_ln.{ suffix } ' , data_torch )]
@@ -6647,8 +6647,8 @@ def _should_be_f32(self, gguf_name: str) -> bool:
66476647 binary-op dtype issues; patch embedding bias is also safer as F32.
66486648 """
66496649 patterns = (
6650- ".ln_1 .weight" , ".ln_1 .bias" ,
6651- ".ln_2 .weight" , ".ln_2 .bias" ,
6650+ ".ln1 .weight" , ".ln1 .bias" ,
6651+ ".ln2 .weight" , ".ln2 .bias" ,
66526652 ".attn_ln.weight" , ".attn_ln.bias" ,
66536653 ".ffn_norm.weight" , ".ffn_norm.bias" ,
66546654 "v.patch_embd.proj.bias" ,
0 commit comments