Skip to content

Commit 542ed6a

Browse files
author
liyang
committed
fix: Gemma2/Gemma3 inference issue caused by ln_1/ln_2 keys (switch to ln1/ln2)
1 parent af8d0bf commit 542ed6a

File tree

2 files changed

+7
-7
lines changed

2 files changed

+7
-7
lines changed

convert_hf_to_gguf.py

Lines changed: 5 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -4897,7 +4897,7 @@ def _xlmroberta_set_vocab(self) -> None:
48974897
with open(tokenizer_config_path, "r", encoding="utf-8") as fp:
48984898
tokenizer_config_json = json.load(fp)
48994899

4900-
add_prefix = tokenizer.add_prefix_space
4900+
add_prefix = getattr(tokenizer, "add_prefix_space", False)
49014901
remove_whitespaces = tokenizer.clean_up_tokenization_spaces
49024902
precompiled_charsmap = b64decode(tokenizer_json["normalizer"]["precompiled_charsmap"])
49034903

@@ -6508,10 +6508,10 @@ def _map_block_tensor(self, layer: int, rest: str, data_torch: Tensor, name: str
65086508
# layer norms
65096509
if rest.startswith('norm1.'):
65106510
suffix = parts[-1]
6511-
return [(f'v.blk.{layer}.ln_1.{suffix}', data_torch)]
6511+
return [(f'v.blk.{layer}.ln1.{suffix}', data_torch)]
65126512
if rest.startswith('norm2.'):
65136513
suffix = parts[-1]
6514-
return [(f'v.blk.{layer}.ln_2.{suffix}', data_torch)]
6514+
return [(f'v.blk.{layer}.ln2.{suffix}', data_torch)]
65156515
if rest.startswith('attn.inner_attn_ln.'):
65166516
suffix = parts[-1]
65176517
return [(f'v.blk.{layer}.attn_ln.{suffix}', data_torch)]
@@ -6647,8 +6647,8 @@ def _should_be_f32(self, gguf_name: str) -> bool:
66476647
binary-op dtype issues; patch embedding bias is also safer as F32.
66486648
"""
66496649
patterns = (
6650-
".ln_1.weight", ".ln_1.bias",
6651-
".ln_2.weight", ".ln_2.bias",
6650+
".ln1.weight", ".ln1.bias",
6651+
".ln2.weight", ".ln2.bias",
66526652
".attn_ln.weight", ".attn_ln.bias",
66536653
".ffn_norm.weight", ".ffn_norm.bias",
66546654
"v.patch_embd.proj.bias",

tools/mtmd/clip-impl.h

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -76,8 +76,8 @@
7676
#define TN_FFN_GATE "%s.blk.%d.ffn_gate.%s"
7777
#define TN_FFN_UP "%s.blk.%d.ffn_up.%s"
7878
#define TN_FFN_NORM "%s.blk.%d.ffn_norm.%s"
79-
#define TN_LN_1 "%s.blk.%d.ln_1.%s" // layer norm
80-
#define TN_LN_2 "%s.blk.%d.ln_2.%s" // layer norm
79+
#define TN_LN_1 "%s.blk.%d.ln1.%s" // layer norm
80+
#define TN_LN_2 "%s.blk.%d.ln2.%s" // layer norm
8181
#define TN_LS_1 "%s.blk.%d.ls1.%s" // layer scale
8282
#define TN_LS_2 "%s.blk.%d.ls2.%s" // layer scale
8383
#define TN_LN_PRE "%s.pre_ln.%s"

0 commit comments

Comments
 (0)