Skip to content

Commit 38ca305

Browse files
fix the layer name (#2247)
1 parent d561f07 commit 38ca305

File tree

2 files changed

+9
-9
lines changed

2 files changed

+9
-9
lines changed

keras_hub/src/utils/transformers/convert_mixtral.py

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -50,19 +50,19 @@ def transpose_and_reshape(x, shape):
5050
# Attention layers
5151
## Query
5252
loader.port_weight(
53-
keras_variable=decoder_layer._self_attention_layer._query_dense.kernel,
53+
keras_variable=decoder_layer._self_attention_layer.query_dense.kernel,
5454
hf_weight_key=f"model.layers.{i}.self_attn.q_proj.weight",
5555
hook_fn=transpose_and_reshape,
5656
)
5757
## Key
5858
loader.port_weight(
59-
keras_variable=decoder_layer._self_attention_layer._key_dense.kernel,
59+
keras_variable=decoder_layer._self_attention_layer.key_dense.kernel,
6060
hf_weight_key=f"model.layers.{i}.self_attn.k_proj.weight",
6161
hook_fn=transpose_and_reshape,
6262
)
6363
## Value
6464
loader.port_weight(
65-
keras_variable=decoder_layer._self_attention_layer._value_dense.kernel,
65+
keras_variable=decoder_layer._self_attention_layer.value_dense.kernel,
6666
hf_weight_key=f"model.layers.{i}.self_attn.v_proj.weight",
6767
hook_fn=transpose_and_reshape,
6868
)

keras_hub/src/utils/transformers/convert_qwen_moe.py

Lines changed: 6 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -64,34 +64,34 @@ def transpose_and_reshape(x, shape):
6464

6565
## Query
6666
loader.port_weight(
67-
keras_variable=decoder_layer._self_attention_layer._query_dense.kernel,
67+
keras_variable=decoder_layer._self_attention_layer.query_dense.kernel,
6868
hf_weight_key=f"model.layers.{i}.self_attn.q_proj.weight",
6969
hook_fn=transpose_and_reshape,
7070
)
7171
loader.port_weight(
72-
keras_variable=decoder_layer._self_attention_layer._query_dense.bias,
72+
keras_variable=decoder_layer._self_attention_layer.query_dense.bias,
7373
hf_weight_key=f"model.layers.{i}.self_attn.q_proj.bias",
7474
hook_fn=transpose_and_reshape,
7575
)
7676
## Key
7777
loader.port_weight(
78-
keras_variable=decoder_layer._self_attention_layer._key_dense.kernel,
78+
keras_variable=decoder_layer._self_attention_layer.key_dense.kernel,
7979
hf_weight_key=f"model.layers.{i}.self_attn.k_proj.weight",
8080
hook_fn=transpose_and_reshape,
8181
)
8282
loader.port_weight(
83-
keras_variable=decoder_layer._self_attention_layer._key_dense.bias,
83+
keras_variable=decoder_layer._self_attention_layer.key_dense.bias,
8484
hf_weight_key=f"model.layers.{i}.self_attn.k_proj.bias",
8585
hook_fn=transpose_and_reshape,
8686
)
8787
## Value
8888
loader.port_weight(
89-
keras_variable=decoder_layer._self_attention_layer._value_dense.kernel,
89+
keras_variable=decoder_layer._self_attention_layer.value_dense.kernel,
9090
hf_weight_key=f"model.layers.{i}.self_attn.v_proj.weight",
9191
hook_fn=transpose_and_reshape,
9292
)
9393
loader.port_weight(
94-
keras_variable=decoder_layer._self_attention_layer._value_dense.bias,
94+
keras_variable=decoder_layer._self_attention_layer.value_dense.bias,
9595
hf_weight_key=f"model.layers.{i}.self_attn.v_proj.bias",
9696
hook_fn=transpose_and_reshape,
9797
)

0 commit comments

Comments
 (0)