Skip to content

Commit ab01cd1

Browse files
authored
[BugFix] Fix glm4_moe_mtp load weights bug (#28805)
Signed-off-by: wuyaoxuehun <798143193@qq.com>
1 parent 577bb34 commit ab01cd1

File tree

1 file changed

+3
-4
lines changed

1 file changed

+3
-4
lines changed

vllm/model_executor/models/glm4_moe_mtp.py

Lines changed: 3 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -256,13 +256,12 @@ def load_weights(self, weights: Iterable[tuple[str, torch.Tensor]]) -> set[str]:
256256

257257
params_dict = dict(self.named_parameters())
258258
loaded_params: set[str] = set()
259-
spec_layer = self.model.mtp_start_layer_idx
260259
for name, loaded_weight in weights:
261260
if name == "lm_head.weight":
262-
name = f"model.layers.{spec_layer}.shard_head.head.weight"
261+
spec_layer = self.model.mtp_start_layer_idx
262+
name = f"model.layers.{spec_layer}.shared_head.head.weight"
263263
elif name == "model.embed_tokens.weight":
264-
# This name is same with local model, rewriting is not needed.
265-
pass
264+
spec_layer = self.model.mtp_start_layer_idx
266265
else:
267266
spec_layer = get_spec_layer_idx_from_weight_name(self.config, name)
268267
if spec_layer is None:

0 commit comments

Comments
 (0)