Skip to content

Commit 41fcead

Browse files
authored
Add qwen3moe meta loading for AutoTP (#7297)
Enable Qwen3-Moe meta loading for AutoTP, for issue #7275 Signed-off-by: ranzhejiang <[email protected]>
1 parent 0e74171 commit 41fcead

File tree

1 file changed

+2
-2
lines changed

1 file changed

+2
-2
lines changed

deepspeed/module_inject/auto_tp.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -137,8 +137,8 @@ def is_load_module(module):
137137
"LPLayerNorm", "SharedEmbedding", "OPTLearnedPositionalEmbedding", "LlamaRMSNorm", "FalconLinear",
138138
"MistralRMSNorm", "T5LayerNorm", "MixtralRMSNorm", "Phi3RotaryEmbedding", "Phi3SuScaledRotaryEmbedding",
139139
"Phi3RMSNorm", "YuanRMSNorm", "YuanRotaryEmbedding", "Phi3LongRoPEScaledRotaryEmbedding", "Qwen2RMSNorm",
140-
"Qwen3RMSNorm", "DeepseekV2RMSNorm", "DeepseekV3RMSNorm", "DeepseekV2YarnRotaryEmbedding",
141-
"DeepseekV3YarnRotaryEmbedding", "MoEGate"
140+
"Qwen3RMSNorm", "Qwen3MoeRMSNorm", "DeepseekV2RMSNorm", "DeepseekV3RMSNorm",
141+
"DeepseekV2YarnRotaryEmbedding", "DeepseekV3YarnRotaryEmbedding", "MoEGate"
142142
]
143143
return module.__class__ in load_layers or module._get_name() in load_layer_names
144144

0 commit comments

Comments
 (0)