We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent e17e448 commit 7bab4bbCopy full SHA for 7bab4bb
benchmarks/kernels/benchmark_moe.py
@@ -509,6 +509,11 @@ def main(args: argparse.Namespace):
509
intermediate_size = config.moe_intermediate_size
510
shard_intermediate_size = 2 * intermediate_size // args.tp_size
511
block_quant_shape = config.quantization_config['weight_block_size']
512
+ elif config.architectures[0] == "Qwen2MoeForCausalLM":
513
+ E = config.num_experts
514
+ topk = config.num_experts_per_tok
515
+ intermediate_size = config.moe_intermediate_size
516
+ shard_intermediate_size = 2 * intermediate_size // args.tp_size
517
else:
518
# Default: Mixtral.
519
E = config.num_local_experts
0 commit comments