Skip to content

Commit 7dd034d

Browse files
Fix 'HPUCompressedTensorsWNA16MoEMethod' object has no attribute 'fused_experts'
Signed-off-by: Paweł Olejniczak <polejniczakx@habana.ai>
1 parent 257dada commit 7dd034d

File tree

2 files changed

+0
-2
lines changed

2 files changed

+0
-2
lines changed

vllm_gaudi/ops/hpu_compressed_tensors.py

Lines changed: 0 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -654,7 +654,6 @@ def apply(
654654
logical_to_physical_map: Optional[torch.Tensor] = None,
655655
logical_replica_count: Optional[torch.Tensor] = None,
656656
) -> Union[torch.Tensor, tuple[torch.Tensor, torch.Tensor]]:
657-
assert self.fused_experts is None
658657

659658
if enable_eplb:
660659
raise NotImplementedError("EPLB not supported for "

vllm_gaudi/ops/hpu_fp8.py

Lines changed: 0 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -72,7 +72,6 @@ def __init__(self, quant_config: Fp8Config, layer: torch.nn.Module):
7272
self.allow_deep_gemm = False
7373

7474
self.topk_indices_dtype = None
75-
self.fused_experts = None
7675

7776
def create_weights(self, *args, **kwargs) -> None:
7877
if hpu_ops.is_hpu_gaudi2:

0 commit comments

Comments
 (0)