Skip to content

Commit 6e91c4e

Browse files
committed
addressing L2 CI errors
1 parent 6d20d2b commit 6e91c4e

File tree

2 files changed

+7
-2
lines changed

2 files changed

+7
-2
lines changed

.github/workflows/linux-test.yml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -80,7 +80,7 @@ jobs:
8080
runs-on: ${{ matrix.validation_runner }}
8181
container:
8282
image: ${{ matrix.container_image }}
83-
options: ${{ matrix.gpu_arch_type == 'cuda' && '--gpus all' || ' ' }}
83+
options: ${{ matrix.gpu_arch_type == 'cuda' && '--gpus all --shm-size=1g' || ' ' }}
8484
# If a build is taking longer than 120 minutes on these runners we need
8585
# to have a conversation
8686
timeout-minutes: 120

tests/py/dynamo/distributed/test_nccl_ops.py

Lines changed: 6 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -11,6 +11,7 @@
1111
)
1212
from parameterized import parameterized
1313
from torch.testing._internal.common_utils import run_tests
14+
from torch_tensorrt._features import ENABLED_FEATURES
1415

1516

1617
def is_distributed_nccl_available():
@@ -75,11 +76,15 @@ def forward(self, x):
7576

7677
class TestNcclOpsConverter(DispatchTestCase):
7778
# 1. Skip if NCCL backend is not available (e.g., Windows, Jetson) - hard requirement
78-
# 2. Don't skip if TRTLLM is unavailable (e.g., CUDA 13) - falls back to PyTorch
79+
# 2. Skip if TRTLLM is unavailable (e.g., CUDA 13) - no converters registered
7980
@unittest.skipIf(
8081
not is_distributed_nccl_available(),
8182
"Skipped: NCCL backend is not available (Windows/Jetson Orin not supported).",
8283
)
84+
@unittest.skipIf(
85+
not ENABLED_FEATURES.trtllm_for_nccl,
86+
"Skipped: TensorRT-LLM plugin for NCCL is not available (e.g., CUDA 13).",
87+
)
8388
@classmethod
8489
def setUpClass(cls):
8590
cls.world_size = int(os.environ.get("OMPI_COMM_WORLD_SIZE", 1))

0 commit comments

Comments
 (0)