File tree Expand file tree Collapse file tree 2 files changed +7
-2
lines changed
tests/py/dynamo/distributed Expand file tree Collapse file tree 2 files changed +7
-2
lines changed Original file line number Diff line number Diff line change 8080 runs-on : ${{ matrix.validation_runner }}
8181 container :
8282 image : ${{ matrix.container_image }}
83- options : ${{ matrix.gpu_arch_type == 'cuda' && '--gpus all' || ' ' }}
83+ options : ${{ matrix.gpu_arch_type == 'cuda' && '--gpus all --shm-size=1g ' || ' ' }}
8484 # If a build is taking longer than 120 minutes on these runners we need
8585 # to have a conversation
8686 timeout-minutes : 120
Original file line number Diff line number Diff line change 1111)
1212from parameterized import parameterized
1313from torch .testing ._internal .common_utils import run_tests
14+ from torch_tensorrt ._features import ENABLED_FEATURES
1415
1516
1617def is_distributed_nccl_available ():
@@ -75,11 +76,15 @@ def forward(self, x):
7576
7677class TestNcclOpsConverter (DispatchTestCase ):
7778 # 1. Skip if NCCL backend is not available (e.g., Windows, Jetson) - hard requirement
78- # 2. Don't skip if TRTLLM is unavailable (e.g., CUDA 13) - falls back to PyTorch
79+ # 2. Skip if TRTLLM is unavailable (e.g., CUDA 13) - no converters registered
7980 @unittest .skipIf (
8081 not is_distributed_nccl_available (),
8182 "Skipped: NCCL backend is not available (Windows/Jetson Orin not supported)." ,
8283 )
84+ @unittest .skipIf (
85+ not ENABLED_FEATURES .trtllm_for_nccl ,
86+ "Skipped: TensorRT-LLM plugin for NCCL is not available (e.g., CUDA 13)." ,
87+ )
8388 @classmethod
8489 def setUpClass (cls ):
8590 cls .world_size = int (os .environ .get ("OMPI_COMM_WORLD_SIZE" , 1 ))
You can’t perform that action at this time.
0 commit comments