[https://nvbugs/5916151][fix] Unwaive test_fused_moe_w4a8_nvfp4_fp8[TRTLLM]#12400
[https://nvbugs/5916151][fix] Unwaive test_fused_moe_w4a8_nvfp4_fp8[TRTLLM]#12400EmmaQiaoCh merged 1 commit intoNVIDIA:mainNVIDIA/TensorRT-LLM:mainfrom xxi-nv:unwaive-nvbug-5916151xxi-nv/TensorRT-LLM:unwaive-nvbug-5916151Copy head branch name to clipboard
Conversation
…RTLLM] The illegal memory access bug (NVBug 5916151) was fixed by PR NVIDIA#11502 which corrected the .cuda() call ordering in the test. Verified on B300 with 7/7 passes (1 initial + 1 CUTLASS + 5 repeat runs). Signed-off-by: xxi <xxi@nvidia.com>
📝 WalkthroughWalkthroughOne integration test waiver entry is removed from the test waiver list, allowing the previously skipped test case Changes
Estimated code review effort🎯 2 (Simple) | ⏱️ ~10 minutes 🚥 Pre-merge checks | ✅ 3✅ Passed checks (3 passed)
✏️ Tip: You can configure your own custom pre-merge checks in the settings. ✨ Finishing Touches🧪 Generate unit tests (beta)
📝 Coding Plan
Comment Tip CodeRabbit can suggest fixes for GitHub Check annotations.Configure the |
|
/bot run --disable-fail-fast |
|
PR_Github #39707 [ run ] triggered by Bot. Commit: |
|
/bot run --stage-list "DGX_B300-4_GPUs-PyTorch-Post-Merge-2" |
|
PR_Github #39710 [ run ] triggered by Bot. Commit: |
|
PR_Github #39710 [ run ] completed with state |
|
/bot run --disable-fail-fast |
|
PR_Github #39734 [ run ] triggered by Bot. Commit: |
|
PR_Github #39734 [ run ] completed with state |
|
/bot run --disable-fail-fast |
|
PR_Github #39847 [ run ] triggered by Bot. Commit: |
|
PR_Github #39847 [ run ] completed with state |
…RTLLM] (NVIDIA#12400) Signed-off-by: xxi <xxi@nvidia.com>
Summary
test_fused_moe_w4a8_nvfp4_fp8[TRTLLM](NVBug 5916151).cuda()call ordering in the test'srun_fused_moe_nvfp4()functionTest plan
Summary by CodeRabbit