Skip to content

Commit c9aa9ef

Browse files
xinhe-nvWong4j
authored andcommitted
[https://nvbugs/5519462][fix] Add failed cases into waives.txt (NVIDIA#7817)
Signed-off-by: xinhe-nv <[email protected]> Signed-off-by: Xin He (SW-GPU) <[email protected]>
1 parent 8320ae5 commit c9aa9ef

File tree

3 files changed

+5
-12
lines changed

3 files changed

+5
-12
lines changed

tests/integration/defs/accuracy/test_disaggregated_serving.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -697,9 +697,9 @@ def test_nixl_backend(self):
697697
task.evaluate(llm)
698698

699699
@pytest.mark.skip_less_device(8)
700+
@skip_pre_hopper
700701
@parametrize_with_ids("overlap_scheduler", [True, False])
701-
@parametrize_with_ids("mtp_nextn",
702-
[0, pytest.param(2, marks=skip_pre_hopper)])
702+
@parametrize_with_ids("mtp_nextn", [0, 2])
703703
@pytest.mark.skip_less_device(8)
704704
def test_auto_dtype(self, overlap_scheduler, mtp_nextn):
705705
ctx_server_config = {"disable_overlap_scheduler": True}

tests/integration/defs/accuracy/test_llm_api_pytorch.py

Lines changed: 3 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -1130,23 +1130,21 @@ def test_nvfp4_tp2(self):
11301130
# This class has extensively parameterized test methods, which yield totally 200 test cases.
11311131
# This is because this model requires high test coverage over the feature combinations.
11321132
# Normally we should not parameterize test methods so extensively -- just test on the typical/important feature combinations.
1133+
@skip_pre_hopper
11331134
class TestDeepSeekV3Lite(LlmapiAccuracyTestHarness):
11341135
MODEL_NAME = "deepseek-ai/DeepSeek-V3-Lite"
11351136
MODEL_PATH = f"{llm_models_root()}/DeepSeek-V3-Lite/bf16"
11361137

11371138
@pytest.mark.skip_less_device_memory(60000)
11381139
# Chunked Prefill for MLA can only be enabled on SM100
1139-
@parametrize_with_ids(
1140-
"enable_chunked_prefill",
1141-
[False, pytest.param(True, marks=skip_pre_hopper)])
1140+
@parametrize_with_ids("enable_chunked_prefill", [False, True])
11421141
@parametrize_with_ids("torch_compile", [False, True])
11431142
@parametrize_with_ids("attention_dp,cuda_graph,overlap_scheduler",
11441143
[(False, False, False), (True, False, False),
11451144
(False, True, False), (False, False, True),
11461145
(False, True, True), (True, True, True)])
11471146
# Only Hopper and Blackwell MLA kernel supports MTP
1148-
@parametrize_with_ids("mtp_nextn",
1149-
[0, pytest.param(2, marks=skip_pre_hopper)])
1147+
@parametrize_with_ids("mtp_nextn", [0, 2])
11501148
def test_bfloat16(self, mtp_nextn, attention_dp, cuda_graph,
11511149
overlap_scheduler, torch_compile, enable_chunked_prefill):
11521150
kv_cache_config = KvCacheConfig(free_gpu_memory_fraction=0.75)

tests/integration/test_lists/waives.txt

Lines changed: 0 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -283,9 +283,7 @@ disaggregated/test_disaggregated.py::test_disaggregated_diff_max_tokens[TinyLlam
283283
disaggregated/test_disaggregated.py::test_disaggregated_deepseek_v3_lite_fp8_tp1_single_gpu_mtp[DeepSeek-V3-Lite-fp8] SKIP (https://nvbugs/5465642)
284284
examples/test_multimodal.py::test_llm_multimodal_general[Mistral-Small-3.1-24B-Instruct-2503-pp:1-tp:1-bfloat16-bs:1-cpp_e2e:False-nb:1] SKIP (https://nvbugs/5431146)
285285
accuracy/test_disaggregated_serving.py::TestQwen3_30B_A3B::test_mixed_ctx_gen_model[ctxpp2gentp2] SKIP (https://nvbugs/5470769)
286-
full:L40S/accuracy/test_disaggregated_serving.py::TestDeepSeekV3Lite::test_auto_dtype[mtp_nextn=0-overlap_scheduler=False] SKIP (https://nvbugs/5347051)
287286
full:L40S/accuracy/test_disaggregated_serving.py::TestLlama3_1_8BInstruct::test_tp_pp_symmetric[MMLU-tp2pp2] SKIP (https://nvbugs/5471108)
288-
full:L20/accuracy/test_disaggregated_serving.py::TestDeepSeekV3Lite::test_auto_dtype[mtp_nextn=0-overlap_scheduler=False] SKIP (https://nvbugs/5347051)
289287
full:L20/accuracy/test_disaggregated_serving.py::TestLlama3_1_8BInstruct::test_tp_pp_symmetric[MMLU-tp2pp2] SKIP (https://nvbugs/5471108)
290288
test_e2e.py::test_multi_nodes_eval[llama4-models/nvidia/Llama-4-Maverick-17B-128E-Instruct-FP8-tp8pp2-mmlu] SKIP (https://nvbugs/5473781)
291289
disaggregated/test_disaggregated_single_gpu.py::test_disaggregated_llama_context_capacity[False-False-DeepSeek-V3-Lite-fp8/fp8] SKIP (https://nvbugs/5477404)
@@ -337,9 +335,6 @@ full:H100/accuracy/test_llm_api_pytorch.py::TestLlama4MaverickInstruct::test_fp8
337335
full:H100/accuracy/test_llm_api_pytorch.py::TestLlama4MaverickInstruct::test_fp8[tp8ep4-cuda_graph=True] SKIP (https://nvbugs/5512734)
338336
full:H100/accuracy/test_llm_api_pytorch.py::TestLlama4MaverickInstruct::test_fp8[tp8-cuda_graph=True] SKIP (https://nvbugs/5512734)
339337
full:H100/accuracy/test_llm_api_pytorch.py::TestLlama4MaverickInstruct::test_fp8_eagle3[tp8-torch_compile=True] SKIP (https://nvbugs/5483534)
340-
full:A100/accuracy/test_disaggregated_serving.py::TestDeepSeekV3Lite::test_auto_dtype[mtp_nextn=0-overlap_scheduler=False] SKIP (https://nvbugs/5519462)
341-
full:A100/accuracy/test_disaggregated_serving.py::TestDeepSeekV3Lite::test_auto_dtype[mtp_nextn=0-overlap_scheduler=True] SKIP (https://nvbugs/5519462)
342-
full:A100/accuracy/test_llm_api_pytorch.py::TestDeepSeekV3Lite::test_bfloat16[mtp_nextn=0-attention_dp=False-cuda_graph=False-overlap_scheduler=False-torch_compile=False-enable_chunked_prefill=False] SKIP (https://nvbugs/5347051)
343338
full:A100/test_e2e.py::test_ptp_quickstart_multimodal[NVILA-8B-FP16-vila/NVILA-8B-video-False] SKIP (https://nvbugs/5453725)
344339
test_e2e.py::test_ptp_scaffolding[DeepSeek-R1-Distill-Qwen-7B-DeepSeek-R1/DeepSeek-R1-Distill-Qwen-7B] SKIP (https://nvbugs/5517260)
345340
accuracy/test_llm_api_pytorch.py::TestLlama4ScoutInstruct::test_fp8_chunked_prefill[tp4ep4-cuda_graph=True] SKIP (https://nvbugs/5522462)

0 commit comments

Comments
 (0)