Skip to content

Commit 2692daa

Browse files
authored
infra: Remove the WAR for test items incompletely (#3313)
* Remove the WAR for test items incompleted Signed-off-by: EmmaQiaoCh <[email protected]> * Complete test item manually Signed-off-by: EmmaQiaoCh <[email protected]> * Fix another test definition file Signed-off-by: EmmaQiaoCh <[email protected]> * Complete test name Signed-off-by: EmmaQiaoCh <[email protected]> * Fix some other test names Signed-off-by: EmmaQiaoCh <[email protected]> * Fix another test name after rebase Signed-off-by: EmmaQiaoCh <[email protected]> * Update name for waived case name, too Signed-off-by: EmmaQiaoCh <[email protected]> * Fix name for multi-gpu tests Signed-off-by: EmmaQiaoCh <[email protected]> * Fix test name after rebase Signed-off-by: EmmaQiaoCh <[email protected]> * Fix another test name Signed-off-by: EmmaQiaoCh <[email protected]> * Fix typo Signed-off-by: EmmaQiaoCh <[email protected]> * Fix test name after rebase Signed-off-by: EmmaQiaoCh <[email protected]> * Fix other qa tests Signed-off-by: EmmaQiaoCh <[email protected]> * Fix tests name after rebase Signed-off-by: qqiao <[email protected]> * Fix name after rebase Signed-off-by: qqiao <[email protected]> * Correct test names in waive.txt Signed-off-by: qqiao <[email protected]> * Add new test_durations file Signed-off-by: qqiao <[email protected]> * Fix names after rebase Signed-off-by: qqiao <[email protected]> * Update test duration to latest Signed-off-by: qqiao <[email protected]> --------- Signed-off-by: EmmaQiaoCh <[email protected]> Signed-off-by: qqiao <[email protected]>
1 parent 403370a commit 2692daa

File tree

11 files changed

+524
-552
lines changed

11 files changed

+524
-552
lines changed

tests/integration/defs/.test_durations

+386-398
Large diffs are not rendered by default.

tests/integration/defs/conftest.py

-16
Original file line numberDiff line numberDiff line change
@@ -1971,22 +1971,6 @@ def get_device_memory():
19711971
return memory
19721972

19731973

1974-
#
1975-
# When test parameters have an empty id, older versions of pytest ignored that parameter when generating the
1976-
# test node's ID completely. This however was actually a bug, and not expected behavior that got fixed in newer
1977-
# versions of pytest:https://github.com/pytest-dev/pytest/pull/6607. TRT test defs however rely on this behavior
1978-
# for quite a few test names. This is a hacky WAR that restores the old behavior back so that the
1979-
# test names do not change. Note: This might break in a future pytest version.
1980-
#
1981-
# TODO: Remove this hack once the test names are fixed.
1982-
#
1983-
1984-
from _pytest.python import CallSpec2
1985-
1986-
CallSpec2.id = property(
1987-
lambda self: "-".join(map(str, filter(None, self._idlist))))
1988-
1989-
19901974
def pytest_addoption(parser):
19911975
parser.addoption(
19921976
"--test-list",

tests/integration/test_lists/qa/examples_test_list.txt

+14-14
Original file line numberDiff line numberDiff line change
@@ -424,18 +424,18 @@ accuracy/test_llm_api_pytorch.py::TestLlama3_3_70BInstruct::test_nvfp4_tp4
424424
accuracy/test_llm_api_pytorch.py::TestMistral7B::test_auto_dtype
425425
accuracy/test_llm_api_pytorch.py::TestMixtral8x7B::test_fp8_tp2
426426
accuracy/test_llm_api_pytorch.py::TestMixtral8x7B::test_nvfp4_tp2
427-
accuracy/test_llm_api_pytorch.py::TestDeepSeekV3Lite::test_bfloat16[]
428-
accuracy/test_llm_api_pytorch.py::TestDeepSeekV3Lite::test_fp8_block_scales[]
427+
accuracy/test_llm_api_pytorch.py::TestDeepSeekV3Lite::test_bfloat16[-]
428+
accuracy/test_llm_api_pytorch.py::TestDeepSeekV3Lite::test_fp8_block_scales[-]
429429
accuracy/test_llm_api_pytorch.py::TestDeepSeekV3Lite::test_nvfp4[]
430430
accuracy/test_llm_api_pytorch.py::TestMinitron4BBaseInstruct::test_fp8_prequantized
431431
accuracy/test_llm_api_pytorch.py::TestNemotronNas::test_auto_dtype_tp8
432432
accuracy/test_llm_api_pytorch.py::TestQwen2_7BInstruct::test_auto_dtype
433-
accuracy/test_llm_api_pytorch.py::TestDeepSeekR1::test_nvfp4_8gpus[tp8-cuda_graph-overlap_scheduler]
434-
accuracy/test_llm_api_pytorch.py::TestDeepSeekR1::test_nvfp4_8gpus[tp8-mtp_nextn=2-cuda_graph-overlap_scheduler]
435-
accuracy/test_llm_api_pytorch.py::TestDeepSeekR1::test_nvfp4_8gpus[tp8ep4-cuda_graph-overlap_scheduler]
436-
accuracy/test_llm_api_pytorch.py::TestDeepSeekR1::test_nvfp4_8gpus[tp8ep4-mtp_nextn=2-cuda_graph-overlap_scheduler]
437-
accuracy/test_llm_api_pytorch.py::TestDeepSeekR1::test_nvfp4_8gpus[tp8ep8-cuda_graph-overlap_scheduler]
438-
accuracy/test_llm_api_pytorch.py::TestDeepSeekR1::test_nvfp4_8gpus[tp8ep8-mtp_nextn=2-cuda_graph-overlap_scheduler]
433+
accuracy/test_llm_api_pytorch.py::TestDeepSeekR1::test_nvfp4_8gpus[tp8---cuda_graph-overlap_scheduler]
434+
accuracy/test_llm_api_pytorch.py::TestDeepSeekR1::test_nvfp4_8gpus[tp8-mtp_nextn=2--cuda_graph-overlap_scheduler]
435+
accuracy/test_llm_api_pytorch.py::TestDeepSeekR1::test_nvfp4_8gpus[tp8ep4---cuda_graph-overlap_scheduler]
436+
accuracy/test_llm_api_pytorch.py::TestDeepSeekR1::test_nvfp4_8gpus[tp8ep4-mtp_nextn=2--cuda_graph-overlap_scheduler]
437+
accuracy/test_llm_api_pytorch.py::TestDeepSeekR1::test_nvfp4_8gpus[tp8ep8---cuda_graph-overlap_scheduler]
438+
accuracy/test_llm_api_pytorch.py::TestDeepSeekR1::test_nvfp4_8gpus[tp8ep8-mtp_nextn=2--cuda_graph-overlap_scheduler]
439439

440440
test_e2e.py::test_benchmark_sanity[bert_base] # 127.18s
441441
test_e2e.py::test_benchmark_sanity[gpt_350m] # 64.06s
@@ -445,14 +445,14 @@ test_e2e.py::test_benchmark_sanity[roberta_base]
445445
test_e2e.py::test_benchmark_sanity[t5_base]
446446
test_e2e.py::test_benchmark_sanity_enable_fp8[gpt_350m]
447447
test_e2e.py::test_benchmark_sanity_enable_fp8[llama_7b]
448-
test_e2e.py::test_llama_e2e[use_cpp_session-remove_input_padding]
449-
test_e2e.py::test_llama_e2e[use_py_session-remove_input_padding]
450-
test_e2e.py::test_llama_e2e[use_py_session]
448+
test_e2e.py::test_llama_e2e[use_cpp_session-remove_input_padding-]
449+
test_e2e.py::test_llama_e2e[use_py_session-remove_input_padding-]
450+
test_e2e.py::test_llama_e2e[use_py_session--]
451451
test_e2e.py::test_llmapi_load_engine_from_build_command[llama-codellama/CodeLlama-7b-Instruct-hf] # 5min
452452
test_e2e.py::test_llmapi_load_engine_from_build_command[llama-llama-models/llama-7b-hf] # 5min
453-
test_e2e.py::test_mistral_e2e[use_cpp_session-remove_input_padding]
454-
test_e2e.py::test_mistral_e2e[use_py_session-remove_input_padding]
455-
test_e2e.py::test_mistral_e2e[use_py_session]
453+
test_e2e.py::test_mistral_e2e[use_cpp_session-remove_input_padding--]
454+
test_e2e.py::test_mistral_e2e[use_py_session-remove_input_padding--]
455+
test_e2e.py::test_mistral_e2e[use_py_session---]
456456
test_e2e.py::test_openai_multi_chat_example
457457
test_e2e.py::test_openai_consistent_chat
458458
test_e2e.py::test_llmapi_server_example

tests/integration/test_lists/qa/llm_sanity_test.txt

+8-8
Original file line numberDiff line numberDiff line change
@@ -91,14 +91,14 @@ examples/test_recurrentgemma.py::test_llm_recurrentgemma_1gpu[use_cpp_session-re
9191
examples/test_redrafter.py::test_llm_redrafter_1gpu[use_cpp_session-redrafter-vicuna-7b-v1.3-bfloat16-dl5-nb8-bs8]
9292
examples/test_redrafter.py::test_llm_redrafter_1gpu[use_py_session-redrafter-vicuna-7b-v1.3-bfloat16-dl5-nb5-bs8]
9393
examples/test_whisper.py::test_llm_whisper_general[large-v3-disable_gemm_plugin-enable_attention_plugin-disable_weight_only-float16-nb:1-use_cpp_runtime]
94-
test_e2e.py::test_llama_e2e[use_cpp_session-remove_input_padding]
95-
test_e2e.py::test_llama_e2e[use_py_session-remove_input_padding]
96-
test_e2e.py::test_llama_e2e[use_py_session]
94+
test_e2e.py::test_llama_e2e[use_cpp_session-remove_input_padding-]
95+
test_e2e.py::test_llama_e2e[use_py_session-remove_input_padding-]
96+
test_e2e.py::test_llama_e2e[use_py_session--]
9797
test_e2e.py::test_llmapi_load_engine_from_build_command[llama-codellama/CodeLlama-7b-Instruct-hf] # 5min
9898
test_e2e.py::test_llmapi_load_engine_from_build_command[llama-llama-models/llama-7b-hf] # 5min
99-
test_e2e.py::test_mistral_e2e[use_cpp_session-remove_input_padding]
100-
test_e2e.py::test_mistral_e2e[use_py_session-remove_input_padding]
101-
test_e2e.py::test_mistral_e2e[use_py_session]
99+
test_e2e.py::test_mistral_e2e[use_cpp_session-remove_input_padding--]
100+
test_e2e.py::test_mistral_e2e[use_py_session-remove_input_padding--]
101+
test_e2e.py::test_mistral_e2e[use_py_session---]
102102
test_e2e.py::test_openai_multi_chat_example
103103
test_e2e.py::test_openai_consistent_chat
104104

@@ -130,8 +130,8 @@ accuracy/test_llm_api_pytorch.py::TestLlama3_3_70BInstruct::test_fp8_tp4
130130
accuracy/test_llm_api_pytorch.py::TestLlama3_3_70BInstruct::test_nvfp4_tp4
131131
accuracy/test_llm_api_pytorch.py::TestMixtral8x7B::test_fp8_tp2
132132
accuracy/test_llm_api_pytorch.py::TestMixtral8x7B::test_nvfp4_tp2
133-
accuracy/test_llm_api_pytorch.py::TestDeepSeekV3Lite::test_bfloat16[]
134-
accuracy/test_llm_api_pytorch.py::TestDeepSeekV3Lite::test_fp8_block_scales[]
133+
accuracy/test_llm_api_pytorch.py::TestDeepSeekV3Lite::test_bfloat16[-]
134+
accuracy/test_llm_api_pytorch.py::TestDeepSeekV3Lite::test_fp8_block_scales[-]
135135
accuracy/test_llm_api_pytorch.py::TestDeepSeekV3Lite::test_nvfp4[]
136136

137137
# Pivot to Pytorch test cases.

tests/integration/test_lists/test-db/l0_a10.yml

+10-10
Original file line numberDiff line numberDiff line change
@@ -34,11 +34,11 @@ l0_a10:
3434
backend: cpp
3535
tests:
3636
# ------------- CPP tests ---------------
37-
- test_cpp.py::test_model[medusa-86]
38-
- test_cpp.py::test_model[redrafter-86]
39-
- test_cpp.py::test_model[mamba-86]
40-
- test_cpp.py::test_model[recurrentgemma-86]
41-
- test_cpp.py::test_model[eagle-86]
37+
- test_cpp.py::test_model[-medusa-86]
38+
- test_cpp.py::test_model[-redrafter-86]
39+
- test_cpp.py::test_model[-mamba-86]
40+
- test_cpp.py::test_model[-recurrentgemma-86]
41+
- test_cpp.py::test_model[-eagle-86]
4242
- condition:
4343
ranges:
4444
system_gpu_count:
@@ -75,9 +75,9 @@ l0_a10:
7575
- test_e2e.py::test_openai_completions_example
7676
- test_e2e.py::test_openai_chat_example
7777
- test_e2e.py::test_openai_chat_multimodal_example
78-
- test_e2e.py::test_trtllm_bench_sanity[non-streaming-FP16-meta-llama/Llama-3.1-8B-llama-3.1-model/Meta-Llama-3.1-8B]
78+
- test_e2e.py::test_trtllm_bench_sanity[--non-streaming-FP16-meta-llama/Llama-3.1-8B-llama-3.1-model/Meta-Llama-3.1-8B]
7979
- test_e2e.py::test_trtllm_bench_latency_sanity[FP16-meta-llama/Llama-3.1-8B-llama-3.1-model/Meta-Llama-3.1-8B]
80-
- test_e2e.py::test_trtllm_bench_request_rate_and_concurrency[enable_concurrency]
80+
- test_e2e.py::test_trtllm_bench_request_rate_and_concurrency[enable_concurrency-]
8181
- unittest/trt/quantization # 18 mins
8282
- accuracy/test_cli_flow.py::TestLlama7B::test_streamingllm # 2 mins
8383
- unittest/trt/functional # 37 mins
@@ -127,9 +127,9 @@ l0_a10:
127127
stage: post_merge
128128
backend: tensorrt
129129
tests:
130-
- test_e2e.py::test_mistral_e2e[use_py_session]
131-
- test_e2e.py::test_mistral_e2e[use_cpp_session-remove_input_padding]
132-
- test_e2e.py::test_mistral_e2e[use_py_session-remove_input_padding]
130+
- test_e2e.py::test_mistral_e2e[use_py_session---]
131+
- test_e2e.py::test_mistral_e2e[use_cpp_session-remove_input_padding--]
132+
- test_e2e.py::test_mistral_e2e[use_py_session-remove_input_padding--]
133133
- examples/test_bert.py::test_llm_bert_general[compare_hf-disable_remove_input_padding-disable_attention_plugin-disable_context_fmha-tp:1-pp:1-float32-BertModel-bert/bert-base-uncased]
134134
- examples/test_bert.py::test_llm_bert_general[compare_hf-enable_remove_input_padding-use_attention_plugin-enable_context_fmha-tp:1-pp:1-float16-RobertaModel-bert/roberta-base]
135135
- examples/test_bert.py::test_llm_bert_general[compare_hf-enable_remove_input_padding-disable_attention_plugin-disable_context_fmha-tp:1-pp:1-float16-RobertaForSequenceClassification-bert/twitter-roberta-base-emotion]

tests/integration/test_lists/test-db/l0_a30.yml

+4-4
Original file line numberDiff line numberDiff line change
@@ -47,9 +47,9 @@ l0_a30:
4747
- test_cpp.py::test_unit_tests[runtime-80]
4848
- test_cpp.py::test_unit_tests[thop-80]
4949
- test_cpp.py::test_unit_tests[utils-80]
50-
- test_cpp.py::test_model[gpt-80]
51-
- test_cpp.py::test_model[gpt_executor-80]
52-
- test_cpp.py::test_model[gpt_tests-80]
50+
- test_cpp.py::test_model[-gpt-80]
51+
- test_cpp.py::test_model[-gpt_executor-80]
52+
- test_cpp.py::test_model[-gpt_tests-80]
5353
- condition:
5454
ranges:
5555
system_gpu_count:
@@ -163,5 +163,5 @@ l0_a30:
163163
stage: post_merge
164164
backend: cpp
165165
tests:
166-
- test_cpp.py::test_model[gpt_session-80]
166+
- test_cpp.py::test_model[-gpt_session-80]
167167
- test_cpp.py::test_benchmarks[gpt-80]

tests/integration/test_lists/test-db/l0_b200.yml

+6-6
Original file line numberDiff line numberDiff line change
@@ -15,12 +15,12 @@ l0_b200:
1515
tests:
1616
# ------------- PyTorch tests ---------------
1717
- accuracy/test_llm_api_pytorch.py::TestLlama3_1_8B::test_nvfp4
18-
- accuracy/test_llm_api_pytorch.py::TestDeepSeekV3Lite::test_bfloat16[]
19-
- accuracy/test_llm_api_pytorch.py::TestDeepSeekV3Lite::test_bfloat16[attention_dp]
20-
- accuracy/test_llm_api_pytorch.py::TestDeepSeekV3Lite::test_bfloat16[cuda_graph]
21-
- accuracy/test_llm_api_pytorch.py::TestDeepSeekV3Lite::test_bfloat16[overlap_scheduler]
22-
- accuracy/test_llm_api_pytorch.py::TestDeepSeekV3Lite::test_bfloat16[attention_dp-cuda_graph-overlap_scheduler]
23-
- accuracy/test_llm_api_pytorch.py::TestDeepSeekV3Lite::test_bfloat16[mtp_nextn=2]
18+
- accuracy/test_llm_api_pytorch.py::TestDeepSeekV3Lite::test_bfloat16[-]
19+
- accuracy/test_llm_api_pytorch.py::TestDeepSeekV3Lite::test_bfloat16[-attention_dp]
20+
- accuracy/test_llm_api_pytorch.py::TestDeepSeekV3Lite::test_bfloat16[-cuda_graph]
21+
- accuracy/test_llm_api_pytorch.py::TestDeepSeekV3Lite::test_bfloat16[-overlap_scheduler]
22+
- accuracy/test_llm_api_pytorch.py::TestDeepSeekV3Lite::test_bfloat16[-attention_dp-cuda_graph-overlap_scheduler]
23+
- accuracy/test_llm_api_pytorch.py::TestDeepSeekV3Lite::test_bfloat16[mtp_nextn=2-]
2424
- accuracy/test_llm_api_pytorch.py::TestDeepSeekV3Lite::test_bfloat16[mtp_nextn=2-attention_dp]
2525
- accuracy/test_llm_api_pytorch.py::TestDeepSeekV3Lite::test_bfloat16[mtp_nextn=2-cuda_graph]
2626
- accuracy/test_llm_api_pytorch.py::TestDeepSeekV3Lite::test_bfloat16[mtp_nextn=2-overlap_scheduler]

0 commit comments

Comments
 (0)