Skip to content

Commit 09a498f

Browse files
committed
reduce test models to be one
Signed-off-by: xuebwang-amd <xuebwang@amd.com>
1 parent 469a6d6 commit 09a498f

File tree

1 file changed

+2
-10
lines changed

1 file changed

+2
-10
lines changed

tests/quantization/test_mixed_precision.py

Lines changed: 2 additions & 10 deletions
Original file line numberDiff line numberDiff line change
@@ -21,7 +21,7 @@
2121

2222
try:
2323
huggingface_hub.list_repo_refs(
24-
"amd/Llama-2-70b-chat-hf-WMXFP4FP8-AMXFP4FP8-AMP-KVFP8")
24+
"amd/Qwen3-8B-WMXFP4FP8-AMXFP4FP8-AMP-KVFP8")
2525
HF_HUB_AMD_ORG_ACCESS = True
2626
except huggingface_hub.errors.RepositoryNotFoundError:
2727
HF_HUB_AMD_ORG_ACCESS = False
@@ -45,14 +45,6 @@ def get_model_args(self) -> str:
4545

4646

4747
TEST_CONFIGS = {
48-
"amd/Llama-2-70b-chat-hf-WMXFP4FP8-AMXFP4FP8-AMP-KVFP8": {
49-
"arc_challenge": 0.53,
50-
"mmlu": 0.60
51-
},
52-
"amd/Mixtral-8x7B-Instruct-v0.1-WMXFP4FP8-AMXFP4FP8-AMP-KVFP8": {
53-
"arc_challenge": 0.62,
54-
"mmlu": 0.68
55-
},
5648
"amd/Qwen3-8B-WMXFP4FP8-AMXFP4FP8-AMP-KVFP8": {
5749
"arc_challenge": 0.52,
5850
"mmlu": 0.72
@@ -73,7 +65,7 @@ def test_mixed_precision_model_accuracies(model_name: str,
7365
model="vllm",
7466
model_args=EvaluationConfig(model_name).get_model_args(),
7567
tasks=list(accuracy_numbers.keys()),
76-
batch_size="auto" if "Qwen3-8B" not in model_name else 8)
68+
batch_size=8)
7769

7870
rtol = 0.05
7971

0 commit comments

Comments
 (0)