Skip to content

Commit 266af2d

Browse files
committed
Skip compressed-tensors quant test on ROCm (erroneously introduced in #6549)
1 parent f6721e8 commit 266af2d

File tree

1 file changed

+7
-2
lines changed

1 file changed

+7
-2
lines changed
Lines changed: 7 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -1,8 +1,13 @@
1+
from vllm.utils import is_hip
2+
13
from ..utils import compare_two_settings
24

35

46
def test_cpu_offload():
57
compare_two_settings("meta-llama/Llama-2-7b-hf", [],
68
["--cpu-offload-gb", "4"])
7-
compare_two_settings("nm-testing/llama7b-one-shot-2_4-w4a16-marlin24-t",
8-
[], ["--cpu-offload-gb", "1"])
9+
if not is_hip():
10+
# compressed-tensors quantization is currently not supported in ROCm.
11+
compare_two_settings(
12+
"nm-testing/llama7b-one-shot-2_4-w4a16-marlin24-t", [],
13+
["--cpu-offload-gb", "1"])

0 commit comments

Comments
 (0)