Skip to content

Commit eba5e57

Browse files
yangw-devmawong-amd
authored andcommitted
add more pytorch related tests for torch nightly (vllm-project#17422)
Signed-off-by: Yang Wang <[email protected]>
1 parent f6f7c5b commit eba5e57

File tree

3 files changed

+14
-4
lines changed

3 files changed

+14
-4
lines changed

.buildkite/test-pipeline.yaml

Lines changed: 4 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -293,6 +293,7 @@ steps:
293293
parallelism: 4
294294

295295
- label: PyTorch Compilation Unit Tests
296+
torch_nightly: true
296297
source_file_dependencies:
297298
- vllm/
298299
- tests/compile
@@ -302,6 +303,7 @@ steps:
302303
- pytest -v -s compile/test_sequence_parallelism.py
303304

304305
- label: PyTorch Fullgraph Smoke Test # 9min
306+
torch_nightly: true
305307
source_file_dependencies:
306308
- vllm/
307309
- tests/compile
@@ -312,6 +314,7 @@ steps:
312314
- pytest -v -s compile/piecewise/test_toy_llama.py
313315

314316
- label: PyTorch Fullgraph Test # 18min
317+
torch_nightly: true
315318
source_file_dependencies:
316319
- vllm/
317320
- tests/compile
@@ -436,6 +439,7 @@ steps:
436439
##### models test #####
437440

438441
- label: Basic Models Test # 24min
442+
torch_nightly: true
439443
source_file_dependencies:
440444
- vllm/
441445
- tests/models

requirements/nightly_torch_test.txt

Lines changed: 7 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -23,5 +23,11 @@ runai-model-streamer-s3==0.11.0
2323
tensorizer>=2.9.0
2424
lm-eval==0.4.8
2525
buildkite-test-collector==0.1.9
26-
2726
lm-eval[api]==0.4.8 # required for model evaluation test
27+
28+
# required for quantization test
29+
bitsandbytes>=0.45.3
30+
31+
# required for minicpmo_26 test
32+
vector_quantize_pytorch
33+
vocos

vllm/sampling_params.py

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -186,9 +186,9 @@ class SamplingParams(
186186
logits_processors: list of functions that modify logits based on
187187
previously generated tokens, and optionally prompt tokens as
188188
a first argument.
189-
truncate_prompt_tokens: If set to -1, will use the truncation size
190-
supported by the model. If set to an integer k, will use only
191-
the last k tokens from the prompt (i.e., left truncation).
189+
truncate_prompt_tokens: If set to -1, will use the truncation size
190+
supported by the model. If set to an integer k, will use only
191+
the last k tokens from the prompt (i.e., left truncation).
192192
Defaults to None (i.e., no truncation).
193193
guided_decoding: If provided, the engine will construct a guided
194194
decoding logits processor from these parameters. Defaults to None.

0 commit comments

Comments
 (0)