Skip to content

Commit aa6d40d

Browse files
committed
fix: get test_mllama.py passing
Signed-off-by: Travis Johnson <[email protected]>
1 parent 62b7772 commit aa6d40d

File tree

1 file changed

+9
-8
lines changed

1 file changed

+9
-8
lines changed

tests/models/encoder_decoder/vision_language/test_mllama.py

Lines changed: 9 additions & 8 deletions
Original file line numberDiff line numberDiff line change
@@ -209,14 +209,15 @@ def _run_test(
209209
# will hurt multiprocessing backend with fork method (the default method).
210210

211211
# max_model_len should be greater than image_feature_size
212-
with vllm_runner(model,
213-
dtype=dtype,
214-
max_model_len=8192,
215-
max_num_seqs=3,
216-
tensor_parallel_size=tensor_parallel_size,
217-
distributed_executor_backend=distributed_executor_backend,
218-
limit_mm_per_prompt={"image": _LIMIT_IMAGE_PER_PROMPT
219-
}) as vllm_model:
212+
with vllm_runner(
213+
model,
214+
dtype=dtype,
215+
max_model_len=19212, # 3 max size images
216+
max_num_seqs=3,
217+
tensor_parallel_size=tensor_parallel_size,
218+
distributed_executor_backend=distributed_executor_backend,
219+
limit_mm_per_prompt={"image":
220+
_LIMIT_IMAGE_PER_PROMPT}) as vllm_model:
220221
vllm_outputs_per_image = [
221222
vllm_model.generate_greedy_logprobs(prompts,
222223
max_tokens,

0 commit comments

Comments
 (0)