Skip to content

Commit 5152935

Browse files
DarkLight1337charlifu
authored andcommitted
[Misc] Avoid use of deprecated AutoModelForVision2Seq (vllm-project#25065)
Signed-off-by: DarkLight1337 <[email protected]> Signed-off-by: charlifu <[email protected]>
1 parent 5938e5f commit 5152935

File tree

1 file changed

+7
-7
lines changed

1 file changed

+7
-7
lines changed

tests/models/multimodal/generation/test_common.py

Lines changed: 7 additions & 7 deletions
Original file line numberDiff line numberDiff line change
@@ -10,7 +10,7 @@
1010

1111
import pytest
1212
from transformers import (AutoModel, AutoModelForImageTextToText,
13-
AutoModelForTextToWaveform, AutoModelForVision2Seq)
13+
AutoModelForTextToWaveform)
1414

1515
from vllm.platforms import current_platform
1616
from vllm.utils import identity
@@ -137,7 +137,7 @@
137137
video_idx_to_prompt=lambda idx: "<|vision_start|><|video_pad|><|vision_end|>", # noqa: E501
138138
max_model_len=4096,
139139
max_num_seqs=2,
140-
auto_cls=AutoModelForVision2Seq,
140+
auto_cls=AutoModelForImageTextToText,
141141
vllm_output_post_proc=model_utils.qwen2_vllm_to_hf_output,
142142
image_size_factors=[(), (0.25,), (0.25, 0.25, 0.25), (0.25, 0.2, 0.15)],
143143
marks=[pytest.mark.core_model, pytest.mark.cpu_model],
@@ -502,7 +502,7 @@
502502
num_video_frames=16,
503503
max_model_len=16384,
504504
hf_model_kwargs=model_utils.llava_onevision_hf_model_kwargs("llava-hf/llava-onevision-qwen2-0.5b-ov-hf"), # noqa: E501
505-
auto_cls=AutoModelForVision2Seq,
505+
auto_cls=AutoModelForImageTextToText,
506506
vllm_output_post_proc=model_utils.llava_onevision_vllm_to_hf_output,
507507
custom_test_opts=[CustomTestOptions(
508508
inputs=custom_inputs.multi_video_multi_aspect_ratio_inputs(
@@ -518,7 +518,7 @@
518518
num_video_frames=16,
519519
max_model_len=4096,
520520
max_num_seqs=2,
521-
auto_cls=AutoModelForVision2Seq,
521+
auto_cls=AutoModelForImageTextToText,
522522
vllm_output_post_proc=model_utils.llava_video_vllm_to_hf_output,
523523
),
524524
"mantis": VLMTestInfo(
@@ -680,7 +680,7 @@
680680
multi_image_prompt="Picture 1: <vlm_image>\nPicture 2: <vlm_image>\nDescribe these two images with one paragraph respectively.", # noqa: E501
681681
max_model_len=4096,
682682
max_num_seqs=2,
683-
auto_cls=AutoModelForVision2Seq,
683+
auto_cls=AutoModelForImageTextToText,
684684
vllm_output_post_proc=model_utils.qwen2_vllm_to_hf_output,
685685
image_size_factors=[(), (0.25,), (0.25, 0.25, 0.25), (0.25, 0.2, 0.15)],
686686
marks=[pytest.mark.cpu_model],
@@ -784,7 +784,7 @@
784784
test_type=VLMTestType.CUSTOM_INPUTS,
785785
max_model_len=16384,
786786
max_num_seqs=2,
787-
auto_cls=AutoModelForVision2Seq,
787+
auto_cls=AutoModelForImageTextToText,
788788
hf_model_kwargs=model_utils.llava_onevision_hf_model_kwargs("llava-hf/llava-onevision-qwen2-0.5b-ov-hf"), # noqa: E501
789789
vllm_output_post_proc=model_utils.llava_onevision_vllm_to_hf_output,
790790
custom_test_opts=[CustomTestOptions(
@@ -800,7 +800,7 @@
800800
test_type=VLMTestType.CUSTOM_INPUTS,
801801
max_model_len=4096,
802802
max_num_seqs=2,
803-
auto_cls=AutoModelForVision2Seq,
803+
auto_cls=AutoModelForImageTextToText,
804804
vllm_output_post_proc=model_utils.qwen2_vllm_to_hf_output,
805805
custom_test_opts=[CustomTestOptions(
806806
inputs=custom_inputs.windows_attention_image_qwen2_5_vl(),

0 commit comments

Comments
 (0)