File tree Expand file tree Collapse file tree 3 files changed +1
-21
lines changed Expand file tree Collapse file tree 3 files changed +1
-21
lines changed Original file line number Diff line number Diff line change 1212import torch_xla .core
1313import torch_xla .core .xla_model
1414
15- from vllm import envs
1615from vllm .attention .layer import MultiHeadAttention
1716from vllm .attention .selector import _cached_get_attn_backend
1817from vllm .platforms import current_platform
1918
20- if not envs .VLLM_USE_V1 :
21- pytest .skip (
22- "Skipping V1 tests. Rerun with `VLLM_USE_V1=1` to test." ,
23- allow_module_level = True ,
24- )
25-
2619
2720@pytest .fixture (autouse = True )
2821def clear_cache ():
Original file line number Diff line number Diff line change 44import openai
55import pytest
66
7- from vllm import envs
87from vllm .multimodal .utils import encode_image_base64 , fetch_image
98from vllm .platforms import current_platform
109
1110from ...entrypoints .openai .test_vision import TEST_IMAGE_URLS
1211from ...utils import RemoteOpenAIServer
1312
14- if not envs .VLLM_USE_V1 :
15- pytest .skip (
16- "Skipping V1 tests. Rerun with `VLLM_USE_V1=1` to test." ,
17- allow_module_level = True ,
18- )
19-
2013
2114@pytest .fixture (scope = "session" )
2215def base64_encoded_image () -> dict [str , str ]:
Original file line number Diff line number Diff line change 44
55import pytest
66
7- from vllm import LLM , envs
7+ from vllm import LLM
88from vllm .platforms import current_platform
99from vllm .sampling_params import SamplingParams
1010
11- if not envs .VLLM_USE_V1 :
12- pytest .skip (
13- "Skipping V1 tests. Rerun with `VLLM_USE_V1=1` to test." ,
14- allow_module_level = True ,
15- )
16-
1711
1812@pytest .mark .parametrize ("model_name" , ["Qwen/Qwen2.5-1.5B-Instruct" ])
1913@pytest .mark .skipif (not current_platform .is_tpu (),
You can’t perform that action at this time.
0 commit comments