File tree Expand file tree Collapse file tree 3 files changed +1
-21
lines changed Expand file tree Collapse file tree 3 files changed +1
-21
lines changed Original file line number Diff line number Diff line change 1212import  torch_xla .core 
1313import  torch_xla .core .xla_model 
1414
15- from  vllm  import  envs 
1615from  vllm .attention .layer  import  MultiHeadAttention 
1716from  vllm .attention .selector  import  _cached_get_attn_backend 
1817from  vllm .platforms  import  current_platform 
1918
20- if  not  envs .VLLM_USE_V1 :
21-     pytest .skip (
22-         "Skipping V1 tests. Rerun with `VLLM_USE_V1=1` to test." ,
23-         allow_module_level = True ,
24-     )
25- 
2619
2720@pytest .fixture (autouse = True ) 
2821def  clear_cache ():
Original file line number Diff line number Diff line change 44import  openai 
55import  pytest 
66
7- from  vllm  import  envs 
87from  vllm .multimodal .utils  import  encode_image_base64 , fetch_image 
98from  vllm .platforms  import  current_platform 
109
1110from  ...entrypoints .openai .test_vision  import  TEST_IMAGE_URLS 
1211from  ...utils  import  RemoteOpenAIServer 
1312
14- if  not  envs .VLLM_USE_V1 :
15-     pytest .skip (
16-         "Skipping V1 tests. Rerun with `VLLM_USE_V1=1` to test." ,
17-         allow_module_level = True ,
18-     )
19- 
2013
2114@pytest .fixture (scope = "session" ) 
2215def  base64_encoded_image () ->  dict [str , str ]:
Original file line number Diff line number Diff line change 44
55import  pytest 
66
7- from  vllm  import  LLM ,  envs 
7+ from  vllm  import  LLM 
88from  vllm .platforms  import  current_platform 
99from  vllm .sampling_params  import  SamplingParams 
1010
11- if  not  envs .VLLM_USE_V1 :
12-     pytest .skip (
13-         "Skipping V1 tests. Rerun with `VLLM_USE_V1=1` to test." ,
14-         allow_module_level = True ,
15-     )
16- 
1711
1812@pytest .mark .parametrize ("model_name" , ["Qwen/Qwen2.5-1.5B-Instruct" ]) 
1913@pytest .mark .skipif (not  current_platform .is_tpu (), 
 
 
   
 
     
   
   
          
    
    
     
    
      
     
     
    You can’t perform that action at this time.
  
 
    
  
    
      
        
     
       
      
     
   
 
    
    
  
 
  
 
     
    
0 commit comments