11# SPDX-License-Identifier: Apache-2.0 
22# SPDX-FileCopyrightText: Copyright contributors to the vLLM project 
3- import  os 
43from  typing  import  Optional 
54
65import  pytest 
@@ -29,20 +28,24 @@ def v1(run_with_both_engines):
2928        # [Decoder-only]  
3029        pytest .param ("BAAI/bge-multilingual-gemma2" , 
3130                     marks = [pytest .mark .core_model ]), 
32-         pytest .param ("intfloat/e5-mistral-7b-instruct" , 
33-                      marks = [pytest .mark .core_model , pytest .mark .cpu_model ]), 
31+         pytest .param ( 
32+             "intfloat/e5-mistral-7b-instruct" , 
33+             # CPU v1 doesn't support sliding window  
34+             marks = [pytest .mark .core_model ]), 
3435        # the qwen models interfere with each other (see PR  
3536        # https://github.com/vllm-project/vllm/pull/18720).  
3637        # To avoid this problem, for now we skip v0 since it will be  
3738        # deprecated anyway.  
3839        pytest .param ("ssmits/Qwen2-7B-Instruct-embed-base" , 
3940                     marks = [pytest .mark .skip_v0 , pytest .mark .cpu_model ]), 
4041        # [Encoder-only]  
41-         pytest .param ("BAAI/bge-base-en-v1.5" , 
42-                      marks = [ 
43-                          pytest .mark .core_model , pytest .mark .cpu_model , 
44-                          pytest .mark .skip_v1  
45-                      ]), 
42+         pytest .param ( 
43+             "BAAI/bge-base-en-v1.5" , 
44+             marks = [ 
45+                 # CPU only supports V1  
46+                 pytest .mark .core_model , 
47+                 pytest .mark .skip_v1  
48+             ]), 
4649        pytest .param ("sentence-transformers/all-MiniLM-L12-v2" , 
4750                     marks = [pytest .mark .skip_v1 ]), 
4851        pytest .param ("intfloat/multilingual-e5-small" , 
@@ -61,10 +64,6 @@ def test_models(
6164    model ,
6265    monkeypatch ,
6366) ->  None :
64-     if  model  ==  "intfloat/e5-mistral-7b-instruct"  and  current_platform .is_cpu (
65-     ) and  os .environ .get ("VLLM_USE_V1" , "0" ) ==  "1" :
66-         pytest .skip ("CPU V1 doesn't support sliding window" )
67- 
6867    if  model  ==  "BAAI/bge-multilingual-gemma2"  and  current_platform .is_rocm ():
6968        # ROCm Triton FA does not currently support sliding window attention 
7069        # switch to use ROCm CK FA backend 
0 commit comments