11# SPDX-License-Identifier: Apache-2.0
22# SPDX-FileCopyrightText: Copyright contributors to the vLLM project
3- import os
43
54import pytest
65
@@ -28,20 +27,24 @@ def v1(run_with_both_engines):
2827 # [Decoder-only]
2928 pytest .param ("BAAI/bge-multilingual-gemma2" ,
3029 marks = [pytest .mark .core_model ]),
31- pytest .param ("intfloat/e5-mistral-7b-instruct" ,
32- marks = [pytest .mark .core_model , pytest .mark .cpu_model ]),
30+ pytest .param (
31+ "intfloat/e5-mistral-7b-instruct" ,
32+ # CPU v1 doesn't support sliding window
33+ marks = [pytest .mark .core_model ]),
3334 # the qwen models interfere with each other (see PR
3435 # https://github.com/vllm-project/vllm/pull/18720).
3536 # To avoid this problem, for now we skip v0 since it will be
3637 # deprecated anyway.
3738 pytest .param ("ssmits/Qwen2-7B-Instruct-embed-base" ,
3839 marks = [pytest .mark .skip_v0 , pytest .mark .cpu_model ]),
3940 # [Encoder-only]
40- pytest .param ("BAAI/bge-base-en-v1.5" ,
41- marks = [
42- pytest .mark .core_model , pytest .mark .cpu_model ,
43- pytest .mark .skip_v1
44- ]),
41+ pytest .param (
42+ "BAAI/bge-base-en-v1.5" ,
43+ marks = [
44+ # CPU only supports V1
45+ pytest .mark .core_model ,
46+ pytest .mark .skip_v1
47+ ]),
4548 pytest .param ("sentence-transformers/all-MiniLM-L12-v2" ,
4649 marks = [pytest .mark .skip_v1 ]),
4750 pytest .param ("intfloat/multilingual-e5-small" ,
@@ -60,10 +63,6 @@ def test_models(
6063 model ,
6164 monkeypatch ,
6265) -> None :
63- if model == "intfloat/e5-mistral-7b-instruct" and current_platform .is_cpu (
64- ) and os .environ .get ("VLLM_USE_V1" , "0" ) == "1" :
65- pytest .skip ("CPU V1 doesn't support sliding window" )
66-
6766 if model == "BAAI/bge-multilingual-gemma2" and current_platform .is_rocm ():
6867 # ROCm Triton FA does not currently support sliding window attention
6968 # switch to use ROCm CK FA backend
0 commit comments