Skip to content

Commit 60cad94

Browse files
authored
[Hardware] correct method signatures for HPU,ROCm,XPU (#18551)
Signed-off-by: Andy Xie <[email protected]>
1 parent 9c1baa5 commit 60cad94

File tree

4 files changed

+16
-16
lines changed

4 files changed

+16
-16
lines changed

vllm/platforms/__init__.py

Lines changed: 4 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -42,7 +42,6 @@ def tpu_platform_plugin() -> Optional[str]:
4242
logger.debug("Confirmed TPU platform is available.")
4343
except Exception as e:
4444
logger.debug("TPU platform is not available because: %s", str(e))
45-
pass
4645

4746
return "vllm.platforms.tpu.TpuPlatform" if is_tpu else None
4847

@@ -112,7 +111,6 @@ def rocm_platform_plugin() -> Optional[str]:
112111
amdsmi.amdsmi_shut_down()
113112
except Exception as e:
114113
logger.debug("ROCm platform is not available because: %s", str(e))
115-
pass
116114

117115
return "vllm.platforms.rocm.RocmPlatform" if is_rocm else None
118116

@@ -130,7 +128,6 @@ def hpu_platform_plugin() -> Optional[str]:
130128
"habana_frameworks is not found.")
131129
except Exception as e:
132130
logger.debug("HPU platform is not available because: %s", str(e))
133-
pass
134131

135132
return "vllm.platforms.hpu.HpuPlatform" if is_hpu else None
136133

@@ -148,7 +145,6 @@ def xpu_platform_plugin() -> Optional[str]:
148145
logger.debug("Confirmed XPU platform is available.")
149146
except Exception as e:
150147
logger.debug("XPU platform is not available because: %s", str(e))
151-
pass
152148

153149
return "vllm.platforms.xpu.XPUPlatform" if is_xpu else None
154150

@@ -170,7 +166,6 @@ def cpu_platform_plugin() -> Optional[str]:
170166

171167
except Exception as e:
172168
logger.debug("CPU platform is not available because: %s", str(e))
173-
pass
174169

175170
return "vllm.platforms.cpu.CpuPlatform" if is_cpu else None
176171

@@ -222,8 +217,11 @@ def resolve_current_platform_cls_qualname() -> str:
222217
platform_cls_qualname = func()
223218
if platform_cls_qualname is not None:
224219
activated_plugins.append(name)
220+
logger.info("Platform plugin %s loaded.", name)
221+
logger.warning(
222+
"Platform plugin %s function's return value is None", name)
225223
except Exception:
226-
pass
224+
logger.exception("Failed to load platform plugin %s", name)
227225

228226
activated_builtin_plugins = list(
229227
set(activated_plugins) & set(builtin_platform_plugins.keys()))

vllm/platforms/hpu.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -39,8 +39,8 @@ def get_attn_backend_cls(cls, selected_backend: _Backend, head_size: int,
3939
def is_async_output_supported(cls, enforce_eager: Optional[bool]) -> bool:
4040
return True
4141

42-
@staticmethod
43-
def inference_mode():
42+
@classmethod
43+
def inference_mode(cls):
4444
return torch.no_grad()
4545

4646
@classmethod

vllm/platforms/rocm.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -217,9 +217,9 @@ def get_device_capability(cls,
217217
major, minor = torch.cuda.get_device_capability(device_id)
218218
return DeviceCapability(major=major, minor=minor)
219219

220-
@staticmethod
220+
@classmethod
221221
@with_amdsmi_context
222-
def is_fully_connected(physical_device_ids: list[int]) -> bool:
222+
def is_fully_connected(cls, physical_device_ids: list[int]) -> bool:
223223
"""
224224
Query if the set of gpus are fully connected by xgmi (1 hop)
225225
"""

vllm/platforms/xpu.py

Lines changed: 8 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -37,15 +37,17 @@ def get_attn_backend_cls(cls, selected_backend: _Backend, head_size: int,
3737
logger.info("Using IPEX attention backend.")
3838
return "vllm.attention.backends.ipex_attn.IpexAttnBackend"
3939

40-
@staticmethod
40+
@classmethod
4141
def get_device_capability(
42-
device_id: int = 0) -> Optional[DeviceCapability]:
42+
cls,
43+
device_id: int = 0,
44+
) -> Optional[DeviceCapability]:
4345
# capacity format differs from cuda's and will cause unexpected
4446
# failure, so use None directly
4547
return None
4648

47-
@staticmethod
48-
def get_device_name(device_id: int = 0) -> str:
49+
@classmethod
50+
def get_device_name(cls, device_id: int = 0) -> str:
4951
return torch.xpu.get_device_name(device_id)
5052

5153
@classmethod
@@ -57,8 +59,8 @@ def get_device_total_memory(cls, device_id: int = 0) -> int:
5759
def is_async_output_supported(cls, enforce_eager: Optional[bool]) -> bool:
5860
return True
5961

60-
@staticmethod
61-
def inference_mode():
62+
@classmethod
63+
def inference_mode(cls):
6264
return torch.no_grad()
6365

6466
@classmethod

0 commit comments

Comments
 (0)