diff --git a/vllm/executor/tpu_executor.py b/vllm/executor/tpu_executor.py index 6627ee6984d..d906a6cc39d 100644 --- a/vllm/executor/tpu_executor.py +++ b/vllm/executor/tpu_executor.py @@ -81,8 +81,7 @@ def initialize_cache( def determine_num_available_blocks(self) -> Tuple[int, int]: """Determine the number of available KV blocks by invoking the - underlying worker. - """ + underlying worker.""" return self.driver_worker.determine_num_available_blocks() def execute_model( @@ -93,16 +92,36 @@ def execute_model( return output def add_lora(self, lora_request: LoRARequest) -> bool: - raise NotImplementedError("LoRA is not implemented for TPU backend.") + raise NotImplementedError( + "LoRA is currently not supported by the TPU backend.") def remove_lora(self, lora_id: int) -> bool: - raise NotImplementedError("LoRA is not implemented for TPU backend.") + raise NotImplementedError( + "LoRA is currently not supported by the TPU backend.") def pin_lora(self, lora_id: int) -> bool: - raise NotImplementedError("LoRA is not implemented for TPU backend.") + raise NotImplementedError( + "LoRA is currently not supported by the TPU backend.") def list_loras(self) -> Set[int]: - raise NotImplementedError("LoRA is not implemented for TPU backend.") + raise NotImplementedError( + "LoRA is currently not supported by the TPU backend.") + + def add_prompt_adapter(self, prompt_adapter_request) -> bool: + raise NotImplementedError( + "Soft prompt is currently not supported by the TPU backend.") + + def remove_prompt_adapter(self, prompt_adapter_id: int) -> bool: + raise NotImplementedError( + "Soft prompt is currently not supported by the TPU backend.") + + def pin_prompt_adapter(self, prompt_adapter_id: int) -> bool: + raise NotImplementedError( + "Soft prompt is currently not supported by the TPU backend.") + + def list_prompt_adapters(self) -> Set[int]: + raise NotImplementedError( + "Soft prompt is currently not supported by the TPU backend.") def check_health(self) -> None: # TPUExecutor will always be healthy as long as it's running.