|
12 | 12 | from torch.cuda import device_count |
13 | 13 |
|
14 | 14 | from tensorrt_llm import LLM as PyTorchLLM |
15 | | -from tensorrt_llm._torch.auto_deploy.llm import LLM as AutoDeployLLM |
16 | 15 | from tensorrt_llm import MultimodalEncoder |
17 | 16 | from tensorrt_llm._tensorrt_engine import LLM |
| 17 | +from tensorrt_llm._torch.auto_deploy.llm import LLM as AutoDeployLLM |
18 | 18 | from tensorrt_llm._utils import mpi_rank |
19 | 19 | from tensorrt_llm.executor.utils import LlmLauncherEnvs |
20 | 20 | from tensorrt_llm.llmapi import (BuildConfig, CapacitySchedulerPolicy, |
@@ -211,10 +211,13 @@ def launch_mm_encoder_server( |
211 | 211 | default="localhost", |
212 | 212 | help="Hostname of the server.") |
213 | 213 | @click.option("--port", type=int, default=8000, help="Port of the server.") |
214 | | -@click.option("--backend", |
215 | | - type=click.Choice(["pytorch", "trt", "_autodeploy"]), |
216 | | - default="pytorch", |
217 | | - help="Set to 'pytorch' for pytorch path and '_autodeploy' for autodeploy path. Default is pytorch path.") |
| 214 | +@click.option( |
| 215 | + "--backend", |
| 216 | + type=click.Choice(["pytorch", "trt", "_autodeploy"]), |
| 217 | + default="pytorch", |
| 218 | + help= |
| 219 | + "Set to 'pytorch' for pytorch path and '_autodeploy' for autodeploy path. Default is pytorch path." |
| 220 | +) |
218 | 221 | @click.option('--log_level', |
219 | 222 | type=click.Choice(severity_map.keys()), |
220 | 223 | default='info', |
|
0 commit comments