We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent 2b0c87e commit 7742d8bCopy full SHA for 7742d8b
tensorrt_llm/scaffolding/worker.py
@@ -4,7 +4,7 @@
4
import openai
5
from transformers import AutoTokenizer
6
7
-from tensorrt_llm._tensorrt_engine import LLM
+from tensorrt_llm import LLM
8
from tensorrt_llm.executor import GenerationExecutor
9
from tensorrt_llm.llmapi.llm_args import KvCacheConfig
10
from tensorrt_llm.sampling_params import SamplingParams
@@ -150,7 +150,6 @@ def init_with_new_llm(
150
)
151
152
llm = LLM(model_dir,
153
- backend=backend,
154
tokenizer=tokenizer,
155
mixed_sampler=True,
156
disable_overlap_scheduler=disable_overlap_scheduler,
0 commit comments