Skip to content

Commit 948b8b9

Browse files
[None][fix] Fix CUDA graph for Qwen2.5-VL (#8047)
Signed-off-by: yechank <[email protected]>
1 parent 1dba9fa commit 948b8b9

File tree

1 file changed

+9
-0
lines changed

1 file changed

+9
-0
lines changed

tensorrt_llm/_torch/pyexecutor/_util.py

Lines changed: 9 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -173,6 +173,15 @@ def _create_dummy_context_requests(
173173
beam_width=max_beam_width, ),
174174
output_config=trtllm.OutputConfig(),
175175
end_id=-1)
176+
if self._model_engine.use_mrope:
177+
request.py_multimodal_data = {
178+
"mrope_config": {
179+
"mrope_position_ids":
180+
torch.zeros(3, 1, input_seq_len, dtype=torch.int32),
181+
"mrope_position_deltas":
182+
torch.zeros(1, 1, dtype=torch.int32)
183+
}
184+
}
176185
requests.append(request)
177186
remaining_tokens -= input_seq_len
178187
if self._mapping.enable_attention_dp:

0 commit comments

Comments
 (0)