Skip to content

Conversation

@Semmer2
Copy link

@Semmer2 Semmer2 commented Nov 6, 2025

Replace Qwen2_5_Omni_Thinker model's original vision tower Qwen2_5_VisionTransformer with AscendQwen2_5_VisionTransoformer. It pads attention qkv weights and bias for better performance.

What this PR does / why we need it?

This PR repalce the vision tower in Qwen2.5-Omni-Thinker model, Qwen2_5_VisionTransformer, with AscendQwen2_5_VisionTransformer, which use QKV padding for padding performance.

Does this PR introduce any user-facing change?

No

How was this patch tested?

@github-actions
Copy link

github-actions bot commented Nov 6, 2025

👋 Hi! Thank you for contributing to the vLLM Ascend project. The following points will speed up your PR merge:‌‌

  • A PR should do only one thing, smaller PRs enable faster reviews.
  • Every PR should include unit tests and end-to-end tests ‌to ensure it works and is not broken by other future PRs.
  • Write the commit message by fulfilling the PR description to help reviewer and future developers understand.

If CI fails, you can run linting and testing checks locally according Contributing and Testing.

Copy link
Contributor

@gemini-code-assist gemini-code-assist bot left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Code Review

This pull request introduces performance optimizations for the Qwen2.5-Omni-Thinker model on Ascend hardware by replacing the standard vision tower with an optimized version that uses QKV padding. The changes involve registering the new Ascend-specific model, implementing the model class which uses AscendQwen2_5_VisionTransformer, and updating the weight loading logic to support the new attn.qkv naming and apply padding.

My review has identified a critical issue in the new model's initialization that would cause a runtime error, and a high-severity issue regarding the robustness of the weight padding logic, which currently only handles one type of QKV layer naming. Please see the detailed comments for suggestions.

@Semmer2 Semmer2 force-pushed the 0.11.0-dev branch 3 times, most recently from 0cb33c6 to e8c0771 Compare November 7, 2025 09:56
Replace Qwen2_5_Omni_Thinker model's original vision tower
Qwen2_5_VisionTransformer with AscendQwen2_5_VisionTransoformer.
It pads attention qkv weights and bias for better performance.

Signed-off-by: Ting FU <[email protected]>
Copy link
Collaborator

@MengqingCao MengqingCao left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

I'm okay with merging this into 0.11.0-dev, lgtm

@MengqingCao MengqingCao merged commit f984256 into vllm-project:v0.11.0-dev Nov 8, 2025
16 checks passed
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants