-
-
Couldn't load subscription status.
- Fork 10.9k
Fix torch version check for SM100 mxfp4 #22535
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
|
👋 Hi! Thank you for contributing to the vLLM project. 💬 Join our developer Slack at https://slack.vllm.ai to discuss your PR in #pr-reviews, coordinate on features in #feat- channels, or join special interest groups in #sig- channels. Just a reminder: PRs would not trigger full CI run by default. Instead, it would only run Once the PR is approved and ready to go, your PR reviewer(s) can run CI to test the changes comprehensively before merging. To run CI, PR reviewers can either: Add 🚀 |
Signed-off-by: Zifei Tong <[email protected]>
49ef8af to
29032dd
Compare
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Code Review
This pull request fixes an issue where the torch version check for mxfp4 was overly broad, applying to all GPU architectures instead of just Hopper. The change correctly scopes this check, along with the Triton kernel check, to only apply to Hopper GPUs. This allows using mxfp4 on newer architectures like Blackwell with different torch versions.
My main feedback is to consider adding explicit checks for Blackwell and newer architectures to make the code more robust and prevent potential runtime errors if their specific requirements are not met. I've provided a suggestion to add these checks.
Signed-off-by: Zifei Tong <[email protected]>
f97e453 to
7efd224
Compare
|
Your understanding is correct. Thank you. |
Signed-off-by: Zifei Tong <[email protected]> Signed-off-by: mgoin <[email protected]> Co-authored-by: mgoin <[email protected]> Signed-off-by: Diego-Castan <[email protected]>
Signed-off-by: Zifei Tong <[email protected]> Signed-off-by: mgoin <[email protected]> Co-authored-by: mgoin <[email protected]>
Signed-off-by: Zifei Tong <[email protected]> Signed-off-by: mgoin <[email protected]> Co-authored-by: mgoin <[email protected]>
Signed-off-by: Zifei Tong <[email protected]> Signed-off-by: mgoin <[email protected]> Co-authored-by: mgoin <[email protected]>
Signed-off-by: Zifei Tong <[email protected]> Signed-off-by: mgoin <[email protected]> Co-authored-by: mgoin <[email protected]> Signed-off-by: Xiao Yu <[email protected]>
Signed-off-by: Zifei Tong <[email protected]> Signed-off-by: mgoin <[email protected]> Co-authored-by: mgoin <[email protected]>
Fixes gpt-oss models on main with blackwell.
My understanding is that mxfp4+blackwell+torch 2.7 is supported.
Essential Elements of an Effective PR Description Checklist
supported_models.mdandexamplesfor a new model.Purpose
Test Plan
Test Result
(Optional) Documentation Update