Skip to content

Conversation

tjohnson31415
Copy link
Contributor

@tjohnson31415 tjohnson31415 commented Mar 13, 2025

Fix for the crash repro reported in this comment. The bug and fix are pretty similar to #12347 in that the problem arises in a batch of mixed text and image requests leading to the attn metadata having some lists with an element for each sequence and others with an element only for each sequence-with-images.

Another fix that came out of #10648

FIX #10648

Copy link

👋 Hi! Thank you for contributing to the vLLM project.

💬 Join our developer Slack at https://slack.vllm.ai to discuss your PR in #pr-reviews, coordinate on features in #feat- channels, or join special interest groups in #sig- channels.

Just a reminder: PRs would not trigger full CI run by default. Instead, it would only run fastcheck CI which starts running only a small and essential subset of CI tests to quickly catch errors. You can run other CI tests on top of those by going to your fastcheck build on Buildkite UI (linked in the PR checks section) and unblock them. If you do not have permission to unblock, ping simon-mo or khluu to add you in our Buildkite org.

Once the PR is approved and ready to go, your PR reviewer(s) can run CI to test the changes comprehensively before merging.

To run CI, PR reviewers can either: Add ready label to the PR or enable auto-merge.

🚀

@tjohnson31415 tjohnson31415 mentioned this pull request Mar 13, 2025
1 task
@tjohnson31415 tjohnson31415 marked this pull request as ready for review March 24, 2025 19:38
Copy link
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

input_processor_for_mllama has been removed and I don't see our use of the Transformers processor doing this same trick to only process the last image group.

I'm working to undersatnd this better, but I think we either should re-implement the "cheat" or remove the extra complexity (i.e. this check comparing to num tiles and remove kv_range_for_decode).

Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

I'm checking with the author of PR #11427 which removes the previous input processor. See the discussion here.
https://vllm-dev.slack.com/archives/C07QCGVDNUF/p1743001000770969

Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

This cheat is implementing here #15564

Copy link

mergify bot commented Mar 28, 2025

This pull request has merge conflicts that must be resolved before it can be
merged. Please rebase the PR, @tjohnson31415.

https://docs.github.com/en/pull-requests/collaborating-with-pull-requests/working-with-forks/syncing-a-fork

@mergify mergify bot added the needs-rebase label Mar 28, 2025
@heheda12345
Copy link
Collaborator

In general this PR looks good. Thanks for the fix. Can you sync this PR with main to get in recent bug fixes for mllama #15564 and #14883 and test it again?

@tjohnson31415
Copy link
Contributor Author

@heheda12345 I rebased and updated the branch.
tests/models/encoder_decoder/vision_language/test_mllama.py tests all pass for me.

Ready for another look!

Copy link
Collaborator

@heheda12345 heheda12345 left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

LGTM in general. Only some nits.

Signed-off-by: Travis Johnson <[email protected]>
@tjohnson31415
Copy link
Contributor Author

@heheda12345 Thanks for the review! I pushed the suggested changes (the linter didn't like l as a var name so used x instead). PTAL

Copy link
Collaborator

@heheda12345 heheda12345 left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Thanks for the bug fix!

@heheda12345 heheda12345 enabled auto-merge (squash) April 11, 2025 18:03
@github-actions github-actions bot added the ready ONLY add when PR is ready to merge/full CI is needed label Apr 11, 2025
@heheda12345 heheda12345 merged commit 71b9cde into vllm-project:main Apr 11, 2025
61 checks passed
@tjohnson31415 tjohnson31415 deleted the fix-mllama-crash branch April 11, 2025 22:08
yangw-dev pushed a commit to yangw-dev/vllm that referenced this pull request Apr 21, 2025
jikunshang pushed a commit to jikunshang/vllm that referenced this pull request Apr 29, 2025
lk-chen pushed a commit to lk-chen/vllm that referenced this pull request Apr 29, 2025
RichardoMrMu pushed a commit to RichardoMrMu/vllm that referenced this pull request May 12, 2025
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

ready ONLY add when PR is ready to merge/full CI is needed

Projects

None yet

Development

Successfully merging this pull request may close these issues.

[Bug]: Llama 3.2 90b crash

2 participants