Skip to content

Conversation

@youkaichao
Copy link
Member

No description provided.

@github-actions
Copy link

👋 Hi! Thank you for contributing to the vLLM project.
Just a reminder: PRs would not trigger full CI run by default. Instead, it would only trigger fastcheck CI to run, which consists only a small and essential subset of tests to quickly catch errors with the flexibility to run extra individual tests on top (you can do this by unblocking test steps in the Buildkite run).

Full CI run is still required to merge this PR so once the PR is ready to go, please make sure to run it. If you need all test signals in between PR commits, you can trigger full CI as well.

To run full CI, you can do one of these:

  • Comment /ready on the PR
  • Add ready label to the PR
  • Enable auto-merge.

🚀

@youkaichao youkaichao requested a review from zhuohan123 July 13, 2024 18:51
@simon-mo
Copy link
Collaborator

If users want to increase the throughput and lower the latency of a small model, they should also use TP

@youkaichao
Copy link
Member Author

If users want to increase the throughput and lower the latency of a small model, they should also use TP

true (mainly for increasing throughput I think). added in 0d99585

@youkaichao youkaichao merged commit 94b82e8 into vllm-project:main Jul 15, 2024
@youkaichao youkaichao deleted the dist_suggest branch July 15, 2024 16:46
Alvant pushed a commit to compressa-ai/vllm that referenced this pull request Oct 26, 2024
LeiWang1999 pushed a commit to LeiWang1999/vllm-bitblas that referenced this pull request Mar 26, 2025
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants