You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: docs/contributing/ci/update_pytorch_version.md
+4-22Lines changed: 4 additions & 22 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -87,7 +87,7 @@ is ineffective.
87
87
88
88
While ongoing efforts like <https://github.com/vllm-project/vllm/issues/17419>
89
89
address the long build time at its source, the current workaround is to set `VLLM_CI_BRANCH`
90
-
to a custom branch provided by @khluu (`VLLM_CI_BRANCH=khluu/use_postmerge_q`)
90
+
to a custom branch provided by @khluu (`VLLM_CI_BRANCH=khluu/long_build`)
91
91
when manually triggering a build on Buildkite. This branch accomplishes two things:
92
92
93
93
1. Increase the timeout limit to 10 hours so that the build doesn't time out.
@@ -100,35 +100,17 @@ to warm it up so that future builds are faster.
100
100
101
101
## Update dependencies
102
102
103
-
Several vLLM dependencies, such as FlashInfer, also depend on PyTorch and need
103
+
Several vLLM dependencies like xFormers depend on PyTorch and need
104
104
to be updated accordingly. Rather than waiting for all of them to publish new
105
105
releases (which would take too much time), they can be built from
106
106
source to unblock the update process.
107
107
108
-
### FlashInfer
109
-
110
-
Here is how to build and install it from source with `torch2.7.0+cu128`in vLLM [Dockerfile](https://github.com/vllm-project/vllm/blob/27bebcd89792d5c4b08af7a65095759526f2f9e1/docker/Dockerfile#L259-L271):
One caveat is that building FlashInfer from source adds approximately 30
120
-
minutes to the vLLM build time. Therefore, it's preferable to cache the wheel in a
121
-
public location for immediate installation, such as [this FlashInfer wheel link](https://download.pytorch.org/whl/cu128/flashinfer/flashinfer_python-0.2.6.post1%2Bcu128torch2.7-cp39-abi3-linux_x86_64.whl). For future releases, contact the PyTorch release
122
-
team if you want to get the package published there.
123
-
124
108
### xFormers
125
109
126
-
Similar to FlashInfer, here is how to build and install xFormers from source:
0 commit comments