🚀 The feature, motivation and pitch
We currently support Ray-based distributed inference, which requires Ray. This issue requests multi-node support for torch.distributed.
Usage Example:
# Server 1
vllm serve model_tag --nnodes 2 --rank 0 --dist-init-addr 192.168.0.1:5000
# Server 2
vllm serve model_tag --nnodes 2 --rank 1 --dist-init-addr 192.168.0.2:5000
Alternatives
No response
Additional context
No response
Before submitting a new issue...