Skip to content
New issue

Have a question about this project? # for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “#”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? # to your account

[V1][PP] Fix & Pin Ray version in requirements-cuda.txt #13436

Merged
merged 1 commit into from
Feb 18, 2025
Merged

Conversation

WoosukKwon
Copy link
Collaborator

@WoosukKwon WoosukKwon commented Feb 17, 2025

Pipeline parallelism in V1 requires ray[adag] instead of ray[default].
Also, because of the API changes in 2.42.0, we have to pin the version to 2.41.0 (or 2.40.0).

NOTE: Importantly, having ray[adag] will add CuPy (cu12) as a dependency. Since PP is not used for all models, we can consider keeping ray[adag] as an optional dependency if it's not acceptable.

Signed-off-by: Woosuk Kwon <woosuk.kwon@berkeley.edu>
Copy link

👋 Hi! Thank you for contributing to the vLLM project.

💬 Join our developer Slack at https://slack.vllm.ai to discuss your PR in #pr-reviews, coordinate on features in #feat- channels, or join special interest groups in #sig- channels.

Just a reminder: PRs would not trigger full CI run by default. Instead, it would only run fastcheck CI which starts running only a small and essential subset of CI tests to quickly catch errors. You can run other CI tests on top of those by going to your fastcheck build on Buildkite UI (linked in the PR checks section) and unblock them. If you do not have permission to unblock, ping simon-mo or khluu to add you in our Buildkite org.

Once the PR is approved and ready to go, your PR reviewer(s) can run CI to test the changes comprehensively before merging.

To run CI, PR reviewers can either: Add ready label to the PR or enable auto-merge.

🚀

@mergify mergify bot added the ci/build label Feb 17, 2025
@comaniac
Copy link
Collaborator

cc @ruisearch42 @richardliaw

@youkaichao
Copy link
Member

i think it’s fine, as long as vllm does not directly use cupy .
one thing about cupy is, it has cupy-cuda11x and cupy-cuda12x . I’m not sure how ray deals with it. will it break the cuda 11.8 build of vllm?

@WoosukKwon
Copy link
Collaborator Author

@youkaichao It seems to use cupy-cu12. However, IIUC, it doesn't break anything on our cu11.8 build unless the user explicitly chooses Ray?

@youkaichao
Copy link
Member

@youkaichao It seems to use cupy-cu12. However, IIUC, it doesn't break anything on our cu11.8 build unless the user explicitly chooses Ray?

sounds good, then it's a ray-related issue, whether they want to support cuda 11.8 . we can go ahead with ray[adag] .

@ruisearch42
Copy link
Collaborator

ray[adag] uses cupy-cuda12x. BTW, there is an issue in ray 2.42 and is being fixed. After that we can upgrade to the latest version with a small API change.

@WoosukKwon WoosukKwon merged commit 9915912 into main Feb 18, 2025
24 checks passed
@WoosukKwon WoosukKwon deleted the v1-ray-pp branch February 18, 2025 05:58
panf2333 pushed a commit to yottalabsai/vllm that referenced this pull request Feb 18, 2025
…#13436)

Signed-off-by: Woosuk Kwon <woosuk.kwon@berkeley.edu>
xjpang pushed a commit to xjpang/vllm that referenced this pull request Feb 20, 2025
…#13436)

Signed-off-by: Woosuk Kwon <woosuk.kwon@berkeley.edu>
kerthcet pushed a commit to kerthcet/vllm that referenced this pull request Feb 21, 2025
…#13436)

Signed-off-by: Woosuk Kwon <woosuk.kwon@berkeley.edu>
@darthhexx
Copy link
Contributor

The issue with pinning to a specific Ray version is that anyone with a long running cluster will not be able to upgrade vLLM services unless they upgrade the entire Ray cluster.

Please can we rather look at a range specifier (i.e. ray[adag]>=2.43.0), once 2.43.0 comes out with the fix?

@ruisearch42
Copy link
Collaborator

Hi @darthhexx , that makes sense. This is a short term fix and the plan is to support a ray version range.

Akshat-Tripathi pushed a commit to krai/vllm that referenced this pull request Mar 3, 2025
…#13436)

Signed-off-by: Woosuk Kwon <woosuk.kwon@berkeley.edu>
lk-chen pushed a commit to lk-chen/vllm that referenced this pull request Mar 5, 2025
…#13436)

Signed-off-by: Woosuk Kwon <woosuk.kwon@berkeley.edu>
Signed-off-by: Linkun Chen <github@lkchen.net>
Said-Akbar pushed a commit to Said-Akbar/vllm-rocm that referenced this pull request Mar 7, 2025
…#13436)

Signed-off-by: Woosuk Kwon <woosuk.kwon@berkeley.edu>
Signed-off-by: saeediy <saidakbarp@gmail.com>
# for free to join this conversation on GitHub. Already have an account? # to comment
Labels
Projects
None yet
Development

Successfully merging this pull request may close these issues.

5 participants