-
-
Notifications
You must be signed in to change notification settings - Fork 6.2k
New issue
Have a question about this project? # for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “#”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? # to your account
[Bug]: I try to use vllm==0.6.5 for GLM4-9b-chat but error "/usr/bin/ld: cannot find -lcuda" #11643
Comments
Maybe the cause is your cuda runtime is old:
|
Thanks, by the way, could you please tell me which version of the CUDA runtime is required? |
I am using version 12.4 |
Same issue with 12.4 for me |
The issue still exists for latest nightly builds. Is someone looking into this? Using Cuda 12.4 |
Same here. I updated our DGX because I figured it was a cuda-issue as well, but that doesn't seem to help. (though I was on 12.4 already before updating). I'm using the docker containers, and it also happens for the newer 0.6.6. |
I've looked into this a bit more. We're using the nvidia-gpu-operator to deploy vllm via helm charts. I'm also no expert in this area at all, but maybe this helps someone smarter than me. I think there's a few separate issues going one. One of them should be fixed in the next version (see #12505). I think for that issue the symlink fix should work, but I think if you've been having the issue before 0.7.0 already, it's something else. I've connected to the containers manually, and calling 0.6.4.post1:
0.7.0:
I'm not sure why there's a downgrade from 12.4 to 12.1 here, these run on the same host system, the only difference here is the vllm container. Maybe more interestingly though is the missing libcuda.so in /usr/lib. Adding a link there as described in #12505 seems to fix the issue. On the host system everything seems correct:
This makes me think this is more likely to be an issue with the nvidia-gpu-operator than with vllm itself, but once again I'm not an expert in this at all, this is just what I noticed. |
I had similar issues, may be this could help. |
Your current environment
The output of `python collect_env.py`
Model Input Dumps
🐛 Describe the bug
I try to use glm4-9b-chat with vllm==0.6.5, but the error say
/usr/bin/ld: cannot find -lcuda
, when I use vllm==0.5.5, it can works.So anybody know what's going on here, what's the problem?
Before submitting a new issue...
The text was updated successfully, but these errors were encountered: