-
-
Notifications
You must be signed in to change notification settings - Fork 6.2k
New issue
Have a question about this project? # for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “#”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? # to your account
[Installation]: Nvidia runtime issue? On new VLLM 0.7.0 #12505
Comments
I hit this as well during
|
@tlrmchlsmth would you have an idea? This seems related to #12424 |
@russellb this looks similar to what you were helping dan with |
Yep, does seem suspicious. Not sure what's going wrong though |
We might need to do something like this instead: |
I hit similar issue on my build instance (Ubuntu 20.04) and here is what i did to workaround the error:
The ld command looks only looking for libraries under certain locations. Since the libcuda.so is not under where it's looking for thus the error. After setting the softlink, vllm can build and run successfully. |
I tried setting LD_LIBRARY_PATH initially but it didn't work for me for some reason. The ld command still complained about -lcuda not found and I had to use the softlink. |
+1 |
This should have been fixed with #12552 so please wait for the next release to include the revert |
I assume this only happens for 0.7.0 for everyone here then, since the reverted change is relatively recent? I'm asking because I've been having this issue ever since 0.6.5 which would suggest a different root cause (or a different issue altogether) as also mentioned here #11643 |
+1 |
thanks it started working normally again in v0.7.1 |
Did it work for anyone else? |
+1 |
It works normally again for me since 0.7.1 |
Your current environment
docker run --runtime nvidia --gpus all -v ~/.cache/huggingface:/root/.cache/huggingface -p 8000:8000 --ipc=host -e VLLM_ENABLE_PREFIX_CACHING=true --name qwen2.5_20250128 vllm/vllm-openai:v0.7.0 --model Qwen/Qwen2.5-72B-Instruct --tensor-parallel-size=4 --gpu-memory-utilization=0.90 --enforce-eager --rope-scaling '{"type": "yarn","factor": 4,"original_max_position_embeddings": 32768}'
error:
/usr/bin/ld: cannot find -lcuda: No such file or directory
How you are installing vllm
docker
Before submitting a new issue...
The text was updated successfully, but these errors were encountered: