Skip to content

Commit e081f01

Browse files
youkaichaosiddharth9820
authored andcommitted
[doc] recommend pip instead of conda (vllm-project#8446)
1 parent 543e30b commit e081f01

File tree

1 file changed

+9
-5
lines changed

1 file changed

+9
-5
lines changed

docs/source/getting_started/installation.rst

+9-5
Original file line numberDiff line numberDiff line change
@@ -26,6 +26,10 @@ You can install vLLM using pip:
2626
$ # Install vLLM with CUDA 12.1.
2727
$ pip install vllm
2828
29+
.. note::
30+
31+
Although we recommend using ``conda`` to create and manage Python environments, it is highly recommended to use ``pip`` to install vLLM. This is because ``pip`` can install ``torch`` with separate library packages like ``NCCL``, while ``conda`` installs ``torch`` with statically linked ``NCCL``. This can cause issues when vLLM tries to use ``NCCL``. See `this issue <https://github.com/vllm-project/vllm/issues/8420>`_ for more details.
32+
2933
.. note::
3034

3135
As of now, vLLM's binaries are compiled with CUDA 12.1 and public PyTorch release versions by default.
@@ -34,7 +38,7 @@ You can install vLLM using pip:
3438
.. code-block:: console
3539
3640
$ # Install vLLM with CUDA 11.8.
37-
$ export VLLM_VERSION=0.4.0
41+
$ export VLLM_VERSION=0.6.1.post1
3842
$ export PYTHON_VERSION=310
3943
$ pip install https://github.com/vllm-project/vllm/releases/download/v${VLLM_VERSION}/vllm-${VLLM_VERSION}+cu118-cp${PYTHON_VERSION}-cp${PYTHON_VERSION}-manylinux1_x86_64.whl --extra-index-url https://download.pytorch.org/whl/cu118
4044
@@ -48,7 +52,7 @@ You can install vLLM using pip:
4852

4953
.. code-block:: console
5054
51-
$ export VLLM_VERSION=0.5.4 # vLLM's main branch version is currently set to latest released tag
55+
$ export VLLM_VERSION=0.6.1.post1 # vLLM's main branch version is currently set to latest released tag
5256
$ pip install https://vllm-wheels.s3.us-west-2.amazonaws.com/nightly/vllm-${VLLM_VERSION}-cp38-abi3-manylinux1_x86_64.whl
5357
$ # You can also access a specific commit
5458
$ # export VLLM_COMMIT=...
@@ -80,11 +84,11 @@ You can also build and install vLLM from source:
8084
8185
.. tip::
8286

83-
Building from source requires quite a lot compilation. If you are building from source for multiple times, it is beneficial to cache the compilation results. For example, you can install `ccache <https://github.com/ccache/ccache>`_ via either `conda install ccache` or `apt install ccache` . As long as `which ccache` command can find the `ccache` binary, it will be used automatically by the build system. After the first build, the subsequent builds will be much faster.
87+
Building from source requires quite a lot compilation. If you are building from source for multiple times, it is beneficial to cache the compilation results. For example, you can install `ccache <https://github.com/ccache/ccache>`_ via either ``conda install ccache`` or ``apt install ccache`` . As long as ``which ccache`` command can find the ``ccache`` binary, it will be used automatically by the build system. After the first build, the subsequent builds will be much faster.
8488

8589
.. tip::
8690
To avoid your system being overloaded, you can limit the number of compilation jobs
87-
to be run simultaneously, via the environment variable `MAX_JOBS`. For example:
91+
to be run simultaneously, via the environment variable ``MAX_JOBS``. For example:
8892

8993
.. code-block:: console
9094
@@ -99,7 +103,7 @@ You can also build and install vLLM from source:
99103
$ # Use `--ipc=host` to make sure the shared memory is large enough.
100104
$ docker run --gpus all -it --rm --ipc=host nvcr.io/nvidia/pytorch:23.10-py3
101105
102-
If you don't want to use docker, it is recommended to have a full installation of CUDA Toolkit. You can download and install it from `the official website <https://developer.nvidia.com/cuda-toolkit-archive>`_. After installation, set the environment variable `CUDA_HOME` to the installation path of CUDA Toolkit, and make sure that the `nvcc` compiler is in your `PATH`, e.g.:
106+
If you don't want to use docker, it is recommended to have a full installation of CUDA Toolkit. You can download and install it from `the official website <https://developer.nvidia.com/cuda-toolkit-archive>`_. After installation, set the environment variable ``CUDA_HOME`` to the installation path of CUDA Toolkit, and make sure that the ``nvcc`` compiler is in your ``PATH``, e.g.:
103107

104108
.. code-block:: console
105109

0 commit comments

Comments
 (0)