Add documentation on how to do incremental builds (#2796)
This commit is contained in:
parent
c81dddb45c
commit
931746bc6d
@ -67,3 +67,13 @@ You can also build and install vLLM from source:
|
|||||||
|
|
||||||
$ # Use `--ipc=host` to make sure the shared memory is large enough.
|
$ # Use `--ipc=host` to make sure the shared memory is large enough.
|
||||||
$ docker run --gpus all -it --rm --ipc=host nvcr.io/nvidia/pytorch:23.10-py3
|
$ docker run --gpus all -it --rm --ipc=host nvcr.io/nvidia/pytorch:23.10-py3
|
||||||
|
|
||||||
|
.. note::
|
||||||
|
If you are developing the C++ backend of vLLM, consider building vLLM with
|
||||||
|
|
||||||
|
.. code-block:: console
|
||||||
|
|
||||||
|
$ python setup.py develop
|
||||||
|
|
||||||
|
since it will give you incremental builds. The downside is that this method
|
||||||
|
is `deprecated by setuptools <https://github.com/pypa/setuptools/issues/917>`_.
|
||||||
|
5
setup.py
5
setup.py
@ -15,6 +15,11 @@ from torch.utils.cpp_extension import BuildExtension, CUDAExtension, CUDA_HOME,
|
|||||||
|
|
||||||
ROOT_DIR = os.path.dirname(__file__)
|
ROOT_DIR = os.path.dirname(__file__)
|
||||||
|
|
||||||
|
# If you are developing the C++ backend of vLLM, consider building vLLM with
|
||||||
|
# `python setup.py develop` since it will give you incremental builds.
|
||||||
|
# The downside is that this method is deprecated, see
|
||||||
|
# https://github.com/pypa/setuptools/issues/917
|
||||||
|
|
||||||
MAIN_CUDA_VERSION = "12.1"
|
MAIN_CUDA_VERSION = "12.1"
|
||||||
|
|
||||||
# Supported NVIDIA GPU architectures.
|
# Supported NVIDIA GPU architectures.
|
||||||
|
Loading…
x
Reference in New Issue
Block a user