62 lines
2.1 KiB
Docker
62 lines
2.1 KiB
Docker
# oneapi 2025.0.2 docker base image use rolling 2448 package. https://dgpu-docs.intel.com/releases/packages.html?release=Rolling+2448.13&os=Ubuntu+22.04, and we don't need install driver manually.
|
|
FROM intel/deep-learning-essentials:2025.0.2-0-devel-ubuntu22.04 AS vllm-base
|
|
|
|
RUN rm /etc/apt/sources.list.d/intel-graphics.list
|
|
|
|
RUN apt-get update -y && \
|
|
apt-get install -y --no-install-recommends --fix-missing \
|
|
curl \
|
|
ffmpeg \
|
|
git \
|
|
libsndfile1 \
|
|
libsm6 \
|
|
libxext6 \
|
|
libgl1 \
|
|
lsb-release \
|
|
numactl \
|
|
python3 \
|
|
python3-dev \
|
|
python3-pip \
|
|
wget
|
|
|
|
WORKDIR /workspace/vllm
|
|
COPY requirements/xpu.txt /workspace/vllm/requirements/xpu.txt
|
|
COPY requirements/common.txt /workspace/vllm/requirements/common.txt
|
|
|
|
RUN --mount=type=cache,target=/root/.cache/pip \
|
|
pip install --no-cache-dir \
|
|
-r requirements/xpu.txt
|
|
|
|
ENV LD_LIBRARY_PATH="$LD_LIBRARY_PATH:/usr/local/lib/"
|
|
|
|
COPY . .
|
|
ARG GIT_REPO_CHECK=0
|
|
RUN --mount=type=bind,source=.git,target=.git \
|
|
if [ "$GIT_REPO_CHECK" != 0 ]; then bash tools/check_repo.sh; fi
|
|
|
|
ENV VLLM_TARGET_DEVICE=xpu
|
|
|
|
RUN --mount=type=cache,target=/root/.cache/pip \
|
|
--mount=type=bind,source=.git,target=.git \
|
|
python3 setup.py install
|
|
|
|
# Please refer xpu doc, we need manually install intel-extension-for-pytorch 2.6.10+xpu due to there are some conflict dependencies with torch 2.6.0+xpu
|
|
# FIXME: This will be fix in ipex 2.7. just leave this here for awareness.
|
|
RUN --mount=type=cache,target=/root/.cache/pip \
|
|
pip install intel-extension-for-pytorch==2.6.10+xpu \
|
|
--extra-index-url=https://pytorch-extension.intel.com/release-whl/stable/xpu/us/
|
|
|
|
CMD ["/bin/bash"]
|
|
|
|
FROM vllm-base AS vllm-openai
|
|
|
|
# install additional dependencies for openai api server
|
|
RUN --mount=type=cache,target=/root/.cache/pip \
|
|
pip install accelerate hf_transfer 'modelscope!=1.15.0'
|
|
|
|
ENV VLLM_USAGE_SOURCE production-docker-image \
|
|
TRITON_XPU_PROFILE 1
|
|
# install development dependencies (for testing)
|
|
RUN python3 -m pip install -e tests/vllm_test_utils
|
|
ENTRYPOINT ["python3", "-m", "vllm.entrypoints.openai.api_server"]
|