# oneapi 2025.0.2 docker base image use rolling 2448 package. https://dgpu-docs.intel.com/releases/packages.html?release=Rolling+2448.13&os=Ubuntu+22.04, and we don't need install driver manually. FROM intel/deep-learning-essentials:2025.0.2-0-devel-ubuntu22.04 AS vllm-base RUN rm /etc/apt/sources.list.d/intel-graphics.list RUN apt-get update -y && \ apt-get install -y --no-install-recommends --fix-missing \ curl \ ffmpeg \ git \ libsndfile1 \ libsm6 \ libxext6 \ libgl1 \ lsb-release \ numactl \ python3 \ python3-dev \ python3-pip \ wget WORKDIR /workspace/vllm COPY requirements/xpu.txt /workspace/vllm/requirements/xpu.txt COPY requirements/common.txt /workspace/vllm/requirements/common.txt RUN --mount=type=cache,target=/root/.cache/pip \ pip install --no-cache-dir \ -r requirements/xpu.txt ENV LD_LIBRARY_PATH="$LD_LIBRARY_PATH:/usr/local/lib/" COPY . . ARG GIT_REPO_CHECK=0 RUN --mount=type=bind,source=.git,target=.git \ if [ "$GIT_REPO_CHECK" != 0 ]; then bash tools/check_repo.sh; fi ENV VLLM_TARGET_DEVICE=xpu RUN --mount=type=cache,target=/root/.cache/pip \ --mount=type=bind,source=.git,target=.git \ python3 setup.py install # Please refer xpu doc, we need manually install intel-extension-for-pytorch 2.6.10+xpu due to there are some conflict dependencies with torch 2.6.0+xpu # FIXME: This will be fix in ipex 2.7. just leave this here for awareness. RUN --mount=type=cache,target=/root/.cache/pip \ pip install intel-extension-for-pytorch==2.6.10+xpu \ --extra-index-url=https://pytorch-extension.intel.com/release-whl/stable/xpu/us/ CMD ["/bin/bash"] FROM vllm-base AS vllm-openai # install additional dependencies for openai api server RUN --mount=type=cache,target=/root/.cache/pip \ pip install accelerate hf_transfer 'modelscope!=1.15.0' ENV VLLM_USAGE_SOURCE production-docker-image \ TRITON_XPU_PROFILE 1 # install development dependencies (for testing) RUN python3 -m pip install -e tests/vllm_test_utils ENTRYPOINT ["python3", "-m", "vllm.entrypoints.openai.api_server"]