diff --git a/Dockerfile b/Dockerfile index e8e1bf5..252fb22 100644 --- a/Dockerfile +++ b/Dockerfile @@ -401,34 +401,34 @@ # RUN mv mkdocs.yaml test_docs/ # #################### TEST IMAGE #################### -# #################### OPENAI API SERVER #################### -# # base openai image with additional requirements, for any subsequent openai-style images -# FROM vllm-base AS vllm-openai-base -# ARG TARGETPLATFORM -# ARG INSTALL_KV_CONNECTORS=false +#################### OPENAI API SERVER #################### +# base openai image with additional requirements, for any subsequent openai-style images +FROM vllm-base AS vllm-openai-base +ARG TARGETPLATFORM +ARG INSTALL_KV_CONNECTORS=false -# ARG PIP_INDEX_URL UV_INDEX_URL -# ARG PIP_EXTRA_INDEX_URL UV_EXTRA_INDEX_URL +ARG PIP_INDEX_URL UV_INDEX_URL +ARG PIP_EXTRA_INDEX_URL UV_EXTRA_INDEX_URL -# # This timeout (in seconds) is necessary when installing some dependencies via uv since it's likely to time out -# # Reference: https://github.com/astral-sh/uv/pull/1694 -# ENV UV_HTTP_TIMEOUT=500 +# This timeout (in seconds) is necessary when installing some dependencies via uv since it's likely to time out +# Reference: https://github.com/astral-sh/uv/pull/1694 +ENV UV_HTTP_TIMEOUT=500 -# COPY ./vllm_v0.10.0/requirements/kv_connectors.txt requirements/kv_connectors.txt +COPY ./vllm_v0.10.0/requirements/kv_connectors.txt requirements/kv_connectors.txt -# # install additional dependencies for openai api server -# RUN --mount=type=cache,target=/root/.cache/uv \ -# if [ "$INSTALL_KV_CONNECTORS" = "true" ]; then \ -# uv pip install --system -r requirements/kv_connectors.txt; \ -# fi; \ -# if [ "$TARGETPLATFORM" = "linux/arm64" ]; then \ -# BITSANDBYTES_VERSION="0.42.0"; \ -# else \ -# BITSANDBYTES_VERSION="0.46.1"; \ -# fi; \ -# uv pip install --system accelerate hf_transfer modelscope "bitsandbytes>=${BITSANDBYTES_VERSION}" 'timm==0.9.10' boto3 runai-model-streamer runai-model-streamer[s3] +# install additional dependencies for openai api server +RUN --mount=type=cache,target=/root/.cache/uv \ + if [ "$INSTALL_KV_CONNECTORS" = "true" ]; then \ + uv pip install --system -r requirements/kv_connectors.txt; \ + fi; \ + if [ "$TARGETPLATFORM" = "linux/arm64" ]; then \ + BITSANDBYTES_VERSION="0.42.0"; \ + else \ + BITSANDBYTES_VERSION="0.46.1"; \ + fi; \ + uv pip install --system accelerate hf_transfer modelscope "bitsandbytes>=${BITSANDBYTES_VERSION}" 'timm==0.9.10' boto3 runai-model-streamer runai-model-streamer[s3] -# ENV VLLM_USAGE_SOURCE production-docker-image +ENV VLLM_USAGE_SOURCE production-docker-image # # define sagemaker first, so it is not default from `docker build` # FROM vllm-openai-base AS vllm-sagemaker