Skip to content

Commit 4a2f8b8

Browse files
authored
[LLM] update llm server dockerfiles (#9940) (#9957)
* [LLM] update llm server dockerfiles * update llm dockerfile
1 parent b79aae3 commit 4a2f8b8

File tree

2 files changed

+16
-14
lines changed

2 files changed

+16
-14
lines changed

llm/server/dockerfiles/Dockerfile_serving_cuda118_cudnn8

Lines changed: 8 additions & 7 deletions
Original file line numberDiff line numberDiff line change
@@ -1,18 +1,19 @@
11
FROM registry.baidubce.com/paddlepaddle/fastdeploy:llm-base-gcc12.3-cuda11.8-cudnn8-nccl2.15.5
22

33
WORKDIR /opt/output/
4-
ENV LD_LIBRARY_PATH="/usr/local/cuda-11.8/compat/:$LD_LIBRARY_PATH"
4+
ENV LD_LIBRARY_PATH="/usr/local/cuda-11.8/lib64:/usr/lib64:/usr/local/cuda-11.8/targets/x86_64-linux/lib/:/opt/nccl-2.15.5-1/build/lib/:/usr/local/nvidia/lib:/usr/local/nvidia/lib64"
55

6-
RUN python3 -m pip install --pre paddlepaddle-gpu -i https://www.paddlepaddle.org.cn/packages/nightly/cu123/ \
7-
&& python3 -m pip install paddlenlp==3.0.0b0 \
8-
&& python3 -m pip install --no-cache-dir sentencepiece pycryptodome tritonclient[all]==2.41.1 \
6+
# 安装 paddlepaddle & paddlenlp & paddlenlp_ops
7+
RUN python3 -m pip install --pre paddlepaddle-gpu -i https://www.paddlepaddle.org.cn/packages/nightly/cu118/ \
8+
&& python3 -m pip install --no-cache-dir --force-reinstall https://paddle-qa.bj.bcebos.com/paddlenlp/wheel/2f85a64edd4aa9911c94ccb5ce53e83ac41ce22b/paddlenlp-3.0.0b3.post20250123-py3-none-any.whl \
99
&& python3 -m pip install --no-cache-dir --force-reinstall https://paddlepaddle-inference-banchmark.bj.bcebos.com/paddlenlp_ops-0.0.0-py3-none-any.whl \
10+
&& python3 -m pip install --no-cache-dir sentencepiece pycryptodome tritonclient[all]==2.41.1 \
1011
&& apt-get clean && rm -rf /var/lib/apt/lists/*
1112

13+
# clone paddle & paddlenlp 源码(代码版本应与上述安装版本对齐)
1214
RUN mkdir -p /opt/source/ && cd /opt/source/ \
1315
&& git clone https://github.com/PaddlePaddle/Paddle.git \
14-
&& git clone https://github.com/PaddlePaddle/PaddleNLP.git \
15-
&& cp -r /opt/source/PaddleNLP/paddlenlp /usr/local/lib/python3.10/dist-packages/ \
16+
&& git clone -b release/3.0-beta4 https://github.com/PaddlePaddle/PaddleNLP.git \
1617
&& python3 -m pip install --no-cache-dir -r PaddleNLP/requirements.txt \
1718
&& python3 -m pip install --no-cache-dir -r PaddleNLP/llm/server/server/requirements.txt
1819

@@ -21,7 +22,7 @@ RUN mkdir -p /opt/output/Serving/llm_model/model/1 \
2122
&& cp /opt/source/PaddleNLP/llm/server/server/scripts/start_server.sh /opt/output/Serving/ \
2223
&& cp /opt/source/PaddleNLP/llm/server/server/scripts/stop_server.sh /opt/output/Serving/
2324

24-
ENV PYTHONPATH="/opt/source/PaddleNLP/llm/server/server"
25+
ENV PYTHONPATH="/opt/source/PaddleNLP/llm/server/server:/opt/source/PaddleNLP"
2526
RUN echo "from server.triton_server import TritonPythonModel" >>/opt/output/Serving/llm_model/model/1/model.py
2627

2728
ENV http_proxy=""

llm/server/dockerfiles/Dockerfile_serving_cuda123_cudnn9 renamed to llm/server/dockerfiles/Dockerfile_serving_cuda124_cudnn9

Lines changed: 8 additions & 7 deletions
Original file line numberDiff line numberDiff line change
@@ -1,18 +1,19 @@
1-
FROM registry.baidubce.com/paddlepaddle/fastdeploy:llm-base-gcc12.3-cuda12.3-cudnn9-nccl2.15.5
1+
FROM registry.baidubce.com/paddlepaddle/fastdeploy:llm-base-gcc12.3-cuda12.4-cudnn9-nccl2.15.5
22

33
WORKDIR /opt/output/
4-
ENV LD_LIBRARY_PATH="/usr/local/cuda-12.3/compat/:$LD_LIBRARY_PATH"
4+
ENV LD_LIBRARY_PATH="/usr/local/cuda-12.4/lib64:/usr/lib64:/usr/local/cuda-12.4/targets/x86_64-linux/lib/:/opt/nccl-2.15.5-1/build/lib/:/usr/local/nvidia/lib:/usr/local/nvidia/lib64"
55

6+
# 安装 paddlepaddle & paddlenlp & paddlenlp_ops
67
RUN python3 -m pip install --pre paddlepaddle-gpu -i https://www.paddlepaddle.org.cn/packages/nightly/cu123/ \
7-
&& python3 -m pip install paddlenlp==3.0.0b0 \
8-
&& python3 -m pip install --no-cache-dir sentencepiece pycryptodome tritonclient[all]==2.41.1 \
8+
&& python3 -m pip install --no-cache-dir --force-reinstall https://paddle-qa.bj.bcebos.com/paddlenlp/wheel/2f85a64edd4aa9911c94ccb5ce53e83ac41ce22b/paddlenlp-3.0.0b3.post20250123-py3-none-any.whl \
99
&& python3 -m pip install --no-cache-dir --force-reinstall https://paddlepaddle-inference-banchmark.bj.bcebos.com/paddlenlp_ops-0.0.0-py3-none-any.whl \
10+
&& python3 -m pip install --no-cache-dir sentencepiece pycryptodome tritonclient[all]==2.41.1 \
1011
&& apt-get clean && rm -rf /var/lib/apt/lists/*
1112

13+
# clone paddle & paddlenlp 源码(代码版本应与上述安装版本对齐)
1214
RUN mkdir -p /opt/source/ && cd /opt/source/ \
1315
&& git clone https://github.com/PaddlePaddle/Paddle.git \
14-
&& git clone https://github.com/PaddlePaddle/PaddleNLP.git \
15-
&& cp -r /opt/source/PaddleNLP/paddlenlp /usr/local/lib/python3.10/dist-packages/ \
16+
&& git clone -b release/3.0-beta4 https://github.com/PaddlePaddle/PaddleNLP.git \
1617
&& python3 -m pip install --no-cache-dir -r PaddleNLP/requirements.txt \
1718
&& python3 -m pip install --no-cache-dir -r PaddleNLP/llm/server/server/requirements.txt
1819

@@ -21,7 +22,7 @@ RUN mkdir -p /opt/output/Serving/llm_model/model/1 \
2122
&& cp /opt/source/PaddleNLP/llm/server/server/scripts/start_server.sh /opt/output/Serving/ \
2223
&& cp /opt/source/PaddleNLP/llm/server/server/scripts/stop_server.sh /opt/output/Serving/
2324

24-
ENV PYTHONPATH="/opt/source/PaddleNLP/llm/server/server"
25+
ENV PYTHONPATH="/opt/source/PaddleNLP/llm/server/server:/opt/source/PaddleNLP"
2526
RUN echo "from server.triton_server import TritonPythonModel" >>/opt/output/Serving/llm_model/model/1/model.py
2627

2728
ENV http_proxy=""

0 commit comments

Comments
 (0)