Skip to content

Commit 599a58f

Browse files
authored
update vllm to latest version for hpu (#713)
* update vllm version for hpu Signed-off-by: Xinyao Wang <[email protected]> * fix bug Signed-off-by: Xinyao Wang <[email protected]> * fix bug Signed-off-by: Xinyao Wang <[email protected]> --------- Signed-off-by: Xinyao Wang <[email protected]>
1 parent ccdd2d0 commit 599a58f

File tree

1 file changed

+12
-15
lines changed

1 file changed

+12
-15
lines changed
Lines changed: 12 additions & 15 deletions
Original file line numberDiff line numberDiff line change
@@ -1,28 +1,25 @@
11
# Copyright (C) 2024 Intel Corporation
22
# SPDX-License-Identifier: Apache-2.0
33

4-
# FROM vault.habana.ai/gaudi-docker/1.16.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.2:latest as hpu
5-
FROM opea/habanalabs:1.16.1-pytorch-installer-2.2.2 as hpu
4+
#FROM vault.habana.ai/gaudi-docker/1.17.0/ubuntu22.04/habanalabs/pytorch-installer-2.3.1:latest
5+
FROM vault.habana.ai/gaudi-docker/1.16.2/ubuntu22.04/habanalabs/pytorch-installer-2.2.2:latest
66

7-
RUN useradd -m -s /bin/bash user && \
8-
mkdir -p /home/user && \
9-
chown -R user /home/user/
10-
ENV LANG=en_US.UTF-8
11-
RUN sed -i 's/#PermitRootLogin prohibit-password/PermitRootLogin yes/' /etc/ssh/sshd_config && \
12-
service ssh restart
13-
USER user
14-
WORKDIR /root
7+
RUN git clone https://github.com/HabanaAI/vllm-fork.git /workspace/vllm
158

16-
RUN pip install --no-cache-dir --upgrade-strategy eager optimum[habana]
9+
# COPY ./ /workspace/vllm
1710

18-
RUN pip install --no-cache-dir -v git+https://github.com/HabanaAI/vllm-fork.git@cf6952d
11+
WORKDIR /workspace/vllm
1912

20-
RUN pip install --no-cache-dir setuptools
13+
RUN pip install -v -r requirements-hpu.txt
2114

2215
ENV no_proxy=localhost,127.0.0.1
16+
ENV PT_HPU_ENABLE_LAZY_COLLECTIVES=true
2317

24-
ENV PT_HPU_LAZY_ACC_PAR_MODE=0
18+
RUN VLLM_TARGET_DEVICE=hpu python3 setup.py install
2519

26-
ENV PT_HPU_ENABLE_LAZY_COLLECTIVES=true
20+
WORKDIR /workspace/
21+
22+
RUN ln -s /workspace/vllm/tests && ln -s /workspace/vllm/examples && ln -s /workspace/vllm/benchmarks
2723

24+
#ENTRYPOINT ["python3", "-m", "vllm.entrypoints.openai.api_server"]
2825
CMD ["/bin/bash"]

0 commit comments

Comments
 (0)