|
1 | 1 | # Copyright (C) 2024 Intel Corporation
|
2 | 2 | # SPDX-License-Identifier: Apache-2.0
|
3 | 3 |
|
4 |
| -# FROM vault.habana.ai/gaudi-docker/1.16.1/ubuntu22.04/habanalabs/pytorch-installer-2.2.2:latest as hpu |
5 |
| -FROM opea/habanalabs:1.16.1-pytorch-installer-2.2.2 as hpu |
| 4 | +#FROM vault.habana.ai/gaudi-docker/1.17.0/ubuntu22.04/habanalabs/pytorch-installer-2.3.1:latest |
| 5 | +FROM vault.habana.ai/gaudi-docker/1.16.2/ubuntu22.04/habanalabs/pytorch-installer-2.2.2:latest |
6 | 6 |
|
7 |
| -RUN useradd -m -s /bin/bash user && \ |
8 |
| - mkdir -p /home/user && \ |
9 |
| - chown -R user /home/user/ |
10 |
| -ENV LANG=en_US.UTF-8 |
11 |
| -RUN sed -i 's/#PermitRootLogin prohibit-password/PermitRootLogin yes/' /etc/ssh/sshd_config && \ |
12 |
| - service ssh restart |
13 |
| -USER user |
14 |
| -WORKDIR /root |
| 7 | +RUN git clone https://github.com/HabanaAI/vllm-fork.git /workspace/vllm |
15 | 8 |
|
16 |
| -RUN pip install --no-cache-dir --upgrade-strategy eager optimum[habana] |
| 9 | +# COPY ./ /workspace/vllm |
17 | 10 |
|
18 |
| -RUN pip install --no-cache-dir -v git+https://github.com/HabanaAI/vllm-fork.git@cf6952d |
| 11 | +WORKDIR /workspace/vllm |
19 | 12 |
|
20 |
| -RUN pip install --no-cache-dir setuptools |
| 13 | +RUN pip install -v -r requirements-hpu.txt |
21 | 14 |
|
22 | 15 | ENV no_proxy=localhost,127.0.0.1
|
| 16 | +ENV PT_HPU_ENABLE_LAZY_COLLECTIVES=true |
23 | 17 |
|
24 |
| -ENV PT_HPU_LAZY_ACC_PAR_MODE=0 |
| 18 | +RUN VLLM_TARGET_DEVICE=hpu python3 setup.py install |
25 | 19 |
|
26 |
| -ENV PT_HPU_ENABLE_LAZY_COLLECTIVES=true |
| 20 | +WORKDIR /workspace/ |
| 21 | + |
| 22 | +RUN ln -s /workspace/vllm/tests && ln -s /workspace/vllm/examples && ln -s /workspace/vllm/benchmarks |
27 | 23 |
|
| 24 | +#ENTRYPOINT ["python3", "-m", "vllm.entrypoints.openai.api_server"] |
28 | 25 | CMD ["/bin/bash"]
|
0 commit comments