docked: adjust
This commit is contained in:
parent
92ae8312af
commit
8c3937b4b5
|
@ -7,7 +7,7 @@ FROM nvidia/cuda:11.8.0-devel-ubuntu22.04 as build
|
|||
RUN apt-get update && \
|
||||
apt-get install -y git python3-pip python3-venv wget unzip && \
|
||||
rm -rf /var/lib/apt/lists/*
|
||||
RUN pip3 install --upgrade pip setuptools wheel
|
||||
RUN pip install --upgrade pip setuptools wheel
|
||||
|
||||
RUN git clone https://git.evulid.cc/cyberes/local-llm-server.git /local-llm-server
|
||||
|
||||
|
@ -28,11 +28,8 @@ RUN /venv/bin/pip3 install --upgrade pip setuptools wheel
|
|||
# version for our CUDA install. (VLLM wants 2.0.1)
|
||||
RUN /venv/bin/pip3 install torch==2.0.1 --index-url https://download.pytorch.org/whl/cu118
|
||||
|
||||
WORKDIR /local-llm-server
|
||||
# WORKDIR /local-llm-server
|
||||
|
||||
# We don't need to rebuild VLLM every time we build the container. But if we need
|
||||
# to, uncomment the following line.
|
||||
# ADD "https://www.random.org/cgi-bin/randbyte?nbytes=10&format=h" skipcache
|
||||
RUN /venv/bin/pip install git+https://github.com/vllm-project/vllm
|
||||
|
||||
FROM nvidia/cuda:11.8.0-base-ubuntu22.04 as base
|
||||
|
|
Reference in New Issue