hotfix: update nccl
This commit is contained in:
parent
e7e3aa6cac
commit
a895029424
|
@ -252,7 +252,10 @@ COPY server/Makefile server/Makefile
|
||||||
RUN cd server && \
|
RUN cd server && \
|
||||||
make gen-server && \
|
make gen-server && \
|
||||||
pip install -r requirements_cuda.txt && \
|
pip install -r requirements_cuda.txt && \
|
||||||
pip install ".[bnb, accelerate, quantize, peft, outlines]" --no-cache-dir
|
pip install ".[bnb, accelerate, quantize, peft, outlines]" --no-cache-dir && \
|
||||||
|
pip install nvidia-nccl-cu12==2.22.3
|
||||||
|
|
||||||
|
ENV LD_PRELOAD=/opt/conda/lib/python3.10/site-packages/nvidia/nccl/lib/libnccl.so.2
|
||||||
|
|
||||||
# Deps before the binaries
|
# Deps before the binaries
|
||||||
# The binaries change on every build given we burn the SHA into them
|
# The binaries change on every build given we burn the SHA into them
|
||||||
|
|
|
@ -30,6 +30,7 @@ install: install-cuda
|
||||||
|
|
||||||
install-cuda: install-server install-flash-attention-v2-cuda install-vllm-cuda install-flash-attention install-fbgemm
|
install-cuda: install-server install-flash-attention-v2-cuda install-vllm-cuda install-flash-attention install-fbgemm
|
||||||
pip install -e ".[bnb]"
|
pip install -e ".[bnb]"
|
||||||
|
pip install nvidia-nccl-cu12==2.22.3
|
||||||
|
|
||||||
install-rocm: install-server install-flash-attention-v2-rocm install-vllm-rocm
|
install-rocm: install-server install-flash-attention-v2-rocm install-vllm-rocm
|
||||||
|
|
||||||
|
|
Loading…
Reference in New Issue