Using the "lockfile". (#2992)
* Using the "lockfile". * Revert dummy modifications. * Lock on python 3.11 * Another attempt. * .. * Bad cache hits. * The good old monkey. * How in the world... * We need the launcher still. * . * .. * Attempt #42 * Don't break all other builds. * Mode max. * Applying to other builds.
This commit is contained in:
parent
c1cf36c0dc
commit
0ef8c8a97a
|
@ -202,8 +202,8 @@ jobs:
|
|||
target: ${{ env.TARGET }}
|
||||
tags: ${{ steps.meta.outputs.tags || steps.meta-pr.outputs.tags }}
|
||||
labels: ${{ steps.meta.outputs.labels || steps.meta-pr.outputs.labels }}
|
||||
cache-from: type=s3,region=us-east-1,bucket=ci-docker-buildx-cache,name=text-generation-inference-cache${{ env.LABEL }},mode=min,access_key_id=${{ secrets.S3_CI_DOCKER_BUILDX_CACHE_ACCESS_KEY_ID }},secret_access_key=${{ secrets.S3_CI_DOCKER_BUILDX_CACHE_SECRET_ACCESS_KEY }},mode=max
|
||||
cache-to: type=s3,region=us-east-1,bucket=ci-docker-buildx-cache,name=text-generation-inference-cache${{ env.LABEL }},mode=min,access_key_id=${{ secrets.S3_CI_DOCKER_BUILDX_CACHE_ACCESS_KEY_ID }},secret_access_key=${{ secrets.S3_CI_DOCKER_BUILDX_CACHE_SECRET_ACCESS_KEY }},mode=max
|
||||
cache-from: type=s3,region=us-east-1,bucket=ci-docker-buildx-cache,name=text-generation-inference-cache${{ env.LABEL }},mode=max,access_key_id=${{ secrets.S3_CI_DOCKER_BUILDX_CACHE_ACCESS_KEY_ID }},secret_access_key=${{ secrets.S3_CI_DOCKER_BUILDX_CACHE_SECRET_ACCESS_KEY }},mode=max
|
||||
cache-to: type=s3,region=us-east-1,bucket=ci-docker-buildx-cache,name=text-generation-inference-cache${{ env.LABEL }},mode=min,access_key_id=${{ secrets.S3_CI_DOCKER_BUILDX_CACHE_ACCESS_KEY_ID }},secret_access_key=${{ secrets.S3_CI_DOCKER_BUILDX_CACHE_SECRET_ACCESS_KEY }},mode=min
|
||||
- name: Final
|
||||
id: final
|
||||
run: |
|
||||
|
|
72
Dockerfile
72
Dockerfile
|
@ -195,50 +195,56 @@ RUN apt-get update && DEBIAN_FRONTEND=noninteractive apt-get install -y --no-ins
|
|||
git \
|
||||
&& rm -rf /var/lib/apt/lists/*
|
||||
|
||||
# Copy conda with PyTorch installed
|
||||
COPY --from=pytorch-install /opt/conda /opt/conda
|
||||
|
||||
# Copy build artifacts from flash attention builder
|
||||
COPY --from=flash-att-builder /usr/src/flash-attention/build/lib.linux-x86_64-cpython-311 /opt/conda/lib/python3.11/site-packages
|
||||
COPY --from=flash-att-builder /usr/src/flash-attention/csrc/layer_norm/build/lib.linux-x86_64-cpython-311 /opt/conda/lib/python3.11/site-packages
|
||||
COPY --from=flash-att-builder /usr/src/flash-attention/csrc/rotary/build/lib.linux-x86_64-cpython-311 /opt/conda/lib/python3.11/site-packages
|
||||
|
||||
# Copy build artifacts from flash attention v2 builder
|
||||
COPY --from=flash-att-v2-builder /opt/conda/lib/python3.11/site-packages/flash_attn_2_cuda.cpython-311-x86_64-linux-gnu.so /opt/conda/lib/python3.11/site-packages
|
||||
|
||||
# Copy build artifacts from custom kernels builder
|
||||
COPY --from=custom-kernels-builder /usr/src/build/lib.linux-x86_64-cpython-311 /opt/conda/lib/python3.11/site-packages
|
||||
# Copy build artifacts from exllama kernels builder
|
||||
COPY --from=exllama-kernels-builder /usr/src/build/lib.linux-x86_64-cpython-311 /opt/conda/lib/python3.11/site-packages
|
||||
# Copy build artifacts from exllamav2 kernels builder
|
||||
COPY --from=exllamav2-kernels-builder /usr/src/exllamav2/build/lib.linux-x86_64-cpython-311 /opt/conda/lib/python3.11/site-packages
|
||||
# Copy build artifacts from awq kernels builder
|
||||
COPY --from=awq-kernels-builder /usr/src/llm-awq/awq/kernels/build/lib.linux-x86_64-cpython-311 /opt/conda/lib/python3.11/site-packages
|
||||
# Copy build artifacts from eetq kernels builder
|
||||
COPY --from=eetq-kernels-builder /usr/src/eetq/build/lib.linux-x86_64-cpython-311 /opt/conda/lib/python3.11/site-packages
|
||||
# Copy build artifacts from lorax punica kernels builder
|
||||
COPY --from=lorax-punica-builder /usr/src/lorax-punica/server/punica_kernels/build/lib.linux-x86_64-cpython-311 /opt/conda/lib/python3.11/site-packages
|
||||
# Copy build artifacts from mamba builder
|
||||
COPY --from=mamba-builder /usr/src/mamba/build/lib.linux-x86_64-cpython-311/ /opt/conda/lib/python3.11/site-packages
|
||||
COPY --from=mamba-builder /usr/src/causal-conv1d/build/lib.linux-x86_64-cpython-311/ /opt/conda/lib/python3.11/site-packages
|
||||
COPY --from=flashinfer-builder /opt/conda/lib/python3.11/site-packages/flashinfer/ /opt/conda/lib/python3.11/site-packages/flashinfer/
|
||||
|
||||
# Install flash-attention dependencies
|
||||
# RUN pip install einops --no-cache-dir
|
||||
|
||||
# Copy conda with PyTorch installed
|
||||
COPY --from=pytorch-install /opt/conda /opt/conda
|
||||
|
||||
# Install server
|
||||
COPY proto proto
|
||||
COPY server server
|
||||
COPY server/Makefile server/Makefile
|
||||
ENV UV_SYSTEM_PYTHON=1
|
||||
RUN cd server && \
|
||||
make gen-server && \
|
||||
python -c "from text_generation_server.pb import generate_pb2" && \
|
||||
pip install -U pip uv && \
|
||||
uv pip install -e ".[attention, bnb, accelerate, compressed-tensors, marlin, moe, quantize, peft, outlines]" --no-cache-dir # && \
|
||||
# uv pip install nvidia-nccl-cu12==2.22.3
|
||||
uv sync --frozen --extra gen --extra attention --extra bnb --extra accelerate --extra compressed-tensors --extra marlin --extra moe --extra quantize --extra peft --extra outlines --no-install-project && \
|
||||
. ./.venv/bin/activate && \
|
||||
make gen-server-raw
|
||||
|
||||
ENV LD_PRELOAD=/opt/conda/lib/python3.11/site-packages/nvidia/nccl/lib/libnccl.so.2
|
||||
RUN cd server && \
|
||||
uv sync --frozen --extra gen --extra attention --extra bnb --extra accelerate --extra compressed-tensors --extra marlin --extra moe --extra quantize --extra peft --extra outlines && \
|
||||
. ./.venv/bin/activate && \
|
||||
pwd && \
|
||||
text-generation-server --help
|
||||
|
||||
# Copy build artifacts from flash attention builder
|
||||
COPY --from=flash-att-builder /usr/src/flash-attention/build/lib.linux-x86_64-cpython-311 /usr/src/server/.venv/lib/python3.11/site-packages
|
||||
COPY --from=flash-att-builder /usr/src/flash-attention/csrc/layer_norm/build/lib.linux-x86_64-cpython-311 /usr/src/server/.venv/lib/python3.11/site-packages
|
||||
COPY --from=flash-att-builder /usr/src/flash-attention/csrc/rotary/build/lib.linux-x86_64-cpython-311 /usr/src/server/.venv/lib/python3.11/site-packages
|
||||
|
||||
# Copy build artifacts from flash attention v2 builder
|
||||
COPY --from=flash-att-v2-builder /opt/conda/lib/python3.11/site-packages/flash_attn_2_cuda.cpython-311-x86_64-linux-gnu.so /usr/src/server/.venv/lib/python3.11/site-packages
|
||||
|
||||
# Copy build artifacts from custom kernels builder
|
||||
COPY --from=custom-kernels-builder /usr/src/build/lib.linux-x86_64-cpython-311 /usr/src/server/.venv/lib/python3.11/site-packages
|
||||
# Copy build artifacts from exllama kernels builder
|
||||
COPY --from=exllama-kernels-builder /usr/src/build/lib.linux-x86_64-cpython-311 /usr/src/server/.venv/lib/python3.11/site-packages
|
||||
# Copy build artifacts from exllamav2 kernels builder
|
||||
COPY --from=exllamav2-kernels-builder /usr/src/exllamav2/build/lib.linux-x86_64-cpython-311 /usr/src/server/.venv/lib/python3.11/site-packages
|
||||
# Copy build artifacts from awq kernels builder
|
||||
COPY --from=awq-kernels-builder /usr/src/llm-awq/awq/kernels/build/lib.linux-x86_64-cpython-311 /usr/src/server/.venv/lib/python3.11/site-packages
|
||||
# Copy build artifacts from eetq kernels builder
|
||||
COPY --from=eetq-kernels-builder /usr/src/eetq/build/lib.linux-x86_64-cpython-311 /usr/src/server/.venv/lib/python3.11/site-packages
|
||||
# Copy build artifacts from lorax punica kernels builder
|
||||
COPY --from=lorax-punica-builder /usr/src/lorax-punica/server/punica_kernels/build/lib.linux-x86_64-cpython-311 /usr/src/server/.venv/lib/python3.11/site-packages
|
||||
# Copy build artifacts from mamba builder
|
||||
COPY --from=mamba-builder /usr/src/mamba/build/lib.linux-x86_64-cpython-311/ /usr/src/server/.venv/lib/python3.11/site-packages
|
||||
COPY --from=mamba-builder /usr/src/causal-conv1d/build/lib.linux-x86_64-cpython-311/ /usr/src/server/.venv/lib/python3.11/site-packages
|
||||
COPY --from=flashinfer-builder /opt/conda/lib/python3.11/site-packages/flashinfer/ /usr/src/server/.venv/lib/python3.11/site-packages/flashinfer/
|
||||
|
||||
|
||||
# ENV LD_PRELOAD=/opt/conda/lib/python3.11/site-packages/nvidia/nccl/lib/libnccl.so.2
|
||||
# Required to find libpython within the rust binaries
|
||||
ENV LD_LIBRARY_PATH="$LD_LIBRARY_PATH:/opt/conda/lib/"
|
||||
# This is needed because exl2 tries to load flash-attn
|
||||
|
|
|
@ -320,9 +320,16 @@ COPY server server
|
|||
COPY server/Makefile server/Makefile
|
||||
ENV UV_SYSTEM_PYTHON=1
|
||||
RUN cd server && \
|
||||
make gen-server && \
|
||||
pip install -U pip uv && \
|
||||
uv pip install -e ".[accelerate, compressed-tensors, peft, outlines]" --no-cache-dir
|
||||
uv sync --frozen --extra gen --extra accelerate --extra compressed-tensors --extra quantize --extra peft --extra outlines --no-install-project && \
|
||||
. ./.venv/bin/activate && \
|
||||
make gen-server-raw
|
||||
|
||||
RUN cd server && \
|
||||
uv sync --frozen --extra gen --extra accelerate --extra compressed-tensors --extra quantize --extra peft --extra outlines && \
|
||||
. ./.venv/bin/activate && \
|
||||
pwd && \
|
||||
text-generation-server --help
|
||||
|
||||
# Install benchmarker
|
||||
COPY --from=builder /usr/src/target/release-opt/text-generation-benchmark /usr/local/bin/text-generation-benchmark
|
||||
|
|
|
@ -215,9 +215,16 @@ COPY server server
|
|||
COPY server/Makefile server/Makefile
|
||||
ENV UV_SYSTEM_PYTHON=1
|
||||
RUN cd server && \
|
||||
make gen-server && \
|
||||
pip install -U pip uv && \
|
||||
uv pip install -e ".[accelerate, compressed-tensors, peft, outlines]" --no-cache-dir
|
||||
uv sync --frozen --extra gen --extra accelerate --extra compressed-tensors --extra quantize --extra peft --extra outlines --no-install-project && \
|
||||
. ./.venv/bin/activate && \
|
||||
make gen-server-raw
|
||||
|
||||
RUN cd server && \
|
||||
uv sync --frozen --extra gen --extra accelerate --extra compressed-tensors --extra quantize --extra peft --extra outlines && \
|
||||
. ./.venv/bin/activate && \
|
||||
pwd && \
|
||||
text-generation-server --help
|
||||
|
||||
# Install benchmarker
|
||||
COPY --from=builder /usr/src/target/release-opt/text-generation-benchmark /usr/local/bin/text-generation-benchmark
|
||||
|
@ -231,5 +238,8 @@ ENV ATTENTION=flashdecoding-ipex
|
|||
ENV PREFIX_CACHING=1
|
||||
ENV PREFILL_CHUNKING=1
|
||||
ENV CUDA_GRAPHS=0
|
||||
ENTRYPOINT ["text-generation-launcher"]
|
||||
COPY ./tgi-entrypoint.sh /tgi-entrypoint.sh
|
||||
RUN chmod +x /tgi-entrypoint.sh
|
||||
|
||||
ENTRYPOINT ["/tgi-entrypoint.sh"]
|
||||
CMD ["--json-output"]
|
||||
|
|
|
@ -16,7 +16,14 @@ unit-tests:
|
|||
gen-server:
|
||||
# Compile protos
|
||||
pip install -U pip uv
|
||||
uv pip install ".[gen]"
|
||||
uv pip install -r requirements_gen.txt
|
||||
mkdir text_generation_server/pb || true
|
||||
python -m grpc_tools.protoc -I../proto/v3 --python_out=text_generation_server/pb \
|
||||
--grpc_python_out=text_generation_server/pb --mypy_out=text_generation_server/pb ../proto/v3/generate.proto
|
||||
find text_generation_server/pb/ -type f -name "*.py" -print0 -exec sed -i -e 's/^\(import.*pb2\)/from . \1/g' {} \;
|
||||
touch text_generation_server/pb/__init__.py
|
||||
|
||||
gen-server-raw:
|
||||
mkdir text_generation_server/pb || true
|
||||
python -m grpc_tools.protoc -I../proto/v3 --python_out=text_generation_server/pb \
|
||||
--grpc_python_out=text_generation_server/pb --mypy_out=text_generation_server/pb ../proto/v3/generate.proto
|
||||
|
@ -35,3 +42,9 @@ install-cuda: install-server install-flash-attention-v2-cuda install-flash-atten
|
|||
uv pip install nvidia-nccl-cu12==2.22.3
|
||||
|
||||
install-rocm: install-server install-flash-attention-v2-rocm install-vllm-rocm
|
||||
|
||||
export-requirements:
|
||||
uv pip compile pyproject.toml --extra gen -o requirements_gen.txt --python-version 3.11
|
||||
uv pip compile pyproject.toml --extra attention --extra bnb --extra accelerate --extra compressed-tensors --extra marlin --extra moe --extra quantize --extra peft --extra outlines -o requirements_cuda.txt --python-version 3.11
|
||||
uv pip compile pyproject.toml --extra accelerate --extra compressed-tensors --extra quantize --extra peft --extra outlines -o requirements_intel.txt --python-version 3.11
|
||||
uv pip compile pyproject.toml --extra accelerate --extra compressed-tensors --extra quantize --extra peft --extra outlines -o requirements_rocm.txt --python-version 3.11
|
||||
|
|
|
@ -30,6 +30,7 @@ dependencies = [
|
|||
"sentencepiece>=0.2.0",
|
||||
"tokenizers>=0.20.3",
|
||||
"typer>=0.15.1",
|
||||
"transformers>=4.48.0"
|
||||
]
|
||||
|
||||
[project.scripts]
|
||||
|
@ -83,5 +84,8 @@ markers = ["private: marks tests as requiring an admin hf token (deselect with '
|
|||
[tool.isort]
|
||||
profile = "black"
|
||||
|
||||
[tool.uv]
|
||||
package = true
|
||||
|
||||
[tool.setuptools.packages.find]
|
||||
include = ["text_generation_server*"]
|
||||
|
|
|
@ -0,0 +1,212 @@
|
|||
# This file was autogenerated by uv via the following command:
|
||||
# uv pip compile pyproject.toml --extra attention --extra bnb -o req.txt
|
||||
attention-kernels @ https://github.com/danieldk/attention-kernels/releases/download/v0.2.0.post2/attention_kernels-0.2.0.post2+cu123torch2.5-cp39-abi3-linux_x86_64.whl
|
||||
# via text-generation-server (pyproject.toml)
|
||||
bitsandbytes==0.45.1
|
||||
# via text-generation-server (pyproject.toml)
|
||||
certifi==2025.1.31
|
||||
# via requests
|
||||
charset-normalizer==3.4.1
|
||||
# via requests
|
||||
click==8.1.8
|
||||
# via typer
|
||||
deprecated==1.2.18
|
||||
# via
|
||||
# opentelemetry-api
|
||||
# opentelemetry-exporter-otlp-proto-grpc
|
||||
# opentelemetry-exporter-otlp-proto-http
|
||||
# opentelemetry-semantic-conventions
|
||||
einops==0.8.0
|
||||
# via text-generation-server (pyproject.toml)
|
||||
filelock==3.17.0
|
||||
# via
|
||||
# huggingface-hub
|
||||
# torch
|
||||
fsspec==2025.2.0
|
||||
# via
|
||||
# huggingface-hub
|
||||
# torch
|
||||
googleapis-common-protos==1.66.0
|
||||
# via
|
||||
# grpcio-status
|
||||
# opentelemetry-exporter-otlp-proto-grpc
|
||||
# opentelemetry-exporter-otlp-proto-http
|
||||
grpc-interceptor==0.15.4
|
||||
# via text-generation-server (pyproject.toml)
|
||||
grpcio==1.70.0
|
||||
# via
|
||||
# text-generation-server (pyproject.toml)
|
||||
# grpc-interceptor
|
||||
# grpcio-reflection
|
||||
# grpcio-status
|
||||
# opentelemetry-exporter-otlp-proto-grpc
|
||||
grpcio-reflection==1.70.0
|
||||
# via text-generation-server (pyproject.toml)
|
||||
grpcio-status==1.70.0
|
||||
# via text-generation-server (pyproject.toml)
|
||||
hf-transfer==0.1.9
|
||||
# via text-generation-server (pyproject.toml)
|
||||
huggingface-hub==0.28.1
|
||||
# via tokenizers
|
||||
idna==3.10
|
||||
# via requests
|
||||
importlib-metadata==8.5.0
|
||||
# via opentelemetry-api
|
||||
jinja2==3.1.5
|
||||
# via torch
|
||||
loguru==0.7.3
|
||||
# via text-generation-server (pyproject.toml)
|
||||
markdown-it-py==3.0.0
|
||||
# via rich
|
||||
markupsafe==3.0.2
|
||||
# via jinja2
|
||||
mdurl==0.1.2
|
||||
# via markdown-it-py
|
||||
mpmath==1.3.0
|
||||
# via sympy
|
||||
networkx==3.4.2
|
||||
# via torch
|
||||
numpy==2.2.2
|
||||
# via
|
||||
# text-generation-server (pyproject.toml)
|
||||
# bitsandbytes
|
||||
# scipy
|
||||
nvidia-cublas-cu12==12.4.5.8
|
||||
# via
|
||||
# nvidia-cudnn-cu12
|
||||
# nvidia-cusolver-cu12
|
||||
# torch
|
||||
nvidia-cuda-cupti-cu12==12.4.127
|
||||
# via torch
|
||||
nvidia-cuda-nvrtc-cu12==12.4.127
|
||||
# via torch
|
||||
nvidia-cuda-runtime-cu12==12.4.127
|
||||
# via torch
|
||||
nvidia-cudnn-cu12==9.1.0.70
|
||||
# via torch
|
||||
nvidia-cufft-cu12==11.2.1.3
|
||||
# via torch
|
||||
nvidia-curand-cu12==10.3.5.147
|
||||
# via torch
|
||||
nvidia-cusolver-cu12==11.6.1.9
|
||||
# via torch
|
||||
nvidia-cusparse-cu12==12.3.1.170
|
||||
# via
|
||||
# nvidia-cusolver-cu12
|
||||
# torch
|
||||
nvidia-cusparselt-cu12==0.6.2
|
||||
# via torch
|
||||
nvidia-nccl-cu12==2.21.5
|
||||
# via torch
|
||||
nvidia-nvjitlink-cu12==12.4.127
|
||||
# via
|
||||
# nvidia-cusolver-cu12
|
||||
# nvidia-cusparse-cu12
|
||||
# torch
|
||||
nvidia-nvtx-cu12==12.4.127
|
||||
# via torch
|
||||
opentelemetry-api==1.30.0
|
||||
# via
|
||||
# text-generation-server (pyproject.toml)
|
||||
# opentelemetry-exporter-otlp-proto-grpc
|
||||
# opentelemetry-exporter-otlp-proto-http
|
||||
# opentelemetry-instrumentation
|
||||
# opentelemetry-instrumentation-grpc
|
||||
# opentelemetry-sdk
|
||||
# opentelemetry-semantic-conventions
|
||||
opentelemetry-exporter-otlp==1.30.0
|
||||
# via text-generation-server (pyproject.toml)
|
||||
opentelemetry-exporter-otlp-proto-common==1.30.0
|
||||
# via
|
||||
# opentelemetry-exporter-otlp-proto-grpc
|
||||
# opentelemetry-exporter-otlp-proto-http
|
||||
opentelemetry-exporter-otlp-proto-grpc==1.30.0
|
||||
# via opentelemetry-exporter-otlp
|
||||
opentelemetry-exporter-otlp-proto-http==1.30.0
|
||||
# via opentelemetry-exporter-otlp
|
||||
opentelemetry-instrumentation==0.51b0
|
||||
# via opentelemetry-instrumentation-grpc
|
||||
opentelemetry-instrumentation-grpc==0.51b0
|
||||
# via text-generation-server (pyproject.toml)
|
||||
opentelemetry-proto==1.30.0
|
||||
# via
|
||||
# opentelemetry-exporter-otlp-proto-common
|
||||
# opentelemetry-exporter-otlp-proto-grpc
|
||||
# opentelemetry-exporter-otlp-proto-http
|
||||
opentelemetry-sdk==1.30.0
|
||||
# via
|
||||
# opentelemetry-exporter-otlp-proto-grpc
|
||||
# opentelemetry-exporter-otlp-proto-http
|
||||
opentelemetry-semantic-conventions==0.51b0
|
||||
# via
|
||||
# opentelemetry-instrumentation
|
||||
# opentelemetry-instrumentation-grpc
|
||||
# opentelemetry-sdk
|
||||
packaging==24.2
|
||||
# via
|
||||
# huggingface-hub
|
||||
# opentelemetry-instrumentation
|
||||
pillow==11.1.0
|
||||
# via text-generation-server (pyproject.toml)
|
||||
prometheus-client==0.21.1
|
||||
# via text-generation-server (pyproject.toml)
|
||||
protobuf==5.29.3
|
||||
# via
|
||||
# text-generation-server (pyproject.toml)
|
||||
# googleapis-common-protos
|
||||
# grpcio-reflection
|
||||
# grpcio-status
|
||||
# opentelemetry-proto
|
||||
py-cpuinfo==9.0.0
|
||||
# via text-generation-server (pyproject.toml)
|
||||
pygments==2.19.1
|
||||
# via rich
|
||||
pyyaml==6.0.2
|
||||
# via huggingface-hub
|
||||
requests==2.32.3
|
||||
# via
|
||||
# huggingface-hub
|
||||
# opentelemetry-exporter-otlp-proto-http
|
||||
rich==13.9.4
|
||||
# via
|
||||
# text-generation-server (pyproject.toml)
|
||||
# typer
|
||||
safetensors==0.5.2
|
||||
# via text-generation-server (pyproject.toml)
|
||||
scipy==1.15.1
|
||||
# via text-generation-server (pyproject.toml)
|
||||
sentencepiece==0.2.0
|
||||
# via text-generation-server (pyproject.toml)
|
||||
setuptools==75.8.0
|
||||
# via torch
|
||||
shellingham==1.5.4
|
||||
# via typer
|
||||
sympy==1.13.1
|
||||
# via torch
|
||||
tokenizers==0.21.0
|
||||
# via text-generation-server (pyproject.toml)
|
||||
torch==2.6.0
|
||||
# via
|
||||
# attention-kernels
|
||||
# bitsandbytes
|
||||
tqdm==4.67.1
|
||||
# via huggingface-hub
|
||||
triton==3.2.0
|
||||
# via torch
|
||||
typer==0.15.1
|
||||
# via text-generation-server (pyproject.toml)
|
||||
typing-extensions==4.12.2
|
||||
# via
|
||||
# huggingface-hub
|
||||
# opentelemetry-sdk
|
||||
# torch
|
||||
# typer
|
||||
urllib3==2.3.0
|
||||
# via requests
|
||||
wrapt==1.17.2
|
||||
# via
|
||||
# deprecated
|
||||
# opentelemetry-instrumentation
|
||||
# opentelemetry-instrumentation-grpc
|
||||
zipp==3.21.0
|
||||
# via importlib-metadata
|
|
@ -1,55 +1,384 @@
|
|||
certifi==2024.8.30 ; python_version >= "3.9" and python_version < "3.13"
|
||||
charset-normalizer==3.4.0 ; python_version >= "3.9" and python_version < "3.13"
|
||||
click==8.1.7 ; python_version >= "3.9" and python_version < "3.13"
|
||||
colorama==0.4.6 ; python_version >= "3.9" and python_version < "3.13" and (sys_platform == "win32" or platform_system == "Windows")
|
||||
deprecated==1.2.14 ; python_version >= "3.9" and python_version < "3.13"
|
||||
einops==0.8.0 ; python_version >= "3.9" and python_version < "3.13"
|
||||
filelock==3.16.1 ; python_version >= "3.9" and python_version < "3.13"
|
||||
fsspec==2024.6.1 ; python_version >= "3.9" and python_version < "3.13"
|
||||
googleapis-common-protos==1.65.0 ; python_version >= "3.9" and python_version < "3.13"
|
||||
grpc-interceptor==0.15.4 ; python_version >= "3.9" and python_version < "3.13"
|
||||
grpcio-reflection==1.62.3 ; python_version >= "3.9" and python_version < "3.13"
|
||||
grpcio-status==1.62.3 ; python_version >= "3.9" and python_version < "3.13"
|
||||
grpcio==1.68.0 ; python_version >= "3.9" and python_version < "3.13"
|
||||
hf-transfer==0.1.8 ; python_version >= "3.9" and python_version < "3.13"
|
||||
huggingface-hub==0.23.5 ; python_version >= "3.9" and python_version < "3.13"
|
||||
idna==3.10 ; python_version >= "3.9" and python_version < "3.13"
|
||||
importlib-metadata==7.1.0 ; python_version >= "3.9" and python_version < "3.13"
|
||||
loguru==0.7.2 ; python_version >= "3.9" and python_version < "3.13"
|
||||
markdown-it-py==3.0.0 ; python_version >= "3.9" and python_version < "3.13"
|
||||
mdurl==0.1.2 ; python_version >= "3.9" and python_version < "3.13"
|
||||
numpy==1.26.4 ; python_version >= "3.9" and python_version < "3.13"
|
||||
opentelemetry-api==1.27.0 ; python_version >= "3.9" and python_version < "3.13"
|
||||
opentelemetry-exporter-otlp-proto-common==1.27.0 ; python_version >= "3.9" and python_version < "3.13"
|
||||
opentelemetry-exporter-otlp-proto-grpc==1.27.0 ; python_version >= "3.9" and python_version < "3.13"
|
||||
opentelemetry-exporter-otlp-proto-http==1.27.0 ; python_version >= "3.9" and python_version < "3.13"
|
||||
opentelemetry-exporter-otlp==1.27.0 ; python_version >= "3.9" and python_version < "3.13"
|
||||
opentelemetry-instrumentation-grpc==0.48b0 ; python_version >= "3.9" and python_version < "3.13"
|
||||
opentelemetry-instrumentation==0.48b0 ; python_version >= "3.9" and python_version < "3.13"
|
||||
opentelemetry-proto==1.27.0 ; python_version >= "3.9" and python_version < "3.13"
|
||||
opentelemetry-sdk==1.27.0 ; python_version >= "3.9" and python_version < "3.13"
|
||||
opentelemetry-semantic-conventions==0.48b0 ; python_version >= "3.9" and python_version < "3.13"
|
||||
packaging==24.1 ; python_version >= "3.9" and python_version < "3.13"
|
||||
pillow==11.0.0 ; python_version >= "3.9" and python_version < "3.13"
|
||||
prometheus-client==0.20.0 ; python_version >= "3.9" and python_version < "3.13"
|
||||
protobuf==4.25.5 ; python_version >= "3.9" and python_version < "3.13"
|
||||
py-cpuinfo==9.0.0 ; python_version >= "3.9" and python_version < "3.13"
|
||||
pygments==2.18.0 ; python_version >= "3.9" and python_version < "3.13"
|
||||
pyyaml==6.0.2 ; python_version >= "3.9" and python_version < "3.13"
|
||||
regex==2024.9.11 ; python_version >= "3.9" and python_version < "3.13"
|
||||
requests==2.32.3 ; python_version >= "3.9" and python_version < "3.13"
|
||||
rich==13.9.4 ; python_version >= "3.9" and python_version < "3.13"
|
||||
safetensors==0.4.5 ; python_version >= "3.9" and python_version < "3.13"
|
||||
scipy==1.13.1 ; python_version >= "3.9" and python_version < "3.13"
|
||||
sentencepiece==0.2.0 ; python_version >= "3.9" and python_version < "3.13"
|
||||
setuptools==75.2.0 ; python_version >= "3.9" and python_version < "3.13"
|
||||
shellingham==1.5.4 ; python_version >= "3.9" and python_version < "3.13"
|
||||
tokenizers==0.20.3 ; python_version >= "3.9" and python_version < "3.13"
|
||||
tqdm==4.66.5 ; python_version >= "3.9" and python_version < "3.13"
|
||||
transformers==4.46.3 ; python_version >= "3.9" and python_version < "3.13"
|
||||
typer==0.12.5 ; python_version >= "3.9" and python_version < "3.13"
|
||||
typing-extensions==4.12.2 ; python_version >= "3.9" and python_version < "3.13"
|
||||
urllib3==2.2.3 ; python_version >= "3.9" and python_version < "3.13"
|
||||
win32-setctime==1.1.0 ; python_version >= "3.9" and python_version < "3.13" and sys_platform == "win32"
|
||||
wrapt==1.16.0 ; python_version >= "3.9" and python_version < "3.13"
|
||||
zipp==3.20.2 ; python_version >= "3.9" and python_version < "3.13"
|
||||
# This file was autogenerated by uv via the following command:
|
||||
# uv pip compile pyproject.toml --extra attention --extra bnb --extra accelerate --extra compressed-tensors --extra marlin --extra moe --extra quantize --extra peft --extra outlines -o requirements_cuda.txt --python-version 3.11
|
||||
accelerate==1.3.0
|
||||
# via
|
||||
# text-generation-server (pyproject.toml)
|
||||
# peft
|
||||
aiohappyeyeballs==2.4.4
|
||||
# via aiohttp
|
||||
aiohttp==3.11.11
|
||||
# via
|
||||
# datasets
|
||||
# fsspec
|
||||
aiosignal==1.3.2
|
||||
# via aiohttp
|
||||
airportsdata==20241001
|
||||
# via outlines
|
||||
annotated-types==0.7.0
|
||||
# via pydantic
|
||||
attention-kernels @ https://github.com/danieldk/attention-kernels/releases/download/v0.2.0.post2/attention_kernels-0.2.0.post2+cu123torch2.5-cp39-abi3-linux_x86_64.whl
|
||||
# via text-generation-server (pyproject.toml)
|
||||
attrs==25.1.0
|
||||
# via
|
||||
# aiohttp
|
||||
# jsonschema
|
||||
# referencing
|
||||
bitsandbytes==0.45.1
|
||||
# via text-generation-server (pyproject.toml)
|
||||
certifi==2024.8.30
|
||||
# via requests
|
||||
charset-normalizer==3.4.0
|
||||
# via requests
|
||||
click==8.1.7
|
||||
# via typer
|
||||
cloudpickle==3.1.1
|
||||
# via outlines
|
||||
compressed-tensors==0.9.1
|
||||
# via text-generation-server (pyproject.toml)
|
||||
datasets==2.21.0
|
||||
# via text-generation-server (pyproject.toml)
|
||||
deprecated==1.2.14
|
||||
# via
|
||||
# opentelemetry-api
|
||||
# opentelemetry-exporter-otlp-proto-grpc
|
||||
# opentelemetry-exporter-otlp-proto-http
|
||||
# opentelemetry-semantic-conventions
|
||||
dill==0.3.8
|
||||
# via
|
||||
# datasets
|
||||
# multiprocess
|
||||
diskcache==5.6.3
|
||||
# via outlines
|
||||
einops==0.8.0
|
||||
# via text-generation-server (pyproject.toml)
|
||||
filelock==3.16.1
|
||||
# via
|
||||
# datasets
|
||||
# huggingface-hub
|
||||
# torch
|
||||
# transformers
|
||||
frozenlist==1.5.0
|
||||
# via
|
||||
# aiohttp
|
||||
# aiosignal
|
||||
fsspec==2024.6.1
|
||||
# via
|
||||
# datasets
|
||||
# huggingface-hub
|
||||
# torch
|
||||
genson==1.3.0
|
||||
# via outlines
|
||||
googleapis-common-protos==1.65.0
|
||||
# via
|
||||
# grpcio-status
|
||||
# opentelemetry-exporter-otlp-proto-grpc
|
||||
# opentelemetry-exporter-otlp-proto-http
|
||||
grpc-interceptor==0.15.4
|
||||
# via text-generation-server (pyproject.toml)
|
||||
grpcio==1.68.0
|
||||
# via
|
||||
# text-generation-server (pyproject.toml)
|
||||
# grpc-interceptor
|
||||
# grpcio-reflection
|
||||
# grpcio-status
|
||||
# opentelemetry-exporter-otlp-proto-grpc
|
||||
grpcio-reflection==1.68.0
|
||||
# via text-generation-server (pyproject.toml)
|
||||
grpcio-status==1.68.0
|
||||
# via text-generation-server (pyproject.toml)
|
||||
hf-transfer==0.1.8
|
||||
# via text-generation-server (pyproject.toml)
|
||||
huggingface-hub==0.28.1
|
||||
# via
|
||||
# accelerate
|
||||
# datasets
|
||||
# peft
|
||||
# tokenizers
|
||||
# transformers
|
||||
idna==3.10
|
||||
# via
|
||||
# requests
|
||||
# yarl
|
||||
importlib-metadata==7.1.0
|
||||
# via opentelemetry-api
|
||||
interegular==0.3.3
|
||||
# via
|
||||
# outlines
|
||||
# outlines-core
|
||||
jinja2==3.1.5
|
||||
# via
|
||||
# outlines
|
||||
# torch
|
||||
jsonschema==4.23.0
|
||||
# via
|
||||
# outlines
|
||||
# outlines-core
|
||||
jsonschema-specifications==2024.10.1
|
||||
# via jsonschema
|
||||
lark==1.2.2
|
||||
# via outlines
|
||||
loguru==0.7.3
|
||||
# via text-generation-server (pyproject.toml)
|
||||
markdown-it-py==3.0.0
|
||||
# via rich
|
||||
markupsafe==3.0.2
|
||||
# via jinja2
|
||||
marlin-kernels @ https://github.com/danieldk/marlin-kernels/releases/download/v0.3.7/marlin_kernels-0.3.7+cu123torch2.5-cp311-cp311-linux_x86_64.whl
|
||||
# via text-generation-server (pyproject.toml)
|
||||
mdurl==0.1.2
|
||||
# via markdown-it-py
|
||||
moe-kernels @ https://github.com/danieldk/moe-kernels/releases/download/v0.8.2/moe_kernels-0.8.2+cu123torch2.5-cp39-abi3-linux_x86_64.whl
|
||||
# via text-generation-server (pyproject.toml)
|
||||
mpmath==1.3.0
|
||||
# via sympy
|
||||
multidict==6.1.0
|
||||
# via
|
||||
# aiohttp
|
||||
# yarl
|
||||
multiprocess==0.70.16
|
||||
# via datasets
|
||||
nest-asyncio==1.6.0
|
||||
# via outlines
|
||||
networkx==3.4.2
|
||||
# via torch
|
||||
numpy==1.26.4
|
||||
# via
|
||||
# text-generation-server (pyproject.toml)
|
||||
# accelerate
|
||||
# bitsandbytes
|
||||
# datasets
|
||||
# outlines
|
||||
# pandas
|
||||
# peft
|
||||
# scipy
|
||||
# transformers
|
||||
nvidia-cublas-cu12==12.4.5.8
|
||||
# via
|
||||
# nvidia-cudnn-cu12
|
||||
# nvidia-cusolver-cu12
|
||||
# torch
|
||||
nvidia-cuda-cupti-cu12==12.4.127
|
||||
# via torch
|
||||
nvidia-cuda-nvrtc-cu12==12.4.127
|
||||
# via torch
|
||||
nvidia-cuda-runtime-cu12==12.4.127
|
||||
# via torch
|
||||
nvidia-cudnn-cu12==9.1.0.70
|
||||
# via torch
|
||||
nvidia-cufft-cu12==11.2.1.3
|
||||
# via torch
|
||||
nvidia-curand-cu12==10.3.5.147
|
||||
# via torch
|
||||
nvidia-cusolver-cu12==11.6.1.9
|
||||
# via torch
|
||||
nvidia-cusparse-cu12==12.3.1.170
|
||||
# via
|
||||
# nvidia-cusolver-cu12
|
||||
# torch
|
||||
nvidia-cusparselt-cu12==0.6.2
|
||||
# via torch
|
||||
nvidia-ml-py==12.570.86
|
||||
# via moe-kernels
|
||||
nvidia-nccl-cu12==2.21.5
|
||||
# via torch
|
||||
nvidia-nvjitlink-cu12==12.4.127
|
||||
# via
|
||||
# nvidia-cusolver-cu12
|
||||
# nvidia-cusparse-cu12
|
||||
# torch
|
||||
nvidia-nvtx-cu12==12.4.127
|
||||
# via torch
|
||||
opentelemetry-api==1.30.0
|
||||
# via
|
||||
# text-generation-server (pyproject.toml)
|
||||
# opentelemetry-exporter-otlp-proto-grpc
|
||||
# opentelemetry-exporter-otlp-proto-http
|
||||
# opentelemetry-instrumentation
|
||||
# opentelemetry-instrumentation-grpc
|
||||
# opentelemetry-sdk
|
||||
# opentelemetry-semantic-conventions
|
||||
opentelemetry-exporter-otlp==1.30.0
|
||||
# via text-generation-server (pyproject.toml)
|
||||
opentelemetry-exporter-otlp-proto-common==1.30.0
|
||||
# via
|
||||
# opentelemetry-exporter-otlp-proto-grpc
|
||||
# opentelemetry-exporter-otlp-proto-http
|
||||
opentelemetry-exporter-otlp-proto-grpc==1.30.0
|
||||
# via opentelemetry-exporter-otlp
|
||||
opentelemetry-exporter-otlp-proto-http==1.30.0
|
||||
# via opentelemetry-exporter-otlp
|
||||
opentelemetry-instrumentation==0.51b0
|
||||
# via opentelemetry-instrumentation-grpc
|
||||
opentelemetry-instrumentation-grpc==0.51b0
|
||||
# via text-generation-server (pyproject.toml)
|
||||
opentelemetry-proto==1.30.0
|
||||
# via
|
||||
# opentelemetry-exporter-otlp-proto-common
|
||||
# opentelemetry-exporter-otlp-proto-grpc
|
||||
# opentelemetry-exporter-otlp-proto-http
|
||||
opentelemetry-sdk==1.30.0
|
||||
# via
|
||||
# opentelemetry-exporter-otlp-proto-grpc
|
||||
# opentelemetry-exporter-otlp-proto-http
|
||||
opentelemetry-semantic-conventions==0.51b0
|
||||
# via
|
||||
# opentelemetry-instrumentation
|
||||
# opentelemetry-instrumentation-grpc
|
||||
# opentelemetry-sdk
|
||||
outlines==0.1.14
|
||||
# via text-generation-server (pyproject.toml)
|
||||
outlines-core==0.1.26
|
||||
# via outlines
|
||||
packaging==24.1
|
||||
# via
|
||||
# accelerate
|
||||
# datasets
|
||||
# huggingface-hub
|
||||
# opentelemetry-instrumentation
|
||||
# peft
|
||||
# transformers
|
||||
pandas==2.2.3
|
||||
# via datasets
|
||||
peft==0.14.0
|
||||
# via text-generation-server (pyproject.toml)
|
||||
pillow==11.1.0
|
||||
# via text-generation-server (pyproject.toml)
|
||||
prometheus-client==0.21.1
|
||||
# via text-generation-server (pyproject.toml)
|
||||
propcache==0.2.1
|
||||
# via
|
||||
# aiohttp
|
||||
# yarl
|
||||
protobuf==5.29.3
|
||||
# via
|
||||
# text-generation-server (pyproject.toml)
|
||||
# googleapis-common-protos
|
||||
# grpcio-reflection
|
||||
# grpcio-status
|
||||
# opentelemetry-proto
|
||||
psutil==6.1.1
|
||||
# via
|
||||
# accelerate
|
||||
# peft
|
||||
py-cpuinfo==9.0.0
|
||||
# via text-generation-server (pyproject.toml)
|
||||
pyarrow==19.0.0
|
||||
# via datasets
|
||||
pycountry==24.6.1
|
||||
# via outlines
|
||||
pydantic==2.10.6
|
||||
# via
|
||||
# compressed-tensors
|
||||
# outlines
|
||||
pydantic-core==2.27.2
|
||||
# via pydantic
|
||||
pygments==2.18.0
|
||||
# via rich
|
||||
python-dateutil==2.9.0.post0
|
||||
# via pandas
|
||||
pytz==2025.1
|
||||
# via pandas
|
||||
pyyaml==6.0.2
|
||||
# via
|
||||
# accelerate
|
||||
# datasets
|
||||
# huggingface-hub
|
||||
# peft
|
||||
# transformers
|
||||
referencing==0.36.2
|
||||
# via
|
||||
# jsonschema
|
||||
# jsonschema-specifications
|
||||
# outlines
|
||||
regex==2024.9.11
|
||||
# via transformers
|
||||
requests==2.32.3
|
||||
# via
|
||||
# datasets
|
||||
# huggingface-hub
|
||||
# opentelemetry-exporter-otlp-proto-http
|
||||
# outlines
|
||||
# transformers
|
||||
rich==13.9.4
|
||||
# via
|
||||
# text-generation-server (pyproject.toml)
|
||||
# typer
|
||||
rpds-py==0.22.3
|
||||
# via
|
||||
# jsonschema
|
||||
# referencing
|
||||
safetensors==0.4.5
|
||||
# via
|
||||
# text-generation-server (pyproject.toml)
|
||||
# accelerate
|
||||
# peft
|
||||
# transformers
|
||||
scipy==1.13.1
|
||||
# via text-generation-server (pyproject.toml)
|
||||
sentencepiece==0.2.0
|
||||
# via text-generation-server (pyproject.toml)
|
||||
shellingham==1.5.4
|
||||
# via typer
|
||||
six==1.17.0
|
||||
# via python-dateutil
|
||||
sympy==1.13.1
|
||||
# via torch
|
||||
texttable==1.7.0
|
||||
# via text-generation-server (pyproject.toml)
|
||||
tokenizers==0.21.0
|
||||
# via
|
||||
# text-generation-server (pyproject.toml)
|
||||
# transformers
|
||||
torch==2.6.0
|
||||
# via
|
||||
# accelerate
|
||||
# attention-kernels
|
||||
# bitsandbytes
|
||||
# compressed-tensors
|
||||
# marlin-kernels
|
||||
# moe-kernels
|
||||
# outlines
|
||||
# peft
|
||||
tqdm==4.66.5
|
||||
# via
|
||||
# datasets
|
||||
# huggingface-hub
|
||||
# outlines
|
||||
# peft
|
||||
# transformers
|
||||
transformers==4.48.2
|
||||
# via
|
||||
# text-generation-server (pyproject.toml)
|
||||
# compressed-tensors
|
||||
# peft
|
||||
triton==3.2.0
|
||||
# via
|
||||
# moe-kernels
|
||||
# torch
|
||||
typer==0.15.1
|
||||
# via text-generation-server (pyproject.toml)
|
||||
typing-extensions==4.12.2
|
||||
# via
|
||||
# huggingface-hub
|
||||
# opentelemetry-sdk
|
||||
# outlines
|
||||
# pydantic
|
||||
# pydantic-core
|
||||
# referencing
|
||||
# torch
|
||||
# typer
|
||||
tzdata==2025.1
|
||||
# via pandas
|
||||
urllib3==2.2.3
|
||||
# via requests
|
||||
wrapt==1.16.0
|
||||
# via
|
||||
# deprecated
|
||||
# opentelemetry-instrumentation
|
||||
# opentelemetry-instrumentation-grpc
|
||||
xxhash==3.5.0
|
||||
# via datasets
|
||||
yarl==1.18.3
|
||||
# via aiohttp
|
||||
zipp==3.20.2
|
||||
# via importlib-metadata
|
||||
|
|
|
@ -0,0 +1,180 @@
|
|||
# This file was autogenerated by uv via the following command:
|
||||
# uv pip compile pyproject.toml --extra gen -o requirements_gen.txt --python-version 3.11
|
||||
certifi==2025.1.31
|
||||
# via requests
|
||||
charset-normalizer==3.4.1
|
||||
# via requests
|
||||
click==8.1.8
|
||||
# via typer
|
||||
deprecated==1.2.18
|
||||
# via
|
||||
# opentelemetry-api
|
||||
# opentelemetry-exporter-otlp-proto-grpc
|
||||
# opentelemetry-exporter-otlp-proto-http
|
||||
# opentelemetry-semantic-conventions
|
||||
einops==0.8.0
|
||||
# via text-generation-server (pyproject.toml)
|
||||
filelock==3.17.0
|
||||
# via
|
||||
# huggingface-hub
|
||||
# transformers
|
||||
fsspec==2025.2.0
|
||||
# via huggingface-hub
|
||||
googleapis-common-protos==1.66.0
|
||||
# via
|
||||
# grpcio-status
|
||||
# opentelemetry-exporter-otlp-proto-grpc
|
||||
# opentelemetry-exporter-otlp-proto-http
|
||||
grpc-interceptor==0.15.4
|
||||
# via text-generation-server (pyproject.toml)
|
||||
grpcio==1.70.0
|
||||
# via
|
||||
# text-generation-server (pyproject.toml)
|
||||
# grpc-interceptor
|
||||
# grpcio-reflection
|
||||
# grpcio-status
|
||||
# grpcio-tools
|
||||
# opentelemetry-exporter-otlp-proto-grpc
|
||||
grpcio-reflection==1.70.0
|
||||
# via text-generation-server (pyproject.toml)
|
||||
grpcio-status==1.70.0
|
||||
# via text-generation-server (pyproject.toml)
|
||||
grpcio-tools==1.70.0
|
||||
# via text-generation-server (pyproject.toml)
|
||||
hf-transfer==0.1.9
|
||||
# via text-generation-server (pyproject.toml)
|
||||
huggingface-hub==0.28.1
|
||||
# via
|
||||
# tokenizers
|
||||
# transformers
|
||||
idna==3.10
|
||||
# via requests
|
||||
importlib-metadata==8.5.0
|
||||
# via opentelemetry-api
|
||||
loguru==0.7.3
|
||||
# via text-generation-server (pyproject.toml)
|
||||
markdown-it-py==3.0.0
|
||||
# via rich
|
||||
mdurl==0.1.2
|
||||
# via markdown-it-py
|
||||
mypy-protobuf==3.6.0
|
||||
# via text-generation-server (pyproject.toml)
|
||||
numpy==2.2.2
|
||||
# via
|
||||
# text-generation-server (pyproject.toml)
|
||||
# scipy
|
||||
# transformers
|
||||
opentelemetry-api==1.30.0
|
||||
# via
|
||||
# text-generation-server (pyproject.toml)
|
||||
# opentelemetry-exporter-otlp-proto-grpc
|
||||
# opentelemetry-exporter-otlp-proto-http
|
||||
# opentelemetry-instrumentation
|
||||
# opentelemetry-instrumentation-grpc
|
||||
# opentelemetry-sdk
|
||||
# opentelemetry-semantic-conventions
|
||||
opentelemetry-exporter-otlp==1.30.0
|
||||
# via text-generation-server (pyproject.toml)
|
||||
opentelemetry-exporter-otlp-proto-common==1.30.0
|
||||
# via
|
||||
# opentelemetry-exporter-otlp-proto-grpc
|
||||
# opentelemetry-exporter-otlp-proto-http
|
||||
opentelemetry-exporter-otlp-proto-grpc==1.30.0
|
||||
# via opentelemetry-exporter-otlp
|
||||
opentelemetry-exporter-otlp-proto-http==1.30.0
|
||||
# via opentelemetry-exporter-otlp
|
||||
opentelemetry-instrumentation==0.51b0
|
||||
# via opentelemetry-instrumentation-grpc
|
||||
opentelemetry-instrumentation-grpc==0.51b0
|
||||
# via text-generation-server (pyproject.toml)
|
||||
opentelemetry-proto==1.30.0
|
||||
# via
|
||||
# opentelemetry-exporter-otlp-proto-common
|
||||
# opentelemetry-exporter-otlp-proto-grpc
|
||||
# opentelemetry-exporter-otlp-proto-http
|
||||
opentelemetry-sdk==1.30.0
|
||||
# via
|
||||
# opentelemetry-exporter-otlp-proto-grpc
|
||||
# opentelemetry-exporter-otlp-proto-http
|
||||
opentelemetry-semantic-conventions==0.51b0
|
||||
# via
|
||||
# opentelemetry-instrumentation
|
||||
# opentelemetry-instrumentation-grpc
|
||||
# opentelemetry-sdk
|
||||
packaging==24.2
|
||||
# via
|
||||
# huggingface-hub
|
||||
# opentelemetry-instrumentation
|
||||
# transformers
|
||||
pillow==11.1.0
|
||||
# via text-generation-server (pyproject.toml)
|
||||
prometheus-client==0.21.1
|
||||
# via text-generation-server (pyproject.toml)
|
||||
protobuf==5.29.3
|
||||
# via
|
||||
# text-generation-server (pyproject.toml)
|
||||
# googleapis-common-protos
|
||||
# grpcio-reflection
|
||||
# grpcio-status
|
||||
# grpcio-tools
|
||||
# mypy-protobuf
|
||||
# opentelemetry-proto
|
||||
py-cpuinfo==9.0.0
|
||||
# via text-generation-server (pyproject.toml)
|
||||
pygments==2.19.1
|
||||
# via rich
|
||||
pyyaml==6.0.2
|
||||
# via
|
||||
# huggingface-hub
|
||||
# transformers
|
||||
regex==2024.11.6
|
||||
# via transformers
|
||||
requests==2.32.3
|
||||
# via
|
||||
# huggingface-hub
|
||||
# opentelemetry-exporter-otlp-proto-http
|
||||
# transformers
|
||||
rich==13.9.4
|
||||
# via
|
||||
# text-generation-server (pyproject.toml)
|
||||
# typer
|
||||
safetensors==0.5.2
|
||||
# via
|
||||
# text-generation-server (pyproject.toml)
|
||||
# transformers
|
||||
scipy==1.15.1
|
||||
# via text-generation-server (pyproject.toml)
|
||||
sentencepiece==0.2.0
|
||||
# via text-generation-server (pyproject.toml)
|
||||
setuptools==75.8.0
|
||||
# via grpcio-tools
|
||||
shellingham==1.5.4
|
||||
# via typer
|
||||
tokenizers==0.21.0
|
||||
# via
|
||||
# text-generation-server (pyproject.toml)
|
||||
# transformers
|
||||
tqdm==4.67.1
|
||||
# via
|
||||
# huggingface-hub
|
||||
# transformers
|
||||
transformers==4.48.2
|
||||
# via text-generation-server (pyproject.toml)
|
||||
typer==0.15.1
|
||||
# via text-generation-server (pyproject.toml)
|
||||
types-protobuf==5.29.1.20241207
|
||||
# via mypy-protobuf
|
||||
typing-extensions==4.12.2
|
||||
# via
|
||||
# huggingface-hub
|
||||
# opentelemetry-sdk
|
||||
# typer
|
||||
urllib3==2.3.0
|
||||
# via requests
|
||||
wrapt==1.17.2
|
||||
# via
|
||||
# deprecated
|
||||
# opentelemetry-instrumentation
|
||||
# opentelemetry-instrumentation-grpc
|
||||
zipp==3.21.0
|
||||
# via importlib-metadata
|
|
@ -1,55 +1,367 @@
|
|||
certifi==2024.8.30 ; python_version >= "3.9" and python_version < "3.13"
|
||||
charset-normalizer==3.4.0 ; python_version >= "3.9" and python_version < "3.13"
|
||||
click==8.1.7 ; python_version >= "3.9" and python_version < "3.13"
|
||||
colorama==0.4.6 ; python_version >= "3.9" and python_version < "3.13" and (sys_platform == "win32" or platform_system == "Windows")
|
||||
deprecated==1.2.14 ; python_version >= "3.9" and python_version < "3.13"
|
||||
einops==0.8.0 ; python_version >= "3.9" and python_version < "3.13"
|
||||
filelock==3.16.1 ; python_version >= "3.9" and python_version < "3.13"
|
||||
fsspec==2024.6.1 ; python_version >= "3.9" and python_version < "3.13"
|
||||
googleapis-common-protos==1.65.0 ; python_version >= "3.9" and python_version < "3.13"
|
||||
grpc-interceptor==0.15.4 ; python_version >= "3.9" and python_version < "3.13"
|
||||
grpcio-reflection==1.62.3 ; python_version >= "3.9" and python_version < "3.13"
|
||||
grpcio-status==1.62.3 ; python_version >= "3.9" and python_version < "3.13"
|
||||
grpcio==1.68.0 ; python_version >= "3.9" and python_version < "3.13"
|
||||
hf-transfer==0.1.8 ; python_version >= "3.9" and python_version < "3.13"
|
||||
huggingface-hub==0.23.5 ; python_version >= "3.9" and python_version < "3.13"
|
||||
idna==3.10 ; python_version >= "3.9" and python_version < "3.13"
|
||||
importlib-metadata==7.1.0 ; python_version >= "3.9" and python_version < "3.13"
|
||||
loguru==0.7.2 ; python_version >= "3.9" and python_version < "3.13"
|
||||
markdown-it-py==3.0.0 ; python_version >= "3.9" and python_version < "3.13"
|
||||
mdurl==0.1.2 ; python_version >= "3.9" and python_version < "3.13"
|
||||
numpy==1.26.4 ; python_version >= "3.9" and python_version < "3.13"
|
||||
opentelemetry-api==1.27.0 ; python_version >= "3.9" and python_version < "3.13"
|
||||
opentelemetry-exporter-otlp-proto-common==1.27.0 ; python_version >= "3.9" and python_version < "3.13"
|
||||
opentelemetry-exporter-otlp-proto-grpc==1.27.0 ; python_version >= "3.9" and python_version < "3.13"
|
||||
opentelemetry-exporter-otlp-proto-http==1.27.0 ; python_version >= "3.9" and python_version < "3.13"
|
||||
opentelemetry-exporter-otlp==1.27.0 ; python_version >= "3.9" and python_version < "3.13"
|
||||
opentelemetry-instrumentation-grpc==0.48b0 ; python_version >= "3.9" and python_version < "3.13"
|
||||
opentelemetry-instrumentation==0.48b0 ; python_version >= "3.9" and python_version < "3.13"
|
||||
opentelemetry-proto==1.27.0 ; python_version >= "3.9" and python_version < "3.13"
|
||||
opentelemetry-sdk==1.27.0 ; python_version >= "3.9" and python_version < "3.13"
|
||||
opentelemetry-semantic-conventions==0.48b0 ; python_version >= "3.9" and python_version < "3.13"
|
||||
packaging==24.1 ; python_version >= "3.9" and python_version < "3.13"
|
||||
pillow==11.0.0 ; python_version >= "3.9" and python_version < "3.13"
|
||||
prometheus-client==0.20.0 ; python_version >= "3.9" and python_version < "3.13"
|
||||
protobuf==4.25.5 ; python_version >= "3.9" and python_version < "3.13"
|
||||
py-cpuinfo==9.0.0 ; python_version >= "3.9" and python_version < "3.13"
|
||||
pygments==2.18.0 ; python_version >= "3.9" and python_version < "3.13"
|
||||
pyyaml==6.0.2 ; python_version >= "3.9" and python_version < "3.13"
|
||||
regex==2024.9.11 ; python_version >= "3.9" and python_version < "3.13"
|
||||
requests==2.32.3 ; python_version >= "3.9" and python_version < "3.13"
|
||||
rich==13.9.4 ; python_version >= "3.9" and python_version < "3.13"
|
||||
safetensors==0.4.5 ; python_version >= "3.9" and python_version < "3.13"
|
||||
scipy==1.13.1 ; python_version >= "3.9" and python_version < "3.13"
|
||||
sentencepiece==0.2.0 ; python_version >= "3.9" and python_version < "3.13"
|
||||
setuptools==75.2.0 ; python_version >= "3.9" and python_version < "3.13"
|
||||
shellingham==1.5.4 ; python_version >= "3.9" and python_version < "3.13"
|
||||
tokenizers==0.20.3 ; python_version >= "3.9" and python_version < "3.13"
|
||||
tqdm==4.66.5 ; python_version >= "3.9" and python_version < "3.13"
|
||||
transformers==4.46.3 ; python_version >= "3.9" and python_version < "3.13"
|
||||
typer==0.12.5 ; python_version >= "3.9" and python_version < "3.13"
|
||||
typing-extensions==4.12.2 ; python_version >= "3.9" and python_version < "3.13"
|
||||
urllib3==2.2.3 ; python_version >= "3.9" and python_version < "3.13"
|
||||
win32-setctime==1.1.0 ; python_version >= "3.9" and python_version < "3.13" and sys_platform == "win32"
|
||||
wrapt==1.16.0 ; python_version >= "3.9" and python_version < "3.13"
|
||||
zipp==3.20.2 ; python_version >= "3.9" and python_version < "3.13"
|
||||
# This file was autogenerated by uv via the following command:
|
||||
# uv pip compile pyproject.toml --extra accelerate --extra compressed-tensors --extra quantize --extra peft --extra outlines -o requirements_intel.txt --python-version 3.11
|
||||
accelerate==1.3.0
|
||||
# via
|
||||
# text-generation-server (pyproject.toml)
|
||||
# peft
|
||||
aiohappyeyeballs==2.4.4
|
||||
# via aiohttp
|
||||
aiohttp==3.11.11
|
||||
# via
|
||||
# datasets
|
||||
# fsspec
|
||||
aiosignal==1.3.2
|
||||
# via aiohttp
|
||||
airportsdata==20241001
|
||||
# via outlines
|
||||
annotated-types==0.7.0
|
||||
# via pydantic
|
||||
attrs==25.1.0
|
||||
# via
|
||||
# aiohttp
|
||||
# jsonschema
|
||||
# referencing
|
||||
certifi==2024.8.30
|
||||
# via requests
|
||||
charset-normalizer==3.4.0
|
||||
# via requests
|
||||
click==8.1.7
|
||||
# via typer
|
||||
cloudpickle==3.1.1
|
||||
# via outlines
|
||||
compressed-tensors==0.9.1
|
||||
# via text-generation-server (pyproject.toml)
|
||||
datasets==2.21.0
|
||||
# via text-generation-server (pyproject.toml)
|
||||
deprecated==1.2.14
|
||||
# via
|
||||
# opentelemetry-api
|
||||
# opentelemetry-exporter-otlp-proto-grpc
|
||||
# opentelemetry-exporter-otlp-proto-http
|
||||
# opentelemetry-semantic-conventions
|
||||
dill==0.3.8
|
||||
# via
|
||||
# datasets
|
||||
# multiprocess
|
||||
diskcache==5.6.3
|
||||
# via outlines
|
||||
einops==0.8.0
|
||||
# via text-generation-server (pyproject.toml)
|
||||
filelock==3.16.1
|
||||
# via
|
||||
# datasets
|
||||
# huggingface-hub
|
||||
# torch
|
||||
# transformers
|
||||
frozenlist==1.5.0
|
||||
# via
|
||||
# aiohttp
|
||||
# aiosignal
|
||||
fsspec==2024.6.1
|
||||
# via
|
||||
# datasets
|
||||
# huggingface-hub
|
||||
# torch
|
||||
genson==1.3.0
|
||||
# via outlines
|
||||
googleapis-common-protos==1.65.0
|
||||
# via
|
||||
# grpcio-status
|
||||
# opentelemetry-exporter-otlp-proto-grpc
|
||||
# opentelemetry-exporter-otlp-proto-http
|
||||
grpc-interceptor==0.15.4
|
||||
# via text-generation-server (pyproject.toml)
|
||||
grpcio==1.68.0
|
||||
# via
|
||||
# text-generation-server (pyproject.toml)
|
||||
# grpc-interceptor
|
||||
# grpcio-reflection
|
||||
# grpcio-status
|
||||
# opentelemetry-exporter-otlp-proto-grpc
|
||||
grpcio-reflection==1.68.0
|
||||
# via text-generation-server (pyproject.toml)
|
||||
grpcio-status==1.68.0
|
||||
# via text-generation-server (pyproject.toml)
|
||||
hf-transfer==0.1.8
|
||||
# via text-generation-server (pyproject.toml)
|
||||
huggingface-hub==0.28.1
|
||||
# via
|
||||
# accelerate
|
||||
# datasets
|
||||
# peft
|
||||
# tokenizers
|
||||
# transformers
|
||||
idna==3.10
|
||||
# via
|
||||
# requests
|
||||
# yarl
|
||||
importlib-metadata==7.1.0
|
||||
# via opentelemetry-api
|
||||
interegular==0.3.3
|
||||
# via
|
||||
# outlines
|
||||
# outlines-core
|
||||
jinja2==3.1.5
|
||||
# via
|
||||
# outlines
|
||||
# torch
|
||||
jsonschema==4.23.0
|
||||
# via
|
||||
# outlines
|
||||
# outlines-core
|
||||
jsonschema-specifications==2024.10.1
|
||||
# via jsonschema
|
||||
lark==1.2.2
|
||||
# via outlines
|
||||
loguru==0.7.3
|
||||
# via text-generation-server (pyproject.toml)
|
||||
markdown-it-py==3.0.0
|
||||
# via rich
|
||||
markupsafe==3.0.2
|
||||
# via jinja2
|
||||
mdurl==0.1.2
|
||||
# via markdown-it-py
|
||||
mpmath==1.3.0
|
||||
# via sympy
|
||||
multidict==6.1.0
|
||||
# via
|
||||
# aiohttp
|
||||
# yarl
|
||||
multiprocess==0.70.16
|
||||
# via datasets
|
||||
nest-asyncio==1.6.0
|
||||
# via outlines
|
||||
networkx==3.4.2
|
||||
# via torch
|
||||
numpy==1.26.4
|
||||
# via
|
||||
# text-generation-server (pyproject.toml)
|
||||
# accelerate
|
||||
# datasets
|
||||
# outlines
|
||||
# pandas
|
||||
# peft
|
||||
# scipy
|
||||
# transformers
|
||||
nvidia-cublas-cu12==12.4.5.8
|
||||
# via
|
||||
# nvidia-cudnn-cu12
|
||||
# nvidia-cusolver-cu12
|
||||
# torch
|
||||
nvidia-cuda-cupti-cu12==12.4.127
|
||||
# via torch
|
||||
nvidia-cuda-nvrtc-cu12==12.4.127
|
||||
# via torch
|
||||
nvidia-cuda-runtime-cu12==12.4.127
|
||||
# via torch
|
||||
nvidia-cudnn-cu12==9.1.0.70
|
||||
# via torch
|
||||
nvidia-cufft-cu12==11.2.1.3
|
||||
# via torch
|
||||
nvidia-curand-cu12==10.3.5.147
|
||||
# via torch
|
||||
nvidia-cusolver-cu12==11.6.1.9
|
||||
# via torch
|
||||
nvidia-cusparse-cu12==12.3.1.170
|
||||
# via
|
||||
# nvidia-cusolver-cu12
|
||||
# torch
|
||||
nvidia-cusparselt-cu12==0.6.2
|
||||
# via torch
|
||||
nvidia-nccl-cu12==2.21.5
|
||||
# via torch
|
||||
nvidia-nvjitlink-cu12==12.4.127
|
||||
# via
|
||||
# nvidia-cusolver-cu12
|
||||
# nvidia-cusparse-cu12
|
||||
# torch
|
||||
nvidia-nvtx-cu12==12.4.127
|
||||
# via torch
|
||||
opentelemetry-api==1.30.0
|
||||
# via
|
||||
# text-generation-server (pyproject.toml)
|
||||
# opentelemetry-exporter-otlp-proto-grpc
|
||||
# opentelemetry-exporter-otlp-proto-http
|
||||
# opentelemetry-instrumentation
|
||||
# opentelemetry-instrumentation-grpc
|
||||
# opentelemetry-sdk
|
||||
# opentelemetry-semantic-conventions
|
||||
opentelemetry-exporter-otlp==1.30.0
|
||||
# via text-generation-server (pyproject.toml)
|
||||
opentelemetry-exporter-otlp-proto-common==1.30.0
|
||||
# via
|
||||
# opentelemetry-exporter-otlp-proto-grpc
|
||||
# opentelemetry-exporter-otlp-proto-http
|
||||
opentelemetry-exporter-otlp-proto-grpc==1.30.0
|
||||
# via opentelemetry-exporter-otlp
|
||||
opentelemetry-exporter-otlp-proto-http==1.30.0
|
||||
# via opentelemetry-exporter-otlp
|
||||
opentelemetry-instrumentation==0.51b0
|
||||
# via opentelemetry-instrumentation-grpc
|
||||
opentelemetry-instrumentation-grpc==0.51b0
|
||||
# via text-generation-server (pyproject.toml)
|
||||
opentelemetry-proto==1.30.0
|
||||
# via
|
||||
# opentelemetry-exporter-otlp-proto-common
|
||||
# opentelemetry-exporter-otlp-proto-grpc
|
||||
# opentelemetry-exporter-otlp-proto-http
|
||||
opentelemetry-sdk==1.30.0
|
||||
# via
|
||||
# opentelemetry-exporter-otlp-proto-grpc
|
||||
# opentelemetry-exporter-otlp-proto-http
|
||||
opentelemetry-semantic-conventions==0.51b0
|
||||
# via
|
||||
# opentelemetry-instrumentation
|
||||
# opentelemetry-instrumentation-grpc
|
||||
# opentelemetry-sdk
|
||||
outlines==0.1.14
|
||||
# via text-generation-server (pyproject.toml)
|
||||
outlines-core==0.1.26
|
||||
# via outlines
|
||||
packaging==24.1
|
||||
# via
|
||||
# accelerate
|
||||
# datasets
|
||||
# huggingface-hub
|
||||
# opentelemetry-instrumentation
|
||||
# peft
|
||||
# transformers
|
||||
pandas==2.2.3
|
||||
# via datasets
|
||||
peft==0.14.0
|
||||
# via text-generation-server (pyproject.toml)
|
||||
pillow==11.1.0
|
||||
# via text-generation-server (pyproject.toml)
|
||||
prometheus-client==0.21.1
|
||||
# via text-generation-server (pyproject.toml)
|
||||
propcache==0.2.1
|
||||
# via
|
||||
# aiohttp
|
||||
# yarl
|
||||
protobuf==5.29.3
|
||||
# via
|
||||
# text-generation-server (pyproject.toml)
|
||||
# googleapis-common-protos
|
||||
# grpcio-reflection
|
||||
# grpcio-status
|
||||
# opentelemetry-proto
|
||||
psutil==6.1.1
|
||||
# via
|
||||
# accelerate
|
||||
# peft
|
||||
py-cpuinfo==9.0.0
|
||||
# via text-generation-server (pyproject.toml)
|
||||
pyarrow==19.0.0
|
||||
# via datasets
|
||||
pycountry==24.6.1
|
||||
# via outlines
|
||||
pydantic==2.10.6
|
||||
# via
|
||||
# compressed-tensors
|
||||
# outlines
|
||||
pydantic-core==2.27.2
|
||||
# via pydantic
|
||||
pygments==2.18.0
|
||||
# via rich
|
||||
python-dateutil==2.9.0.post0
|
||||
# via pandas
|
||||
pytz==2025.1
|
||||
# via pandas
|
||||
pyyaml==6.0.2
|
||||
# via
|
||||
# accelerate
|
||||
# datasets
|
||||
# huggingface-hub
|
||||
# peft
|
||||
# transformers
|
||||
referencing==0.36.2
|
||||
# via
|
||||
# jsonschema
|
||||
# jsonschema-specifications
|
||||
# outlines
|
||||
regex==2024.9.11
|
||||
# via transformers
|
||||
requests==2.32.3
|
||||
# via
|
||||
# datasets
|
||||
# huggingface-hub
|
||||
# opentelemetry-exporter-otlp-proto-http
|
||||
# outlines
|
||||
# transformers
|
||||
rich==13.9.4
|
||||
# via
|
||||
# text-generation-server (pyproject.toml)
|
||||
# typer
|
||||
rpds-py==0.22.3
|
||||
# via
|
||||
# jsonschema
|
||||
# referencing
|
||||
safetensors==0.4.5
|
||||
# via
|
||||
# text-generation-server (pyproject.toml)
|
||||
# accelerate
|
||||
# peft
|
||||
# transformers
|
||||
scipy==1.13.1
|
||||
# via text-generation-server (pyproject.toml)
|
||||
sentencepiece==0.2.0
|
||||
# via text-generation-server (pyproject.toml)
|
||||
shellingham==1.5.4
|
||||
# via typer
|
||||
six==1.17.0
|
||||
# via python-dateutil
|
||||
sympy==1.13.1
|
||||
# via torch
|
||||
texttable==1.7.0
|
||||
# via text-generation-server (pyproject.toml)
|
||||
tokenizers==0.21.0
|
||||
# via
|
||||
# text-generation-server (pyproject.toml)
|
||||
# transformers
|
||||
torch==2.6.0
|
||||
# via
|
||||
# accelerate
|
||||
# compressed-tensors
|
||||
# outlines
|
||||
# peft
|
||||
tqdm==4.66.5
|
||||
# via
|
||||
# datasets
|
||||
# huggingface-hub
|
||||
# outlines
|
||||
# peft
|
||||
# transformers
|
||||
transformers==4.48.2
|
||||
# via
|
||||
# text-generation-server (pyproject.toml)
|
||||
# compressed-tensors
|
||||
# peft
|
||||
triton==3.2.0
|
||||
# via torch
|
||||
typer==0.15.1
|
||||
# via text-generation-server (pyproject.toml)
|
||||
typing-extensions==4.12.2
|
||||
# via
|
||||
# huggingface-hub
|
||||
# opentelemetry-sdk
|
||||
# outlines
|
||||
# pydantic
|
||||
# pydantic-core
|
||||
# referencing
|
||||
# torch
|
||||
# typer
|
||||
tzdata==2025.1
|
||||
# via pandas
|
||||
urllib3==2.2.3
|
||||
# via requests
|
||||
wrapt==1.16.0
|
||||
# via
|
||||
# deprecated
|
||||
# opentelemetry-instrumentation
|
||||
# opentelemetry-instrumentation-grpc
|
||||
xxhash==3.5.0
|
||||
# via datasets
|
||||
yarl==1.18.3
|
||||
# via aiohttp
|
||||
zipp==3.20.2
|
||||
# via importlib-metadata
|
||||
|
|
|
@ -1,55 +1,367 @@
|
|||
certifi==2024.8.30 ; python_version >= "3.9" and python_version < "3.13"
|
||||
charset-normalizer==3.4.0 ; python_version >= "3.9" and python_version < "3.13"
|
||||
click==8.1.7 ; python_version >= "3.9" and python_version < "3.13"
|
||||
colorama==0.4.6 ; python_version >= "3.9" and python_version < "3.13" and (sys_platform == "win32" or platform_system == "Windows")
|
||||
deprecated==1.2.14 ; python_version >= "3.9" and python_version < "3.13"
|
||||
einops==0.8.0 ; python_version >= "3.9" and python_version < "3.13"
|
||||
filelock==3.16.1 ; python_version >= "3.9" and python_version < "3.13"
|
||||
fsspec==2024.6.1 ; python_version >= "3.9" and python_version < "3.13"
|
||||
googleapis-common-protos==1.65.0 ; python_version >= "3.9" and python_version < "3.13"
|
||||
grpc-interceptor==0.15.4 ; python_version >= "3.9" and python_version < "3.13"
|
||||
grpcio-reflection==1.62.3 ; python_version >= "3.9" and python_version < "3.13"
|
||||
grpcio-status==1.62.3 ; python_version >= "3.9" and python_version < "3.13"
|
||||
grpcio==1.68.0 ; python_version >= "3.9" and python_version < "3.13"
|
||||
hf-transfer==0.1.8 ; python_version >= "3.9" and python_version < "3.13"
|
||||
huggingface-hub==0.23.5 ; python_version >= "3.9" and python_version < "3.13"
|
||||
idna==3.10 ; python_version >= "3.9" and python_version < "3.13"
|
||||
importlib-metadata==7.1.0 ; python_version >= "3.9" and python_version < "3.13"
|
||||
loguru==0.7.2 ; python_version >= "3.9" and python_version < "3.13"
|
||||
markdown-it-py==3.0.0 ; python_version >= "3.9" and python_version < "3.13"
|
||||
mdurl==0.1.2 ; python_version >= "3.9" and python_version < "3.13"
|
||||
numpy==1.26.4 ; python_version >= "3.9" and python_version < "3.13"
|
||||
opentelemetry-api==1.27.0 ; python_version >= "3.9" and python_version < "3.13"
|
||||
opentelemetry-exporter-otlp-proto-common==1.27.0 ; python_version >= "3.9" and python_version < "3.13"
|
||||
opentelemetry-exporter-otlp-proto-grpc==1.27.0 ; python_version >= "3.9" and python_version < "3.13"
|
||||
opentelemetry-exporter-otlp-proto-http==1.27.0 ; python_version >= "3.9" and python_version < "3.13"
|
||||
opentelemetry-exporter-otlp==1.27.0 ; python_version >= "3.9" and python_version < "3.13"
|
||||
opentelemetry-instrumentation-grpc==0.48b0 ; python_version >= "3.9" and python_version < "3.13"
|
||||
opentelemetry-instrumentation==0.48b0 ; python_version >= "3.9" and python_version < "3.13"
|
||||
opentelemetry-proto==1.27.0 ; python_version >= "3.9" and python_version < "3.13"
|
||||
opentelemetry-sdk==1.27.0 ; python_version >= "3.9" and python_version < "3.13"
|
||||
opentelemetry-semantic-conventions==0.48b0 ; python_version >= "3.9" and python_version < "3.13"
|
||||
packaging==24.1 ; python_version >= "3.9" and python_version < "3.13"
|
||||
pillow==11.0.0 ; python_version >= "3.9" and python_version < "3.13"
|
||||
prometheus-client==0.20.0 ; python_version >= "3.9" and python_version < "3.13"
|
||||
protobuf==4.25.5 ; python_version >= "3.9" and python_version < "3.13"
|
||||
py-cpuinfo==9.0.0 ; python_version >= "3.9" and python_version < "3.13"
|
||||
pygments==2.18.0 ; python_version >= "3.9" and python_version < "3.13"
|
||||
pyyaml==6.0.2 ; python_version >= "3.9" and python_version < "3.13"
|
||||
regex==2024.9.11 ; python_version >= "3.9" and python_version < "3.13"
|
||||
requests==2.32.3 ; python_version >= "3.9" and python_version < "3.13"
|
||||
rich==13.9.4 ; python_version >= "3.9" and python_version < "3.13"
|
||||
safetensors==0.4.5 ; python_version >= "3.9" and python_version < "3.13"
|
||||
scipy==1.13.1 ; python_version >= "3.9" and python_version < "3.13"
|
||||
sentencepiece==0.2.0 ; python_version >= "3.9" and python_version < "3.13"
|
||||
setuptools==75.2.0 ; python_version >= "3.9" and python_version < "3.13"
|
||||
shellingham==1.5.4 ; python_version >= "3.9" and python_version < "3.13"
|
||||
tokenizers==0.20.3 ; python_version >= "3.9" and python_version < "3.13"
|
||||
tqdm==4.66.5 ; python_version >= "3.9" and python_version < "3.13"
|
||||
transformers==4.46.3 ; python_version >= "3.9" and python_version < "3.13"
|
||||
typer==0.12.5 ; python_version >= "3.9" and python_version < "3.13"
|
||||
typing-extensions==4.12.2 ; python_version >= "3.9" and python_version < "3.13"
|
||||
urllib3==2.2.3 ; python_version >= "3.9" and python_version < "3.13"
|
||||
win32-setctime==1.1.0 ; python_version >= "3.9" and python_version < "3.13" and sys_platform == "win32"
|
||||
wrapt==1.16.0 ; python_version >= "3.9" and python_version < "3.13"
|
||||
zipp==3.20.2 ; python_version >= "3.9" and python_version < "3.13"
|
||||
# This file was autogenerated by uv via the following command:
|
||||
# uv pip compile pyproject.toml --extra accelerate --extra compressed-tensors --extra quantize --extra peft --extra outlines -o requirements_rocm.txt --python-version 3.11
|
||||
accelerate==1.3.0
|
||||
# via
|
||||
# text-generation-server (pyproject.toml)
|
||||
# peft
|
||||
aiohappyeyeballs==2.4.4
|
||||
# via aiohttp
|
||||
aiohttp==3.11.11
|
||||
# via
|
||||
# datasets
|
||||
# fsspec
|
||||
aiosignal==1.3.2
|
||||
# via aiohttp
|
||||
airportsdata==20241001
|
||||
# via outlines
|
||||
annotated-types==0.7.0
|
||||
# via pydantic
|
||||
attrs==25.1.0
|
||||
# via
|
||||
# aiohttp
|
||||
# jsonschema
|
||||
# referencing
|
||||
certifi==2024.8.30
|
||||
# via requests
|
||||
charset-normalizer==3.4.0
|
||||
# via requests
|
||||
click==8.1.7
|
||||
# via typer
|
||||
cloudpickle==3.1.1
|
||||
# via outlines
|
||||
compressed-tensors==0.9.1
|
||||
# via text-generation-server (pyproject.toml)
|
||||
datasets==2.21.0
|
||||
# via text-generation-server (pyproject.toml)
|
||||
deprecated==1.2.14
|
||||
# via
|
||||
# opentelemetry-api
|
||||
# opentelemetry-exporter-otlp-proto-grpc
|
||||
# opentelemetry-exporter-otlp-proto-http
|
||||
# opentelemetry-semantic-conventions
|
||||
dill==0.3.8
|
||||
# via
|
||||
# datasets
|
||||
# multiprocess
|
||||
diskcache==5.6.3
|
||||
# via outlines
|
||||
einops==0.8.0
|
||||
# via text-generation-server (pyproject.toml)
|
||||
filelock==3.16.1
|
||||
# via
|
||||
# datasets
|
||||
# huggingface-hub
|
||||
# torch
|
||||
# transformers
|
||||
frozenlist==1.5.0
|
||||
# via
|
||||
# aiohttp
|
||||
# aiosignal
|
||||
fsspec==2024.6.1
|
||||
# via
|
||||
# datasets
|
||||
# huggingface-hub
|
||||
# torch
|
||||
genson==1.3.0
|
||||
# via outlines
|
||||
googleapis-common-protos==1.65.0
|
||||
# via
|
||||
# grpcio-status
|
||||
# opentelemetry-exporter-otlp-proto-grpc
|
||||
# opentelemetry-exporter-otlp-proto-http
|
||||
grpc-interceptor==0.15.4
|
||||
# via text-generation-server (pyproject.toml)
|
||||
grpcio==1.68.0
|
||||
# via
|
||||
# text-generation-server (pyproject.toml)
|
||||
# grpc-interceptor
|
||||
# grpcio-reflection
|
||||
# grpcio-status
|
||||
# opentelemetry-exporter-otlp-proto-grpc
|
||||
grpcio-reflection==1.68.0
|
||||
# via text-generation-server (pyproject.toml)
|
||||
grpcio-status==1.68.0
|
||||
# via text-generation-server (pyproject.toml)
|
||||
hf-transfer==0.1.8
|
||||
# via text-generation-server (pyproject.toml)
|
||||
huggingface-hub==0.28.1
|
||||
# via
|
||||
# accelerate
|
||||
# datasets
|
||||
# peft
|
||||
# tokenizers
|
||||
# transformers
|
||||
idna==3.10
|
||||
# via
|
||||
# requests
|
||||
# yarl
|
||||
importlib-metadata==7.1.0
|
||||
# via opentelemetry-api
|
||||
interegular==0.3.3
|
||||
# via
|
||||
# outlines
|
||||
# outlines-core
|
||||
jinja2==3.1.5
|
||||
# via
|
||||
# outlines
|
||||
# torch
|
||||
jsonschema==4.23.0
|
||||
# via
|
||||
# outlines
|
||||
# outlines-core
|
||||
jsonschema-specifications==2024.10.1
|
||||
# via jsonschema
|
||||
lark==1.2.2
|
||||
# via outlines
|
||||
loguru==0.7.3
|
||||
# via text-generation-server (pyproject.toml)
|
||||
markdown-it-py==3.0.0
|
||||
# via rich
|
||||
markupsafe==3.0.2
|
||||
# via jinja2
|
||||
mdurl==0.1.2
|
||||
# via markdown-it-py
|
||||
mpmath==1.3.0
|
||||
# via sympy
|
||||
multidict==6.1.0
|
||||
# via
|
||||
# aiohttp
|
||||
# yarl
|
||||
multiprocess==0.70.16
|
||||
# via datasets
|
||||
nest-asyncio==1.6.0
|
||||
# via outlines
|
||||
networkx==3.4.2
|
||||
# via torch
|
||||
numpy==1.26.4
|
||||
# via
|
||||
# text-generation-server (pyproject.toml)
|
||||
# accelerate
|
||||
# datasets
|
||||
# outlines
|
||||
# pandas
|
||||
# peft
|
||||
# scipy
|
||||
# transformers
|
||||
nvidia-cublas-cu12==12.4.5.8
|
||||
# via
|
||||
# nvidia-cudnn-cu12
|
||||
# nvidia-cusolver-cu12
|
||||
# torch
|
||||
nvidia-cuda-cupti-cu12==12.4.127
|
||||
# via torch
|
||||
nvidia-cuda-nvrtc-cu12==12.4.127
|
||||
# via torch
|
||||
nvidia-cuda-runtime-cu12==12.4.127
|
||||
# via torch
|
||||
nvidia-cudnn-cu12==9.1.0.70
|
||||
# via torch
|
||||
nvidia-cufft-cu12==11.2.1.3
|
||||
# via torch
|
||||
nvidia-curand-cu12==10.3.5.147
|
||||
# via torch
|
||||
nvidia-cusolver-cu12==11.6.1.9
|
||||
# via torch
|
||||
nvidia-cusparse-cu12==12.3.1.170
|
||||
# via
|
||||
# nvidia-cusolver-cu12
|
||||
# torch
|
||||
nvidia-cusparselt-cu12==0.6.2
|
||||
# via torch
|
||||
nvidia-nccl-cu12==2.21.5
|
||||
# via torch
|
||||
nvidia-nvjitlink-cu12==12.4.127
|
||||
# via
|
||||
# nvidia-cusolver-cu12
|
||||
# nvidia-cusparse-cu12
|
||||
# torch
|
||||
nvidia-nvtx-cu12==12.4.127
|
||||
# via torch
|
||||
opentelemetry-api==1.30.0
|
||||
# via
|
||||
# text-generation-server (pyproject.toml)
|
||||
# opentelemetry-exporter-otlp-proto-grpc
|
||||
# opentelemetry-exporter-otlp-proto-http
|
||||
# opentelemetry-instrumentation
|
||||
# opentelemetry-instrumentation-grpc
|
||||
# opentelemetry-sdk
|
||||
# opentelemetry-semantic-conventions
|
||||
opentelemetry-exporter-otlp==1.30.0
|
||||
# via text-generation-server (pyproject.toml)
|
||||
opentelemetry-exporter-otlp-proto-common==1.30.0
|
||||
# via
|
||||
# opentelemetry-exporter-otlp-proto-grpc
|
||||
# opentelemetry-exporter-otlp-proto-http
|
||||
opentelemetry-exporter-otlp-proto-grpc==1.30.0
|
||||
# via opentelemetry-exporter-otlp
|
||||
opentelemetry-exporter-otlp-proto-http==1.30.0
|
||||
# via opentelemetry-exporter-otlp
|
||||
opentelemetry-instrumentation==0.51b0
|
||||
# via opentelemetry-instrumentation-grpc
|
||||
opentelemetry-instrumentation-grpc==0.51b0
|
||||
# via text-generation-server (pyproject.toml)
|
||||
opentelemetry-proto==1.30.0
|
||||
# via
|
||||
# opentelemetry-exporter-otlp-proto-common
|
||||
# opentelemetry-exporter-otlp-proto-grpc
|
||||
# opentelemetry-exporter-otlp-proto-http
|
||||
opentelemetry-sdk==1.30.0
|
||||
# via
|
||||
# opentelemetry-exporter-otlp-proto-grpc
|
||||
# opentelemetry-exporter-otlp-proto-http
|
||||
opentelemetry-semantic-conventions==0.51b0
|
||||
# via
|
||||
# opentelemetry-instrumentation
|
||||
# opentelemetry-instrumentation-grpc
|
||||
# opentelemetry-sdk
|
||||
outlines==0.1.14
|
||||
# via text-generation-server (pyproject.toml)
|
||||
outlines-core==0.1.26
|
||||
# via outlines
|
||||
packaging==24.1
|
||||
# via
|
||||
# accelerate
|
||||
# datasets
|
||||
# huggingface-hub
|
||||
# opentelemetry-instrumentation
|
||||
# peft
|
||||
# transformers
|
||||
pandas==2.2.3
|
||||
# via datasets
|
||||
peft==0.14.0
|
||||
# via text-generation-server (pyproject.toml)
|
||||
pillow==11.1.0
|
||||
# via text-generation-server (pyproject.toml)
|
||||
prometheus-client==0.21.1
|
||||
# via text-generation-server (pyproject.toml)
|
||||
propcache==0.2.1
|
||||
# via
|
||||
# aiohttp
|
||||
# yarl
|
||||
protobuf==5.29.3
|
||||
# via
|
||||
# text-generation-server (pyproject.toml)
|
||||
# googleapis-common-protos
|
||||
# grpcio-reflection
|
||||
# grpcio-status
|
||||
# opentelemetry-proto
|
||||
psutil==6.1.1
|
||||
# via
|
||||
# accelerate
|
||||
# peft
|
||||
py-cpuinfo==9.0.0
|
||||
# via text-generation-server (pyproject.toml)
|
||||
pyarrow==19.0.0
|
||||
# via datasets
|
||||
pycountry==24.6.1
|
||||
# via outlines
|
||||
pydantic==2.10.6
|
||||
# via
|
||||
# compressed-tensors
|
||||
# outlines
|
||||
pydantic-core==2.27.2
|
||||
# via pydantic
|
||||
pygments==2.18.0
|
||||
# via rich
|
||||
python-dateutil==2.9.0.post0
|
||||
# via pandas
|
||||
pytz==2025.1
|
||||
# via pandas
|
||||
pyyaml==6.0.2
|
||||
# via
|
||||
# accelerate
|
||||
# datasets
|
||||
# huggingface-hub
|
||||
# peft
|
||||
# transformers
|
||||
referencing==0.36.2
|
||||
# via
|
||||
# jsonschema
|
||||
# jsonschema-specifications
|
||||
# outlines
|
||||
regex==2024.9.11
|
||||
# via transformers
|
||||
requests==2.32.3
|
||||
# via
|
||||
# datasets
|
||||
# huggingface-hub
|
||||
# opentelemetry-exporter-otlp-proto-http
|
||||
# outlines
|
||||
# transformers
|
||||
rich==13.9.4
|
||||
# via
|
||||
# text-generation-server (pyproject.toml)
|
||||
# typer
|
||||
rpds-py==0.22.3
|
||||
# via
|
||||
# jsonschema
|
||||
# referencing
|
||||
safetensors==0.4.5
|
||||
# via
|
||||
# text-generation-server (pyproject.toml)
|
||||
# accelerate
|
||||
# peft
|
||||
# transformers
|
||||
scipy==1.13.1
|
||||
# via text-generation-server (pyproject.toml)
|
||||
sentencepiece==0.2.0
|
||||
# via text-generation-server (pyproject.toml)
|
||||
shellingham==1.5.4
|
||||
# via typer
|
||||
six==1.17.0
|
||||
# via python-dateutil
|
||||
sympy==1.13.1
|
||||
# via torch
|
||||
texttable==1.7.0
|
||||
# via text-generation-server (pyproject.toml)
|
||||
tokenizers==0.21.0
|
||||
# via
|
||||
# text-generation-server (pyproject.toml)
|
||||
# transformers
|
||||
torch==2.6.0
|
||||
# via
|
||||
# accelerate
|
||||
# compressed-tensors
|
||||
# outlines
|
||||
# peft
|
||||
tqdm==4.66.5
|
||||
# via
|
||||
# datasets
|
||||
# huggingface-hub
|
||||
# outlines
|
||||
# peft
|
||||
# transformers
|
||||
transformers==4.48.2
|
||||
# via
|
||||
# text-generation-server (pyproject.toml)
|
||||
# compressed-tensors
|
||||
# peft
|
||||
triton==3.2.0
|
||||
# via torch
|
||||
typer==0.15.1
|
||||
# via text-generation-server (pyproject.toml)
|
||||
typing-extensions==4.12.2
|
||||
# via
|
||||
# huggingface-hub
|
||||
# opentelemetry-sdk
|
||||
# outlines
|
||||
# pydantic
|
||||
# pydantic-core
|
||||
# referencing
|
||||
# torch
|
||||
# typer
|
||||
tzdata==2025.1
|
||||
# via pandas
|
||||
urllib3==2.2.3
|
||||
# via requests
|
||||
wrapt==1.16.0
|
||||
# via
|
||||
# deprecated
|
||||
# opentelemetry-instrumentation
|
||||
# opentelemetry-instrumentation-grpc
|
||||
xxhash==3.5.0
|
||||
# via datasets
|
||||
yarl==1.18.3
|
||||
# via aiohttp
|
||||
zipp==3.20.2
|
||||
# via importlib-metadata
|
||||
|
|
|
@ -997,15 +997,15 @@ wheels = [
|
|||
|
||||
[[package]]
|
||||
name = "moe-kernels"
|
||||
version = "0.8.0"
|
||||
source = { url = "https://github.com/danieldk/moe-kernels/releases/download/v0.8.0/moe_kernels-0.8.0+cu123torch2.5-cp39-abi3-linux_x86_64.whl" }
|
||||
version = "0.8.2"
|
||||
source = { url = "https://github.com/danieldk/moe-kernels/releases/download/v0.8.2/moe_kernels-0.8.2+cu123torch2.5-cp39-abi3-linux_x86_64.whl" }
|
||||
dependencies = [
|
||||
{ name = "nvidia-ml-py" },
|
||||
{ name = "torch" },
|
||||
{ name = "triton" },
|
||||
]
|
||||
wheels = [
|
||||
{ url = "https://github.com/danieldk/moe-kernels/releases/download/v0.8.0/moe_kernels-0.8.0+cu123torch2.5-cp39-abi3-linux_x86_64.whl", hash = "sha256:92c4e083c037a325458e731dda6770790495cab273c9bbf5f50fb8e262c099de" },
|
||||
{ url = "https://github.com/danieldk/moe-kernels/releases/download/v0.8.2/moe_kernels-0.8.2+cu123torch2.5-cp39-abi3-linux_x86_64.whl", hash = "sha256:1ed5b26f52339d25ea2513e99e8b6239cf1921af3eac54e03a46bb8f8efb380b" },
|
||||
]
|
||||
|
||||
[package.metadata]
|
||||
|
@ -1308,7 +1308,6 @@ name = "nvidia-cublas-cu12"
|
|||
version = "12.4.5.8"
|
||||
source = { registry = "https://pypi.org/simple" }
|
||||
wheels = [
|
||||
{ url = "https://files.pythonhosted.org/packages/7f/7f/7fbae15a3982dc9595e49ce0f19332423b260045d0a6afe93cdbe2f1f624/nvidia_cublas_cu12-12.4.5.8-py3-none-manylinux2014_aarch64.whl", hash = "sha256:0f8aa1706812e00b9f19dfe0cdb3999b092ccb8ca168c0db5b8ea712456fd9b3", size = 363333771 },
|
||||
{ url = "https://files.pythonhosted.org/packages/ae/71/1c91302526c45ab494c23f61c7a84aa568b8c1f9d196efa5993957faf906/nvidia_cublas_cu12-12.4.5.8-py3-none-manylinux2014_x86_64.whl", hash = "sha256:2fc8da60df463fdefa81e323eef2e36489e1c94335b5358bcb38360adf75ac9b", size = 363438805 },
|
||||
]
|
||||
|
||||
|
@ -1317,7 +1316,6 @@ name = "nvidia-cuda-cupti-cu12"
|
|||
version = "12.4.127"
|
||||
source = { registry = "https://pypi.org/simple" }
|
||||
wheels = [
|
||||
{ url = "https://files.pythonhosted.org/packages/93/b5/9fb3d00386d3361b03874246190dfec7b206fd74e6e287b26a8fcb359d95/nvidia_cuda_cupti_cu12-12.4.127-py3-none-manylinux2014_aarch64.whl", hash = "sha256:79279b35cf6f91da114182a5ce1864997fd52294a87a16179ce275773799458a", size = 12354556 },
|
||||
{ url = "https://files.pythonhosted.org/packages/67/42/f4f60238e8194a3106d06a058d494b18e006c10bb2b915655bd9f6ea4cb1/nvidia_cuda_cupti_cu12-12.4.127-py3-none-manylinux2014_x86_64.whl", hash = "sha256:9dec60f5ac126f7bb551c055072b69d85392b13311fcc1bcda2202d172df30fb", size = 13813957 },
|
||||
]
|
||||
|
||||
|
@ -1326,7 +1324,6 @@ name = "nvidia-cuda-nvrtc-cu12"
|
|||
version = "12.4.127"
|
||||
source = { registry = "https://pypi.org/simple" }
|
||||
wheels = [
|
||||
{ url = "https://files.pythonhosted.org/packages/77/aa/083b01c427e963ad0b314040565ea396f914349914c298556484f799e61b/nvidia_cuda_nvrtc_cu12-12.4.127-py3-none-manylinux2014_aarch64.whl", hash = "sha256:0eedf14185e04b76aa05b1fea04133e59f465b6f960c0cbf4e37c3cb6b0ea198", size = 24133372 },
|
||||
{ url = "https://files.pythonhosted.org/packages/2c/14/91ae57cd4db3f9ef7aa99f4019cfa8d54cb4caa7e00975df6467e9725a9f/nvidia_cuda_nvrtc_cu12-12.4.127-py3-none-manylinux2014_x86_64.whl", hash = "sha256:a178759ebb095827bd30ef56598ec182b85547f1508941a3d560eb7ea1fbf338", size = 24640306 },
|
||||
]
|
||||
|
||||
|
@ -1335,7 +1332,6 @@ name = "nvidia-cuda-runtime-cu12"
|
|||
version = "12.4.127"
|
||||
source = { registry = "https://pypi.org/simple" }
|
||||
wheels = [
|
||||
{ url = "https://files.pythonhosted.org/packages/a1/aa/b656d755f474e2084971e9a297def515938d56b466ab39624012070cb773/nvidia_cuda_runtime_cu12-12.4.127-py3-none-manylinux2014_aarch64.whl", hash = "sha256:961fe0e2e716a2a1d967aab7caee97512f71767f852f67432d572e36cb3a11f3", size = 894177 },
|
||||
{ url = "https://files.pythonhosted.org/packages/ea/27/1795d86fe88ef397885f2e580ac37628ed058a92ed2c39dc8eac3adf0619/nvidia_cuda_runtime_cu12-12.4.127-py3-none-manylinux2014_x86_64.whl", hash = "sha256:64403288fa2136ee8e467cdc9c9427e0434110899d07c779f25b5c068934faa5", size = 883737 },
|
||||
]
|
||||
|
||||
|
@ -1358,7 +1354,6 @@ dependencies = [
|
|||
{ name = "nvidia-nvjitlink-cu12" },
|
||||
]
|
||||
wheels = [
|
||||
{ url = "https://files.pythonhosted.org/packages/7a/8a/0e728f749baca3fbeffad762738276e5df60851958be7783af121a7221e7/nvidia_cufft_cu12-11.2.1.3-py3-none-manylinux2014_aarch64.whl", hash = "sha256:5dad8008fc7f92f5ddfa2101430917ce2ffacd86824914c82e28990ad7f00399", size = 211422548 },
|
||||
{ url = "https://files.pythonhosted.org/packages/27/94/3266821f65b92b3138631e9c8e7fe1fb513804ac934485a8d05776e1dd43/nvidia_cufft_cu12-11.2.1.3-py3-none-manylinux2014_x86_64.whl", hash = "sha256:f083fc24912aa410be21fa16d157fed2055dab1cc4b6934a0e03cba69eb242b9", size = 211459117 },
|
||||
]
|
||||
|
||||
|
@ -1367,7 +1362,6 @@ name = "nvidia-curand-cu12"
|
|||
version = "10.3.5.147"
|
||||
source = { registry = "https://pypi.org/simple" }
|
||||
wheels = [
|
||||
{ url = "https://files.pythonhosted.org/packages/80/9c/a79180e4d70995fdf030c6946991d0171555c6edf95c265c6b2bf7011112/nvidia_curand_cu12-10.3.5.147-py3-none-manylinux2014_aarch64.whl", hash = "sha256:1f173f09e3e3c76ab084aba0de819c49e56614feae5c12f69883f4ae9bb5fad9", size = 56314811 },
|
||||
{ url = "https://files.pythonhosted.org/packages/8a/6d/44ad094874c6f1b9c654f8ed939590bdc408349f137f9b98a3a23ccec411/nvidia_curand_cu12-10.3.5.147-py3-none-manylinux2014_x86_64.whl", hash = "sha256:a88f583d4e0bb643c49743469964103aa59f7f708d862c3ddb0fc07f851e3b8b", size = 56305206 },
|
||||
]
|
||||
|
||||
|
@ -1381,7 +1375,6 @@ dependencies = [
|
|||
{ name = "nvidia-nvjitlink-cu12" },
|
||||
]
|
||||
wheels = [
|
||||
{ url = "https://files.pythonhosted.org/packages/46/6b/a5c33cf16af09166845345275c34ad2190944bcc6026797a39f8e0a282e0/nvidia_cusolver_cu12-11.6.1.9-py3-none-manylinux2014_aarch64.whl", hash = "sha256:d338f155f174f90724bbde3758b7ac375a70ce8e706d70b018dd3375545fc84e", size = 127634111 },
|
||||
{ url = "https://files.pythonhosted.org/packages/3a/e1/5b9089a4b2a4790dfdea8b3a006052cfecff58139d5a4e34cb1a51df8d6f/nvidia_cusolver_cu12-11.6.1.9-py3-none-manylinux2014_x86_64.whl", hash = "sha256:19e33fa442bcfd085b3086c4ebf7e8debc07cfe01e11513cc6d332fd918ac260", size = 127936057 },
|
||||
]
|
||||
|
||||
|
@ -1393,7 +1386,6 @@ dependencies = [
|
|||
{ name = "nvidia-nvjitlink-cu12" },
|
||||
]
|
||||
wheels = [
|
||||
{ url = "https://files.pythonhosted.org/packages/96/a9/c0d2f83a53d40a4a41be14cea6a0bf9e668ffcf8b004bd65633f433050c0/nvidia_cusparse_cu12-12.3.1.170-py3-none-manylinux2014_aarch64.whl", hash = "sha256:9d32f62896231ebe0480efd8a7f702e143c98cfaa0e8a76df3386c1ba2b54df3", size = 207381987 },
|
||||
{ url = "https://files.pythonhosted.org/packages/db/f7/97a9ea26ed4bbbfc2d470994b8b4f338ef663be97b8f677519ac195e113d/nvidia_cusparse_cu12-12.3.1.170-py3-none-manylinux2014_x86_64.whl", hash = "sha256:ea4f11a2904e2a8dc4b1833cc1b5181cde564edd0d5cd33e3c168eff2d1863f1", size = 207454763 },
|
||||
]
|
||||
|
||||
|
@ -1419,7 +1411,6 @@ name = "nvidia-nvjitlink-cu12"
|
|||
version = "12.4.127"
|
||||
source = { registry = "https://pypi.org/simple" }
|
||||
wheels = [
|
||||
{ url = "https://files.pythonhosted.org/packages/02/45/239d52c05074898a80a900f49b1615d81c07fceadd5ad6c4f86a987c0bc4/nvidia_nvjitlink_cu12-12.4.127-py3-none-manylinux2014_aarch64.whl", hash = "sha256:4abe7fef64914ccfa909bc2ba39739670ecc9e820c83ccc7a6ed414122599b83", size = 20552510 },
|
||||
{ url = "https://files.pythonhosted.org/packages/ff/ff/847841bacfbefc97a00036e0fce5a0f086b640756dc38caea5e1bb002655/nvidia_nvjitlink_cu12-12.4.127-py3-none-manylinux2014_x86_64.whl", hash = "sha256:06b3b9b25bf3f8af351d664978ca26a16d2c5127dbd53c0497e28d1fb9611d57", size = 21066810 },
|
||||
]
|
||||
|
||||
|
@ -1428,7 +1419,6 @@ name = "nvidia-nvtx-cu12"
|
|||
version = "12.4.127"
|
||||
source = { registry = "https://pypi.org/simple" }
|
||||
wheels = [
|
||||
{ url = "https://files.pythonhosted.org/packages/06/39/471f581edbb7804b39e8063d92fc8305bdc7a80ae5c07dbe6ea5c50d14a5/nvidia_nvtx_cu12-12.4.127-py3-none-manylinux2014_aarch64.whl", hash = "sha256:7959ad635db13edf4fc65c06a6e9f9e55fc2f92596db928d169c0bb031e88ef3", size = 100417 },
|
||||
{ url = "https://files.pythonhosted.org/packages/87/20/199b8713428322a2f22b722c62b8cc278cc53dffa9705d744484b5035ee9/nvidia_nvtx_cu12-12.4.127-py3-none-manylinux2014_x86_64.whl", hash = "sha256:781e950d9b9f60d8241ccea575b32f5105a5baf4c2351cab5256a24869f12a1a", size = 99144 },
|
||||
]
|
||||
|
||||
|
@ -2656,7 +2646,7 @@ wheels = [
|
|||
[[package]]
|
||||
name = "text-generation-server"
|
||||
version = "2.0.5.dev0"
|
||||
source = { virtual = "." }
|
||||
source = { editable = "." }
|
||||
dependencies = [
|
||||
{ name = "einops" },
|
||||
{ name = "grpc-interceptor" },
|
||||
|
@ -2680,6 +2670,7 @@ dependencies = [
|
|||
{ name = "scipy", version = "1.15.1", source = { registry = "https://pypi.org/simple" }, marker = "python_full_version >= '3.10'" },
|
||||
{ name = "sentencepiece" },
|
||||
{ name = "tokenizers" },
|
||||
{ name = "transformers" },
|
||||
{ name = "typer" },
|
||||
]
|
||||
|
||||
|
@ -2746,7 +2737,7 @@ requires-dist = [
|
|||
{ name = "marlin-kernels", marker = "python_full_version == '3.10.*' and extra == 'marlin'", url = "https://github.com/danieldk/marlin-kernels/releases/download/v0.3.7/marlin_kernels-0.3.7+cu123torch2.5-cp310-cp310-linux_x86_64.whl" },
|
||||
{ name = "marlin-kernels", marker = "python_full_version == '3.11.*' and extra == 'marlin'", url = "https://github.com/danieldk/marlin-kernels/releases/download/v0.3.7/marlin_kernels-0.3.7+cu123torch2.5-cp311-cp311-linux_x86_64.whl" },
|
||||
{ name = "marlin-kernels", marker = "python_full_version == '3.12.*' and extra == 'marlin'", url = "https://github.com/danieldk/marlin-kernels/releases/download/v0.3.7/marlin_kernels-0.3.7+cu123torch2.5-cp312-cp312-linux_x86_64.whl" },
|
||||
{ name = "moe-kernels", marker = "extra == 'moe'", url = "https://github.com/danieldk/moe-kernels/releases/download/v0.8.0/moe_kernels-0.8.0+cu123torch2.5-cp39-abi3-linux_x86_64.whl" },
|
||||
{ name = "moe-kernels", marker = "extra == 'moe'", url = "https://github.com/danieldk/moe-kernels/releases/download/v0.8.2/moe_kernels-0.8.2+cu123torch2.5-cp39-abi3-linux_x86_64.whl" },
|
||||
{ name = "mypy-protobuf", marker = "extra == 'gen'", specifier = ">=3.6.0" },
|
||||
{ name = "numpy", specifier = ">=1.26,<3" },
|
||||
{ name = "opentelemetry-api", specifier = ">=1.27.0" },
|
||||
|
@ -2765,6 +2756,7 @@ requires-dist = [
|
|||
{ name = "sentencepiece", specifier = ">=0.2.0" },
|
||||
{ name = "texttable", marker = "extra == 'quantize'", specifier = ">=1.6.7,<2" },
|
||||
{ name = "tokenizers", specifier = ">=0.20.3" },
|
||||
{ name = "transformers", specifier = ">=4.48.0" },
|
||||
{ name = "typer", specifier = ">=0.15.1" },
|
||||
]
|
||||
|
||||
|
@ -2902,7 +2894,7 @@ wheels = [
|
|||
|
||||
[[package]]
|
||||
name = "transformers"
|
||||
version = "4.48.0"
|
||||
version = "4.48.2"
|
||||
source = { registry = "https://pypi.org/simple" }
|
||||
dependencies = [
|
||||
{ name = "filelock" },
|
||||
|
@ -2917,9 +2909,9 @@ dependencies = [
|
|||
{ name = "tokenizers" },
|
||||
{ name = "tqdm" },
|
||||
]
|
||||
sdist = { url = "https://files.pythonhosted.org/packages/ea/71/93a6331682d6f15adf7d646956db0c43e5f1759bbbd05f2ef53029bae107/transformers-4.48.0.tar.gz", hash = "sha256:03fdfcbfb8b0367fb6c9fbe9d1c9aa54dfd847618be9b52400b2811d22799cb1", size = 8372101 }
|
||||
sdist = { url = "https://files.pythonhosted.org/packages/c5/cf/1093586e09c8d889d2f6b8ffe6a1369e1e179eb7b8e732fc0f348a8fe58f/transformers-4.48.2.tar.gz", hash = "sha256:dcfb73473e61f22fb3366fe2471ed2e42779ecdd49527a1bdf1937574855d516", size = 8370945 }
|
||||
wheels = [
|
||||
{ url = "https://files.pythonhosted.org/packages/45/d6/a69764e89fc5c2c957aa473881527c8c35521108d553df703e9ba703daeb/transformers-4.48.0-py3-none-any.whl", hash = "sha256:6d3de6d71cb5f2a10f9775ccc17abce9620195caaf32ec96542bd2a6937f25b0", size = 9673380 },
|
||||
{ url = "https://files.pythonhosted.org/packages/bd/40/902c95a2a6f5d2d120c940ac4bd1f937c01035af529803c13d65ca33c2d1/transformers-4.48.2-py3-none-any.whl", hash = "sha256:493bc5b0268b116eff305edf6656367fc89cf570e7a9d5891369e04751db698a", size = 9667774 },
|
||||
]
|
||||
|
||||
[[package]]
|
||||
|
|
|
@ -2,4 +2,5 @@
|
|||
|
||||
ldconfig 2>/dev/null || echo 'unable to refresh ld cache, not a big deal in most cases'
|
||||
|
||||
source ./server/.venv/bin/activate
|
||||
exec text-generation-launcher $@
|
||||
|
|
Loading…
Reference in New Issue