diff --git a/.github/workflows/tests.yaml b/.github/workflows/tests.yaml index 27caff1..476ae44 100644 --- a/.github/workflows/tests.yaml +++ b/.github/workflows/tests.yaml @@ -60,7 +60,7 @@ jobs: - name: Run server tests run: | pip install pytest - pytest -sv server/tests + HF_HUB_ENABLE_HF_TRANSFER=1 pytest -sv server/tests - name: Run Rust tests run: | cargo test diff --git a/Makefile b/Makefile index 0821773..cf7b867 100644 --- a/Makefile +++ b/Makefile @@ -19,7 +19,7 @@ integration-tests: install-router install-launcher cargo test python-tests: - cd server && pytest tests + cd server && HF_HUB_ENABLE_HF_TRANSFER=1 pytest tests run-bloom-560m: text-generation-launcher --model-id bigscience/bloom-560m --num-shard 2 diff --git a/server/Makefile b/server/Makefile index 74ce514..4a77dbc 100644 --- a/server/Makefile +++ b/server/Makefile @@ -1,3 +1,5 @@ +transformers_commit := 712d62e83c28236c7f39af690e7792a54288dbd9 + gen-server: # Compile protos pip install grpcio-tools==1.51.1 --no-cache-dir @@ -10,11 +12,11 @@ install-transformers: # Install specific version of transformers with custom cuda kernels pip uninstall transformers -y || true rm -rf transformers || true - rm -rf transformers-text_generation_inference || true - curl -L -O https://github.com/OlivierDehaene/transformers/archive/refs/heads/text_generation_inference.zip - unzip text_generation_inference.zip - rm text_generation_inference.zip - mv transformers-text_generation_inference transformers + rm -rf transformers-$(transformers_commit) || true + curl -L -O https://github.com/OlivierDehaene/transformers/archive/$(transformers_commit).zip + unzip $(transformers_commit).zip + rm $(transformers_commit).zip + mv transformers-$(transformers_commit) transformers cd transformers && python setup.py install install-torch: