Putting the fix for vllm for CIt
This commit is contained in:
parent
512ed5ca4c
commit
81704d28e8
|
@ -145,7 +145,7 @@ jobs:
|
||||||
pytest -s -vv integration-tests
|
pytest -s -vv integration-tests
|
||||||
integration-tests-rocm:
|
integration-tests-rocm:
|
||||||
concurrency:
|
concurrency:
|
||||||
group: ${{ github.workflow }}-${{ github.job }}-${{ github.head_ref || github.run_id }}
|
group: ${{ github.workflow }}-${{ github.job }}-rocm-${{ github.head_ref || github.run_id }}
|
||||||
cancel-in-progress: true
|
cancel-in-progress: true
|
||||||
runs-on: [amd-gpu-tgi, multi-gpu, mi250]
|
runs-on: [amd-gpu-tgi, multi-gpu, mi250]
|
||||||
needs:
|
needs:
|
||||||
|
|
|
@ -1,5 +1,5 @@
|
||||||
commit_cuda := b5dfc61db88a81069e45b44f7cc99bd9e62a60fa
|
commit_cuda := b5dfc61db88a81069e45b44f7cc99bd9e62a60fa
|
||||||
commit_rocm := ca6913b3c2ffacdcb7d15e914dc34adbc6c89479
|
commit_rocm := 559200c1a028de990c1ddea761b0ccd62109e3a0
|
||||||
build-vllm-cuda:
|
build-vllm-cuda:
|
||||||
if [ ! -d 'vllm' ]; then \
|
if [ ! -d 'vllm' ]; then \
|
||||||
pip install -U ninja packaging --no-cache-dir && \
|
pip install -U ninja packaging --no-cache-dir && \
|
||||||
|
|
Loading…
Reference in New Issue