hf_text-generation-inference/.github/workflows
Daniël de Kok 22fb1be588
Fix cache block size for flash decoding (#2351)
* Fix cache block size for flash decoding

This seems to have been accidentally dropped during the TRT-LLM
PR rebase.

* Also run CI on changes to `backends`
2024-08-01 15:38:57 +02:00
..
autodocs.yaml Rebase TRT-llm (#2331) 2024-07-31 10:33:10 +02:00
build.yaml fix: fix buildkit config in ci 2024-07-29 09:25:56 +02:00
build_documentation.yaml New runner. Manual squash. (#2110) 2024-06-24 18:08:34 +02:00
build_pr_documentation.yaml New runner. Manual squash. (#2110) 2024-06-24 18:08:34 +02:00
ci_build.yaml Fix cache block size for flash decoding (#2351) 2024-08-01 15:38:57 +02:00
client-tests.yaml Removing IPEX_AVAIL. (#2115) 2024-06-25 13:20:57 +02:00
integration_tests.yaml Removing IPEX_AVAIL. (#2115) 2024-06-25 13:20:57 +02:00
load_test.yaml re-push to internal registry (#2242) 2024-07-20 05:06:40 +00:00
stale.yaml New runner. Manual squash. (#2110) 2024-06-24 18:08:34 +02:00
tests.yaml Removing IPEX_AVAIL. (#2115) 2024-06-25 13:20:57 +02:00
trufflehog.yaml New runner. Manual squash. (#2110) 2024-06-24 18:08:34 +02:00
upload_pr_documentation.yaml New runner. Manual squash. (#2110) 2024-06-24 18:08:34 +02:00