0c9b6cdd76
* Choosing input/total tokens automatically based on available VRAM? * Update doc. * Remove generated files. * Trying to fix non chunking targets. * Attempt #2 * fix. * QuantLinear is rocm compatible. * Much simpler logic after the overhead. * Updating logic + non flash. * Revert doc text. * Simple updates. * Fix integration mt0 (transformers update). |
||
---|---|---|
.. | ||
source | ||
README.md | ||
index.html | ||
openapi.json |
README.md
Documentation available at: https://huggingface.co/docs/text-generation-inference
Release
When making a release, please update the latest version in the documentation with:
export OLD_VERSION="2\.0\.3"
export NEW_VERSION="2\.0\.4"
find . -name '*.md' -exec sed -i -e "s/$OLD_VERSION/$NEW_VERSION/g" {} \;