Update speculation.md
This commit is contained in:
parent
b6922d48de
commit
960cc95a0e
|
@ -26,7 +26,7 @@ You can check a few existing fine-tunes for popular models:
|
||||||
- [text-generation-inference/Mistral-7B-Instruct-v0.2-medusa](https://huggingface.co/text-generation-inference/Mistral-7B-Instruct-v0.2-medusa)
|
- [text-generation-inference/Mistral-7B-Instruct-v0.2-medusa](https://huggingface.co/text-generation-inference/Mistral-7B-Instruct-v0.2-medusa)
|
||||||
|
|
||||||
|
|
||||||
In order to create your own medusa heads for your own finetune, you should check own the original medusa repo. [https://github.com/FasterDecoding/Medusa](https://github.com/FasterDecoding/Medusa)
|
In order to create your own medusa heads for your own finetune, you should check own the original medusa repo. [https://github.com/FasterDecoding/Medusa](https://github.com/FasterDecoding/Medusa/pull/83).
|
||||||
|
|
||||||
|
|
||||||
In order to use medusa models in TGI, simply point to a medusa enabled model, and everything will load automatically.
|
In order to use medusa models in TGI, simply point to a medusa enabled model, and everything will load automatically.
|
||||||
|
|
Loading…
Reference in New Issue