FIxing the CI.
This commit is contained in:
parent
c7ab1810d4
commit
cb0a29484d
|
@ -2094,4 +2094,4 @@
|
||||||
"description": "Hugging Face Text Generation Inference API"
|
"description": "Hugging Face Text Generation Inference API"
|
||||||
}
|
}
|
||||||
]
|
]
|
||||||
}
|
}
|
||||||
|
|
|
@ -88,7 +88,7 @@ There is also an async version of the client, `AsyncInferenceClient`, based on `
|
||||||
|
|
||||||
You can directly use the OpenAI [Python](https://github.com/openai/openai-python) or [JS](https://github.com/openai/openai-node) clients to interact with TGI.
|
You can directly use the OpenAI [Python](https://github.com/openai/openai-python) or [JS](https://github.com/openai/openai-node) clients to interact with TGI.
|
||||||
|
|
||||||
Install the OpenAI Python package via pip.
|
Install the OpenAI Python package via pip.
|
||||||
|
|
||||||
```bash
|
```bash
|
||||||
pip install openai
|
pip install openai
|
||||||
|
@ -145,7 +145,7 @@ def inference(message, history):
|
||||||
stream=True,
|
stream=True,
|
||||||
max_tokens=1024,
|
max_tokens=1024,
|
||||||
)
|
)
|
||||||
|
|
||||||
for chunk in output:
|
for chunk in output:
|
||||||
partial_message += chunk.choices[0].delta.content
|
partial_message += chunk.choices[0].delta.content
|
||||||
yield partial_message
|
yield partial_message
|
||||||
|
@ -196,4 +196,4 @@ To serve both ChatUI and TGI in same environment, simply add your own endpoints
|
||||||
}
|
}
|
||||||
```
|
```
|
||||||
|
|
||||||
![ChatUI](https://huggingface.co/datasets/huggingface/documentation-images/resolve/main/chatui_screen.png)
|
![ChatUI](https://huggingface.co/datasets/huggingface/documentation-images/resolve/main/chatui_screen.png)
|
||||||
|
|
Loading…
Reference in New Issue