hf_text-generation-inference/server
OlivierDehaene b1482d9048
breaking(router): modify /generate API to only return generated text (#50)
@njhill, @yk FYI

generated_text was concatenated to the user prompt for legacy reason. We
want to remove this behaviour as we don't think it is useful and even
detrimonial to usability.

We also remove the unused Vec.
2023-02-02 15:02:04 +01:00
..
tests breaking(router): modify /generate API to only return generated text (#50) 2023-02-02 15:02:04 +01:00
text_generation breaking(router): modify /generate API to only return generated text (#50) 2023-02-02 15:02:04 +01:00
.gitignore feat(server): Support all AutoModelForCausalLM on a best effort basis 2022-10-28 19:24:00 +02:00
Makefile fix(dockerfile): fix docker build (#32) 2023-01-24 19:52:39 +01:00
README.md feat(server): Use safetensors 2022-10-22 20:00:15 +02:00
poetry.lock fix(server): fix seeding with multiple shards (#44) 2023-01-31 16:01:15 +01:00
pyproject.toml fix(server): fix seeding with multiple shards (#44) 2023-01-31 16:01:15 +01:00

README.md

BLOOM Inference Python gRPC Server

A Python gRPC server for BLOOM Inference

Install

make install

Run

make run-dev