hf_text-generation-inference/server
OlivierDehaene 2ad895a6cc
feat(server): allow gpt-neox models with odd vocab sizes to be sharded (#48)
2023-02-01 14:43:59 +01:00
..
tests feat(server): Support GPT-Neox (#39) 2023-01-31 18:53:56 +01:00
text_generation feat(server): allow gpt-neox models with odd vocab sizes to be sharded (#48) 2023-02-01 14:43:59 +01:00
.gitignore feat(server): Support all AutoModelForCausalLM on a best effort basis 2022-10-28 19:24:00 +02:00
Makefile fix(dockerfile): fix docker build (#32) 2023-01-24 19:52:39 +01:00
README.md feat(server): Use safetensors 2022-10-22 20:00:15 +02:00
poetry.lock fix(server): fix seeding with multiple shards (#44) 2023-01-31 16:01:15 +01:00
pyproject.toml fix(server): fix seeding with multiple shards (#44) 2023-01-31 16:01:15 +01:00

README.md

BLOOM Inference Python gRPC Server

A Python gRPC server for BLOOM Inference

Install

make install

Run

make run-dev