hf_text-generation-inference/server/tests/conftest.py

42 lines
988 B
Python
Raw Normal View History

2022-12-08 10:49:33 -07:00
import pytest
from transformers import AutoTokenizer
from text_generation.pb import generate_pb2
@pytest.fixture
def default_pb_parameters():
2022-12-15 09:03:56 -07:00
return generate_pb2.NextTokenChooserParameters(
2022-12-08 10:49:33 -07:00
temperature=1.0,
top_k=0,
top_p=1.0,
do_sample=False,
)
2022-12-12 10:25:22 -07:00
@pytest.fixture
def default_pb_stop_parameters():
return generate_pb2.StoppingCriteriaParameters(stop_sequences=[], max_new_tokens=10)
2022-12-08 10:49:33 -07:00
@pytest.fixture(scope="session")
def bloom_560m_tokenizer():
return AutoTokenizer.from_pretrained("bigscience/bloom-560m", padding_side="left")
@pytest.fixture(scope="session")
def gpt2_tokenizer():
tokenizer = AutoTokenizer.from_pretrained("gpt2", padding_side="left")
tokenizer.pad_token_id = 50256
return tokenizer
@pytest.fixture(scope="session")
def mt0_small_tokenizer():
tokenizer = AutoTokenizer.from_pretrained(
"bigscience/mt0-small", padding_side="left"
)
tokenizer.bos_token_id = 0
return tokenizer