2023-03-07 10:52:22 -07:00
|
|
|
import pytest
|
|
|
|
|
|
|
|
from text_generation import Client, AsyncClient
|
|
|
|
from text_generation.errors import NotFoundError, ValidationError
|
2023-06-02 09:12:30 -06:00
|
|
|
from text_generation.types import FinishReason, InputToken
|
2023-03-07 10:52:22 -07:00
|
|
|
|
|
|
|
|
2023-03-09 08:05:33 -07:00
|
|
|
def test_generate(flan_t5_xxl_url, hf_headers):
|
|
|
|
client = Client(flan_t5_xxl_url, hf_headers)
|
2023-06-02 09:12:30 -06:00
|
|
|
response = client.generate("test", max_new_tokens=1, decoder_input_details=True)
|
2023-03-07 10:52:22 -07:00
|
|
|
|
2023-03-09 08:05:33 -07:00
|
|
|
assert response.generated_text == ""
|
2023-03-07 10:52:22 -07:00
|
|
|
assert response.details.finish_reason == FinishReason.Length
|
|
|
|
assert response.details.generated_tokens == 1
|
|
|
|
assert response.details.seed is None
|
|
|
|
assert len(response.details.prefill) == 1
|
2023-06-02 09:12:30 -06:00
|
|
|
assert response.details.prefill[0] == InputToken(id=0, text="<pad>", logprob=None)
|
2023-03-07 10:52:22 -07:00
|
|
|
assert len(response.details.tokens) == 1
|
2023-05-16 12:22:11 -06:00
|
|
|
assert response.details.tokens[0].id == 3
|
2023-06-02 09:12:30 -06:00
|
|
|
assert response.details.tokens[0].text == " "
|
2023-05-16 12:22:11 -06:00
|
|
|
assert not response.details.tokens[0].special
|
2023-03-07 10:52:22 -07:00
|
|
|
|
|
|
|
|
2024-01-29 09:02:57 -07:00
|
|
|
def test_generate_max_new_tokens_not_set(flan_t5_xxl_url, hf_headers):
|
|
|
|
client = Client(flan_t5_xxl_url, hf_headers)
|
|
|
|
response = client.generate("test", decoder_input_details=True)
|
|
|
|
|
|
|
|
assert response.generated_text != ""
|
|
|
|
assert response.details.finish_reason == FinishReason.EndOfSequenceToken
|
|
|
|
assert response.details.generated_tokens > 1
|
|
|
|
assert response.details.seed is None
|
|
|
|
assert len(response.details.prefill) == 1
|
|
|
|
assert response.details.prefill[0] == InputToken(id=0, text="<pad>", logprob=None)
|
|
|
|
assert len(response.details.tokens) > 1
|
|
|
|
assert response.details.tokens[0].id == 3
|
|
|
|
assert response.details.tokens[0].text == " "
|
|
|
|
assert not response.details.tokens[0].special
|
|
|
|
|
|
|
|
|
2023-03-09 08:05:33 -07:00
|
|
|
def test_generate_best_of(flan_t5_xxl_url, hf_headers):
|
|
|
|
client = Client(flan_t5_xxl_url, hf_headers)
|
2023-06-02 09:12:30 -06:00
|
|
|
response = client.generate(
|
|
|
|
"test", max_new_tokens=1, best_of=2, do_sample=True, decoder_input_details=True
|
|
|
|
)
|
2023-03-09 08:05:33 -07:00
|
|
|
|
|
|
|
assert response.details.seed is not None
|
|
|
|
assert response.details.best_of_sequences is not None
|
|
|
|
assert len(response.details.best_of_sequences) == 1
|
|
|
|
assert response.details.best_of_sequences[0].seed is not None
|
|
|
|
|
|
|
|
|
2023-03-07 10:52:22 -07:00
|
|
|
def test_generate_not_found(fake_url, hf_headers):
|
|
|
|
client = Client(fake_url, hf_headers)
|
|
|
|
with pytest.raises(NotFoundError):
|
|
|
|
client.generate("test")
|
|
|
|
|
|
|
|
|
|
|
|
def test_generate_validation_error(flan_t5_xxl_url, hf_headers):
|
|
|
|
client = Client(flan_t5_xxl_url, hf_headers)
|
|
|
|
with pytest.raises(ValidationError):
|
|
|
|
client.generate("test", max_new_tokens=10_000)
|
|
|
|
|
|
|
|
|
2023-03-09 08:05:33 -07:00
|
|
|
def test_generate_stream(flan_t5_xxl_url, hf_headers):
|
|
|
|
client = Client(flan_t5_xxl_url, hf_headers)
|
2023-03-07 10:52:22 -07:00
|
|
|
responses = [
|
|
|
|
response for response in client.generate_stream("test", max_new_tokens=1)
|
|
|
|
]
|
|
|
|
|
|
|
|
assert len(responses) == 1
|
|
|
|
response = responses[0]
|
|
|
|
|
2023-03-09 08:05:33 -07:00
|
|
|
assert response.generated_text == ""
|
2023-03-07 10:52:22 -07:00
|
|
|
assert response.details.finish_reason == FinishReason.Length
|
|
|
|
assert response.details.generated_tokens == 1
|
|
|
|
assert response.details.seed is None
|
|
|
|
|
|
|
|
|
|
|
|
def test_generate_stream_not_found(fake_url, hf_headers):
|
|
|
|
client = Client(fake_url, hf_headers)
|
|
|
|
with pytest.raises(NotFoundError):
|
|
|
|
list(client.generate_stream("test"))
|
|
|
|
|
|
|
|
|
|
|
|
def test_generate_stream_validation_error(flan_t5_xxl_url, hf_headers):
|
|
|
|
client = Client(flan_t5_xxl_url, hf_headers)
|
|
|
|
with pytest.raises(ValidationError):
|
|
|
|
list(client.generate_stream("test", max_new_tokens=10_000))
|
|
|
|
|
|
|
|
|
|
|
|
@pytest.mark.asyncio
|
2023-03-09 08:05:33 -07:00
|
|
|
async def test_generate_async(flan_t5_xxl_url, hf_headers):
|
|
|
|
client = AsyncClient(flan_t5_xxl_url, hf_headers)
|
2023-06-02 09:12:30 -06:00
|
|
|
response = await client.generate(
|
|
|
|
"test", max_new_tokens=1, decoder_input_details=True
|
|
|
|
)
|
2023-03-07 10:52:22 -07:00
|
|
|
|
2023-03-09 08:05:33 -07:00
|
|
|
assert response.generated_text == ""
|
2023-03-07 10:52:22 -07:00
|
|
|
assert response.details.finish_reason == FinishReason.Length
|
|
|
|
assert response.details.generated_tokens == 1
|
|
|
|
assert response.details.seed is None
|
|
|
|
assert len(response.details.prefill) == 1
|
2023-06-02 09:12:30 -06:00
|
|
|
assert response.details.prefill[0] == InputToken(id=0, text="<pad>", logprob=None)
|
2023-03-07 10:52:22 -07:00
|
|
|
assert len(response.details.tokens) == 1
|
2023-05-16 12:22:11 -06:00
|
|
|
assert response.details.tokens[0].id == 3
|
2023-06-02 09:12:30 -06:00
|
|
|
assert response.details.tokens[0].text == " "
|
2023-05-16 12:22:11 -06:00
|
|
|
assert not response.details.tokens[0].special
|
2023-03-07 10:52:22 -07:00
|
|
|
|
|
|
|
|
2023-04-27 11:16:35 -06:00
|
|
|
@pytest.mark.asyncio
|
|
|
|
async def test_generate_async_best_of(flan_t5_xxl_url, hf_headers):
|
|
|
|
client = AsyncClient(flan_t5_xxl_url, hf_headers)
|
|
|
|
response = await client.generate(
|
2023-06-02 09:12:30 -06:00
|
|
|
"test", max_new_tokens=1, best_of=2, do_sample=True, decoder_input_details=True
|
2023-04-27 11:16:35 -06:00
|
|
|
)
|
|
|
|
|
|
|
|
assert response.details.seed is not None
|
|
|
|
assert response.details.best_of_sequences is not None
|
|
|
|
assert len(response.details.best_of_sequences) == 1
|
|
|
|
assert response.details.best_of_sequences[0].seed is not None
|
|
|
|
|
|
|
|
|
2023-03-07 10:52:22 -07:00
|
|
|
@pytest.mark.asyncio
|
|
|
|
async def test_generate_async_not_found(fake_url, hf_headers):
|
|
|
|
client = AsyncClient(fake_url, hf_headers)
|
|
|
|
with pytest.raises(NotFoundError):
|
|
|
|
await client.generate("test")
|
|
|
|
|
|
|
|
|
|
|
|
@pytest.mark.asyncio
|
|
|
|
async def test_generate_async_validation_error(flan_t5_xxl_url, hf_headers):
|
|
|
|
client = AsyncClient(flan_t5_xxl_url, hf_headers)
|
|
|
|
with pytest.raises(ValidationError):
|
|
|
|
await client.generate("test", max_new_tokens=10_000)
|
|
|
|
|
|
|
|
|
|
|
|
@pytest.mark.asyncio
|
2023-03-09 08:05:33 -07:00
|
|
|
async def test_generate_stream_async(flan_t5_xxl_url, hf_headers):
|
|
|
|
client = AsyncClient(flan_t5_xxl_url, hf_headers)
|
2023-03-07 10:52:22 -07:00
|
|
|
responses = [
|
|
|
|
response async for response in client.generate_stream("test", max_new_tokens=1)
|
|
|
|
]
|
|
|
|
|
|
|
|
assert len(responses) == 1
|
|
|
|
response = responses[0]
|
|
|
|
|
2023-03-09 08:05:33 -07:00
|
|
|
assert response.generated_text == ""
|
2023-03-07 10:52:22 -07:00
|
|
|
assert response.details.finish_reason == FinishReason.Length
|
|
|
|
assert response.details.generated_tokens == 1
|
|
|
|
assert response.details.seed is None
|
|
|
|
|
|
|
|
|
|
|
|
@pytest.mark.asyncio
|
|
|
|
async def test_generate_stream_async_not_found(fake_url, hf_headers):
|
|
|
|
client = AsyncClient(fake_url, hf_headers)
|
|
|
|
with pytest.raises(NotFoundError):
|
|
|
|
async for _ in client.generate_stream("test"):
|
|
|
|
pass
|
|
|
|
|
|
|
|
|
|
|
|
@pytest.mark.asyncio
|
|
|
|
async def test_generate_stream_async_validation_error(flan_t5_xxl_url, hf_headers):
|
|
|
|
client = AsyncClient(flan_t5_xxl_url, hf_headers)
|
|
|
|
with pytest.raises(ValidationError):
|
|
|
|
async for _ in client.generate_stream("test", max_new_tokens=10_000):
|
|
|
|
pass
|