feat: add simple idefics3 test
This commit is contained in:
parent
f9dfd36b92
commit
35c64b267a
|
@ -338,6 +338,7 @@ def launcher(event_loop):
|
|||
dtype: Optional[str] = None,
|
||||
revision: Optional[str] = None,
|
||||
max_input_length: Optional[int] = None,
|
||||
max_input_tokens: Optional[int] = None,
|
||||
max_batch_prefill_tokens: Optional[int] = None,
|
||||
max_total_tokens: Optional[int] = None,
|
||||
lora_adapters: Optional[List[str]] = None,
|
||||
|
@ -383,6 +384,9 @@ def launcher(event_loop):
|
|||
if max_input_length:
|
||||
args.append("--max-input-length")
|
||||
args.append(str(max_input_length))
|
||||
if max_input_tokens:
|
||||
args.append("--max-input-tokens")
|
||||
args.append(str(max_input_tokens))
|
||||
if max_batch_prefill_tokens:
|
||||
args.append("--max-batch-prefill-tokens")
|
||||
args.append(str(max_batch_prefill_tokens))
|
||||
|
|
|
@ -0,0 +1,73 @@
|
|||
{
|
||||
"details": {
|
||||
"best_of_sequences": null,
|
||||
"finish_reason": "length",
|
||||
"generated_tokens": 10,
|
||||
"prefill": [],
|
||||
"seed": null,
|
||||
"tokens": [
|
||||
{
|
||||
"id": 578,
|
||||
"logprob": -0.2475586,
|
||||
"special": false,
|
||||
"text": " The"
|
||||
},
|
||||
{
|
||||
"id": 2217,
|
||||
"logprob": -0.017303467,
|
||||
"special": false,
|
||||
"text": " image"
|
||||
},
|
||||
{
|
||||
"id": 62991,
|
||||
"logprob": -0.7368164,
|
||||
"special": false,
|
||||
"text": " depicts"
|
||||
},
|
||||
{
|
||||
"id": 279,
|
||||
"logprob": -0.39990234,
|
||||
"special": false,
|
||||
"text": " the"
|
||||
},
|
||||
{
|
||||
"id": 89675,
|
||||
"logprob": -0.34350586,
|
||||
"special": false,
|
||||
"text": " Statue"
|
||||
},
|
||||
{
|
||||
"id": 315,
|
||||
"logprob": -0.0002901554,
|
||||
"special": false,
|
||||
"text": " of"
|
||||
},
|
||||
{
|
||||
"id": 32492,
|
||||
"logprob": -0.0009598732,
|
||||
"special": false,
|
||||
"text": " Liberty"
|
||||
},
|
||||
{
|
||||
"id": 11,
|
||||
"logprob": -0.2355957,
|
||||
"special": false,
|
||||
"text": ","
|
||||
},
|
||||
{
|
||||
"id": 264,
|
||||
"logprob": -0.66503906,
|
||||
"special": false,
|
||||
"text": " a"
|
||||
},
|
||||
{
|
||||
"id": 97937,
|
||||
"logprob": -0.9199219,
|
||||
"special": false,
|
||||
"text": " colossal"
|
||||
}
|
||||
],
|
||||
"top_tokens": null
|
||||
},
|
||||
"generated_text": " The image depicts the Statue of Liberty, a colossal"
|
||||
}
|
|
@ -2,23 +2,19 @@ import pytest
|
|||
import base64
|
||||
|
||||
|
||||
# TODO fix the server parsser to count inline image tokens correctly
|
||||
def get_chicken():
|
||||
with open("integration-tests/images/chicken_on_money.png", "rb") as image_file:
|
||||
encoded_string = base64.b64encode(image_file.read())
|
||||
return f"data:image/png;base64,{encoded_string.decode('utf-8')}"
|
||||
|
||||
|
||||
def get_cow_beach():
|
||||
with open("integration-tests/images/cow_beach.png", "rb") as image_file:
|
||||
encoded_string = base64.b64encode(image_file.read())
|
||||
return f"data:image/png;base64,{encoded_string.decode('utf-8')}"
|
||||
|
||||
|
||||
@pytest.fixture(scope="module")
|
||||
def flash_idefics3_next_handle(launcher):
|
||||
with launcher(
|
||||
"HuggingFaceM4/Idefics3-8B-Llama3",
|
||||
max_total_tokens=3000,
|
||||
max_batch_prefill_tokens=2501,
|
||||
max_input_tokens=2500,
|
||||
) as handle:
|
||||
yield handle
|
||||
|
||||
|
@ -29,76 +25,40 @@ async def flash_idefics3_next(flash_idefics3_next_handle):
|
|||
return flash_idefics3_next_handle.client
|
||||
|
||||
|
||||
# TODO: dont skip when token issue is resolved
|
||||
@pytest.mark.skip
|
||||
@pytest.mark.asyncio
|
||||
@pytest.mark.private
|
||||
async def test_flash_idefics3_next_simple(flash_idefics3_next, response_snapshot):
|
||||
async def test_flash_idefics3_next_simple_base64(
|
||||
flash_idefics3_next, response_snapshot
|
||||
):
|
||||
chicken = get_chicken()
|
||||
query = "Write me a short story"
|
||||
response = await flash_idefics3_next.generate(
|
||||
f"User:![]({chicken})Write me a short story<end_of_utterance> \nAssistant:",
|
||||
f"<|begin_of_text|><|begin_of_text|>User:![]({chicken}){query}<end_of_utterance>\nAssistant:",
|
||||
max_new_tokens=10,
|
||||
)
|
||||
assert (
|
||||
response.generated_text == " A chicken is sitting on a pile of money."
|
||||
), f"{repr(response.generated_text)}"
|
||||
assert response.details.generated_tokens == 10
|
||||
assert response == response_snapshot
|
||||
# assert response.details.generated_tokens == 10
|
||||
# assert response == response_snapshot
|
||||
|
||||
|
||||
@pytest.mark.asyncio
|
||||
@pytest.mark.private
|
||||
async def test_flash_idefics3_two_images(flash_idefics3_next, response_snapshot):
|
||||
chicken = get_chicken()
|
||||
cow_beach = get_cow_beach()
|
||||
async def test_flash_idefics3_next_simple_url(flash_idefics3_next, response_snapshot):
|
||||
ny_skyline = "https://cdn.britannica.com/61/93061-050-99147DCE/Statue-of-Liberty-Island-New-York-Bay.jpg"
|
||||
query = "What is in this image?"
|
||||
response = await flash_idefics3_next.generate(
|
||||
f"User:![]({chicken})![]({cow_beach})Where are the cow and chicken?<end_of_utterance> \nAssistant:",
|
||||
max_new_tokens=20,
|
||||
f"<|begin_of_text|><|begin_of_text|>User:![]({ny_skyline}){query}<end_of_utterance>\nAssistant:",
|
||||
max_new_tokens=10,
|
||||
seed=1337,
|
||||
)
|
||||
print(response)
|
||||
assert (
|
||||
response.generated_text
|
||||
== " The cow is standing on the beach and the chicken is sitting on a pile of money."
|
||||
== " The image depicts the Statue of Liberty, a colossal"
|
||||
), f"{repr(response.generated_text)}"
|
||||
assert response.details.generated_tokens == 19
|
||||
assert response == response_snapshot
|
||||
|
||||
|
||||
@pytest.mark.asyncio
|
||||
@pytest.mark.private
|
||||
async def test_flash_idefics3_next_all_params(flash_idefics3_next, response_snapshot):
|
||||
response = await flash_idefics3_next.generate(
|
||||
"Test request",
|
||||
max_new_tokens=10,
|
||||
repetition_penalty=1.2,
|
||||
return_full_text=True,
|
||||
stop_sequences=["test"],
|
||||
temperature=0.5,
|
||||
top_p=0.9,
|
||||
top_k=10,
|
||||
truncate=5,
|
||||
typical_p=0.9,
|
||||
watermark=True,
|
||||
decoder_input_details=True,
|
||||
seed=0,
|
||||
)
|
||||
|
||||
assert response.details.generated_tokens == 10
|
||||
assert response == response_snapshot
|
||||
|
||||
|
||||
@pytest.mark.asyncio
|
||||
@pytest.mark.private
|
||||
async def test_flash_idefics3_next_load(
|
||||
flash_idefics3_next, generate_load, response_snapshot
|
||||
):
|
||||
chicken = get_chicken()
|
||||
responses = await generate_load(
|
||||
flash_idefics3_next,
|
||||
f"User:![]({chicken})Write me a short story<end_of_utterance> \nAssistant:",
|
||||
max_new_tokens=10,
|
||||
n=4,
|
||||
)
|
||||
generated_texts = [r.generated_text for r in responses]
|
||||
assert generated_texts[0] == " A chicken is sitting on a pile of money."
|
||||
assert len(generated_texts) == 4
|
||||
assert all([r.generated_text == generated_texts[0] for r in responses])
|
||||
|
||||
assert responses == response_snapshot
|
||||
|
|
Loading…
Reference in New Issue