From baf2adfb694b908a0d9f184c03e659c1d7c07f39 Mon Sep 17 00:00:00 2001 From: drbh Date: Tue, 21 May 2024 21:31:26 +0000 Subject: [PATCH] fix: run pre-commit --- clients/python/text_generation/client.py | 2 +- clients/python/text_generation/types.py | 2 +- router/src/server.rs | 6 +++--- 3 files changed, 5 insertions(+), 5 deletions(-) diff --git a/clients/python/text_generation/client.py b/clients/python/text_generation/client.py index acc239ac..12966747 100644 --- a/clients/python/text_generation/client.py +++ b/clients/python/text_generation/client.py @@ -982,4 +982,4 @@ class AsyncClient: except ValidationError: # If we failed to parse the payload, then it is an error payload raise parse_error(resp.status, json_payload) - yield response \ No newline at end of file + yield response diff --git a/clients/python/text_generation/types.py b/clients/python/text_generation/types.py index 3436d94e..eb872ee6 100644 --- a/clients/python/text_generation/types.py +++ b/clients/python/text_generation/types.py @@ -453,4 +453,4 @@ class StreamResponse(BaseModel): # Inference API currently deployed model class DeployedModel(BaseModel): model_id: str - sha: str \ No newline at end of file + sha: str diff --git a/router/src/server.rs b/router/src/server.rs index d5c529ee..e7570ded 100644 --- a/router/src/server.rs +++ b/router/src/server.rs @@ -597,7 +597,7 @@ async fn completions( let span = tracing::Span::current(); metrics::increment_counter!("tgi_request_count"); - let CompletionRequest { + let CompletionRequest { max_tokens, seed, stop, @@ -642,13 +642,13 @@ async fn completions( } let generate_requests: Vec = req - .prompt + .prompt .iter() .map(|prompt| GenerateRequest { inputs: prompt.to_string(), parameters: GenerateParameters { best_of: None, - temperature: temperature, + temperature, repetition_penalty: req.repetition_penalty, frequency_penalty: req.frequency_penalty, top_k: None,