From 06b52c7648f602802b5ceeda152f404799d19dd0 Mon Sep 17 00:00:00 2001 From: Cyberes Date: Tue, 29 Aug 2023 14:53:03 -0600 Subject: [PATCH] forgot to remove a snippet --- llm_server/routes/v1/generate.py | 3 ++- 1 file changed, 2 insertions(+), 1 deletion(-) diff --git a/llm_server/routes/v1/generate.py b/llm_server/routes/v1/generate.py index cc02979..b025da6 100644 --- a/llm_server/routes/v1/generate.py +++ b/llm_server/routes/v1/generate.py @@ -78,6 +78,7 @@ def generate(): end_time = time.time() elapsed_time = end_time - start_time + # TODO: why is this if block sitting here if (not success or not response) and opts.mode == 'oobabooga': # Ooba doesn't return any error messages backend_response = format_sillytavern_err(f'Failed to reach the backend ({opts.mode}): {error_msg}', 'error') @@ -133,7 +134,7 @@ def generate(): else: raise Exception redis.incr('proompts') - log_prompt(client_ip, token, request_json_body['prompt'], backend_response if not backend_err else '', elapsed_time if not backend_err else None, parameters, dict(request.headers), response.status_code if response else 0, response_json_body.get('details', {}).get('generated_tokens'), is_error=backend_err) + log_prompt(client_ip, token, request_json_body['prompt'], backend_response, elapsed_time if not backend_err else None, parameters, dict(request.headers), response.status_code if response else 0, response_json_body.get('details', {}).get('generated_tokens'), is_error=backend_err) return jsonify({ **response_json_body }), 200