From cef88b866a32432af5d2ef40524885ab14ef200a Mon Sep 17 00:00:00 2001 From: Cyberes Date: Tue, 29 Aug 2023 15:52:58 -0600 Subject: [PATCH] fix wrong response status code --- llm_server/routes/v1/generate.py | 11 +++++++++-- 1 file changed, 9 insertions(+), 2 deletions(-) diff --git a/llm_server/routes/v1/generate.py b/llm_server/routes/v1/generate.py index 9660360..fe38426 100644 --- a/llm_server/routes/v1/generate.py +++ b/llm_server/routes/v1/generate.py @@ -78,6 +78,12 @@ def generate(): end_time = time.time() elapsed_time = end_time - start_time + # Be extra careful when getting attributes from the response object + try: + response_status_code = response.status_code + except: + response_status_code = 0 + # TODO: why is this if block sitting here if (not success or not response) and opts.mode == 'oobabooga': # Ooba doesn't return any error messages @@ -116,7 +122,7 @@ def generate(): error_type_string = 'returned an error' if opts.mode == 'oobabooga' else f'returned {indefinite_article(error_type)} {error_type} error' backend_response = format_sillytavern_err( f'Backend ({opts.mode}) {error_type_string}: {response_json_body.get("error")}', - 'error') + f'HTTP CODE {response_status_code}') response_json_body = { 'results': [ { @@ -136,7 +142,8 @@ def generate(): raise Exception if not backend_err: redis.incr('proompts') - log_prompt(client_ip, token, request_json_body['prompt'], backend_response, elapsed_time if not backend_err else None, parameters, dict(request.headers), response.status_code if response else 0, response_json_body.get('details', {}).get('generated_tokens'), is_error=backend_err) + + log_prompt(client_ip, token, request_json_body['prompt'], backend_response, elapsed_time if not backend_err else None, parameters, dict(request.headers), response_status_code, response_json_body.get('details', {}).get('generated_tokens'), is_error=backend_err) return jsonify({ **response_json_body }), 200