From 06ae8adf0d79ec1526a428dfa9a91d6323175652 Mon Sep 17 00:00:00 2001 From: Cyberes Date: Tue, 22 Aug 2023 21:14:12 -0600 Subject: [PATCH] add backend name to error messages --- llm_server/opts.py | 2 +- llm_server/routes/v1/generate.py | 4 ++-- 2 files changed, 3 insertions(+), 3 deletions(-) diff --git a/llm_server/opts.py b/llm_server/opts.py index a20a364..9e961cb 100644 --- a/llm_server/opts.py +++ b/llm_server/opts.py @@ -1,4 +1,4 @@ -# Global settings that never change after startup +# Global settings that never change after startup/init running_model = 'none' concurrent_gens = 3 diff --git a/llm_server/routes/v1/generate.py b/llm_server/routes/v1/generate.py index 14c67b5..b8968ab 100644 --- a/llm_server/routes/v1/generate.py +++ b/llm_server/routes/v1/generate.py @@ -42,7 +42,7 @@ def generate(): success, response, error_msg = generator(request_json_body) if not success: if opts.mode == 'oobabooga': - backend_response = format_sillytavern_err(f'Failed to reach the backend: {error_msg}', 'error') + backend_response = format_sillytavern_err(f'Failed to reach the backend ({opts.mode}): {error_msg}', 'error') response_json_body = { 'results': [ { @@ -64,7 +64,7 @@ def generate(): backend_response = safe_list_get(response_json_body.get('results', []), 0, {}).get('text') if not backend_response: if opts.mode == 'oobabooga': - backend_response = format_sillytavern_err(f'Backend returned an empty string. This can happen when your parameters are incorrect. Make sure your context size is no greater than {opts.context_size}.', 'error') + backend_response = format_sillytavern_err(f'Backend ({opts.mode}) returned an empty string. This can happen when your parameters are incorrect. Make sure your context size is no greater than {opts.context_size}.', 'error') response_json_body['results'][0]['text'] = backend_response else: raise Exception