adjust netdata json, don't log error messages during generationg

This commit is contained in:
Cyberes 2023-08-24 22:53:06 -06:00
parent 06173f900e
commit e5aca7b09d
2 changed files with 5 additions and 3 deletions

View File

@ -26,4 +26,4 @@ def get_power_states():
print('Failed to fetch Netdata metrics:', e)
return output
gpu_num += 1
return output
return {'power_states': output}

View File

@ -82,19 +82,21 @@ def generate():
**response_json_body
}), 200
response_valid_json, response_json_body = validate_json(response)
backend_err = False
if response_valid_json:
redis.incr('proompts')
backend_response = safe_list_get(response_json_body.get('results', []), 0, {}).get('text')
if not backend_response:
if opts.mode == 'oobabooga':
backend_err = True
backend_response = format_sillytavern_err(
f'Backend (oobabooga) returned an empty string. This can happen when your parameters are incorrect. Make sure your context size is no greater than {opts.context_size}. Furthermore, oobabooga does not support concurrent requests so all users have to wait in line and the backend server may have glitched for a moment. Please try again.',
f'Backend (oobabooga) returned an empty string. This is usually due to an error on the backend during inference. Make sure your context size is no greater than {opts.context_size}. Please try again.',
'error')
response_json_body['results'][0]['text'] = backend_response
else:
raise Exception
log_prompt(client_ip, token, request_json_body['prompt'], backend_response, elapsed_time, parameters, dict(request.headers), response.status_code)
log_prompt(client_ip, token, request_json_body['prompt'], backend_response if not backend_err else None, elapsed_time, parameters, dict(request.headers), response.status_code)
return jsonify({
**response_json_body
}), 200