This repository has been archived on 2024-10-27. You can view files and clone it, but cannot push or open issues or pull requests.
local-llm-server/llm_server/routes/ooba_request_handler.py

58 lines
2.5 KiB
Python
Raw Normal View History

from typing import Tuple
2023-09-12 16:40:09 -06:00
import flask
2023-09-27 14:48:47 -06:00
from flask import jsonify, request
2023-09-12 16:40:09 -06:00
from llm_server import opts
from llm_server.database.database import do_db_log
from llm_server.database.log_to_db import log_to_db
2023-09-12 16:40:09 -06:00
from llm_server.routes.helpers.client import format_sillytavern_err
from llm_server.routes.request_handler import RequestHandler
class OobaRequestHandler(RequestHandler):
def __init__(self, *args, **kwargs):
super().__init__(*args, **kwargs)
2023-10-01 16:04:53 -06:00
def handle_request(self, return_ok: bool = True):
assert not self.used
2023-10-04 13:18:47 -06:00
if self.offline:
2023-10-04 16:04:03 -06:00
msg = 'The model you requested is not a valid choice. Please retry your query.'
2023-10-04 13:18:47 -06:00
print(msg)
2023-10-04 13:21:43 -06:00
self.handle_error(msg)
2023-09-12 16:40:09 -06:00
request_valid, invalid_response = self.validate_request()
if not request_valid:
return invalid_response
2023-09-12 16:40:09 -06:00
# Reconstruct the request JSON with the validated parameters and prompt.
prompt = self.request_json_body.get('prompt', '')
llm_request = {**self.parameters, 'prompt': prompt}
_, backend_response = self.generate_response(llm_request)
2023-10-01 16:04:53 -06:00
if return_ok:
# Always return 200 so ST displays our error messages
return backend_response[0], 200
else:
# The OpenAI route needs to detect 429 errors.
return backend_response
2023-09-12 16:40:09 -06:00
2023-09-28 01:34:15 -06:00
def handle_ratelimited(self, do_log: bool = True):
msg = f'Ratelimited: you are only allowed to have {opts.simultaneous_requests_per_ip} simultaneous requests at a time. Please complete your other requests before sending another.'
2023-09-27 14:48:47 -06:00
backend_response = self.handle_error(msg)
2023-09-28 01:34:15 -06:00
if do_log:
log_to_db(self.client_ip, self.token, self.request_json_body.get('prompt', ''), backend_response[0].data.decode('utf-8'), None, self.parameters, dict(self.request.headers), 429, self.request.url, self.backend_url, is_error=True)
2023-10-01 16:04:53 -06:00
return backend_response[0], 429
2023-09-27 14:48:47 -06:00
def handle_error(self, error_msg: str, error_type: str = 'error') -> Tuple[flask.Response, int]:
disable_st_error_formatting = request.headers.get('LLM-ST-Errors', False) == 'true'
if disable_st_error_formatting:
2023-09-27 14:48:47 -06:00
# TODO: how to format this
response_msg = error_msg
else:
2023-09-30 19:41:50 -06:00
response_msg = format_sillytavern_err(error_msg, error_type=error_type, backend_url=self.backend_url)
return jsonify({
2023-09-27 14:48:47 -06:00
'results': [{'text': response_msg}]
2023-09-27 14:36:49 -06:00
}), 200 # return 200 so we don't trigger an error message in the client's ST