From fd5796ed07fd9a9fe9f5dd1a39be1d5a85d8ec34 Mon Sep 17 00:00:00 2001 From: Cyberes Date: Wed, 23 Aug 2023 16:08:52 -0600 Subject: [PATCH] oops --- llm_server/routes/v1/proxy.py | 5 ++--- 1 file changed, 2 insertions(+), 3 deletions(-) diff --git a/llm_server/routes/v1/proxy.py b/llm_server/routes/v1/proxy.py index 207ba92..5b68d9b 100644 --- a/llm_server/routes/v1/proxy.py +++ b/llm_server/routes/v1/proxy.py @@ -8,7 +8,6 @@ from llm_server.routes.v1.generate import concurrent_semaphore from . import bp from .. import stats from ..cache import cache -from ..helpers.http import cache_control from ..stats import SemaphoreCheckerThread from ...llm.info import get_running_model @@ -16,7 +15,7 @@ from ...llm.info import get_running_model @bp.route('/stats', methods=['GET']) @cache.cached(timeout=5, query_string=True) def get_stats(): - model_list = get_running_model() # will return False when the fetch fails + model_list, error = get_running_model() # will return False when the fetch fails if isinstance(model_list, bool): online = False else: @@ -31,7 +30,7 @@ def get_stats(): }, 'online': online, 'mode': opts.mode, - 'model': get_running_model(), + 'model': model_list, 'endpoints': { 'blocking': f'https://{request.headers.get("Host")}/{opts.frontend_api_client.strip("/")}', },