From 5540112607c368fcd8bdb81ab8297e4adda2e381 Mon Sep 17 00:00:00 2001 From: Cyberes Date: Thu, 5 Oct 2023 19:09:25 -0600 Subject: [PATCH] t --- llm_server/routes/queue.py | 2 ++ llm_server/workers/inferencer.py | 2 ++ 2 files changed, 4 insertions(+) diff --git a/llm_server/routes/queue.py b/llm_server/routes/queue.py index e4c49c5..c8cf2e6 100644 --- a/llm_server/routes/queue.py +++ b/llm_server/routes/queue.py @@ -36,6 +36,8 @@ class RedisPriorityQueue: print(f'Rejecting request from {item[1]} - {ip_count} requests in progress.') return None # reject the request + print('--->', event.event_id) + self.redis.zadd('queue', {json.dumps((item, event.event_id, selected_model)): -priority}) self.increment_ip_count(item[1], 'queued_ip_count') return event diff --git a/llm_server/workers/inferencer.py b/llm_server/workers/inferencer.py index e023c86..5e1232f 100644 --- a/llm_server/workers/inferencer.py +++ b/llm_server/workers/inferencer.py @@ -33,6 +33,8 @@ def worker(): increment_ip_count(client_ip, 'processing_ips') incr_active_workers(selected_model, backend_url) + print('<---', event_id.event_id) + if not request_json_body: # This was a dummy request from the websocket handlers. # We're going to let the websocket handler decrement