From 95d781725ef4562a1232950de3e73f74155c6ff8 Mon Sep 17 00:00:00 2001 From: Cyberes Date: Wed, 4 Oct 2023 12:42:18 -0600 Subject: [PATCH] t --- llm_server/routes/queue.py | 2 +- llm_server/workers/moderator.py | 1 - server.py | 2 -- 3 files changed, 1 insertion(+), 4 deletions(-) diff --git a/llm_server/routes/queue.py b/llm_server/routes/queue.py index b9be18d..e374311 100644 --- a/llm_server/routes/queue.py +++ b/llm_server/routes/queue.py @@ -50,7 +50,7 @@ class RedisPriorityQueue: b = self.redis.hget('queued_ip_count', item[1]) self.decrement_ip_count(client_ip, 'queued_ip_count') a = self.redis.hget('queued_ip_count', item[1]) - + print(item[1], a, b) return item diff --git a/llm_server/workers/moderator.py b/llm_server/workers/moderator.py index 64ba527..16ad0a2 100644 --- a/llm_server/workers/moderator.py +++ b/llm_server/workers/moderator.py @@ -52,7 +52,6 @@ def moderation_worker(): _, categories = check_moderation_endpoint(msg) redis_moderation.rpush('queue:flagged_categories', json.dumps((tag, categories))) except: - print(result) traceback.print_exc() continue diff --git a/server.py b/server.py index a3074e6..10c5b8c 100644 --- a/server.py +++ b/server.py @@ -97,8 +97,6 @@ create_db() @app.route('/api/openai') @flask_cache.cached(timeout=10) def home(): - print(len(priority_queue)) - base_client_api = redis.get('base_client_api', dtype=str) stats = generate_stats() model_choices, default_model = get_model_choices()