From 7acaa3c885e4f418849138ba53a49c06065c96f8 Mon Sep 17 00:00:00 2001 From: Cyberes Date: Wed, 4 Oct 2023 13:32:54 -0600 Subject: [PATCH] g --- llm_server/llm/vllm/tokenize.py | 1 - 1 file changed, 1 deletion(-) diff --git a/llm_server/llm/vllm/tokenize.py b/llm_server/llm/vllm/tokenize.py index d5a1b71..1e3c2a1 100644 --- a/llm_server/llm/vllm/tokenize.py +++ b/llm_server/llm/vllm/tokenize.py @@ -28,7 +28,6 @@ def tokenize(prompt: str, backend_url: str) -> int: return j['length'] except Exception as e: print(f'Failed to tokenize using VLLM -', f'{e.__class__.__name__}: {e}') - raise Exception return len(tokenizer.encode(chunk)) + 10 # Use a ThreadPoolExecutor to send all chunks to the server at once