From 81452ec64322d7b19d2ad5343921aaee3f7d7d08 Mon Sep 17 00:00:00 2001 From: Cyberes Date: Thu, 21 Sep 2023 20:13:29 -0600 Subject: [PATCH] adjust vllm info --- llm_server/llm/vllm/info.py | 4 +++- llm_server/llm/vllm/tokenize.py | 1 - 2 files changed, 3 insertions(+), 2 deletions(-) diff --git a/llm_server/llm/vllm/info.py b/llm_server/llm/vllm/info.py index e873a30..996c614 100644 --- a/llm_server/llm/vllm/info.py +++ b/llm_server/llm/vllm/info.py @@ -1,8 +1,10 @@ -vllm_info = """

Important: This endpoint is running vllm-gptq and not all Oobabooga parameters are supported.

+vllm_info = """

Important: This endpoint is running vllm and not all Oobabooga parameters are supported.

Supported Parameters: """ \ No newline at end of file diff --git a/llm_server/llm/vllm/tokenize.py b/llm_server/llm/vllm/tokenize.py index 9913603..dd9553b 100644 --- a/llm_server/llm/vllm/tokenize.py +++ b/llm_server/llm/vllm/tokenize.py @@ -14,6 +14,5 @@ def tokenize(prompt: str) -> int: j = r.json() return j['length'] except: - # Fall back to whatever the superclass is doing. print(traceback.format_exc()) return len(tokenizer.encode(prompt)) + 10