diff --git a/llm_server/llm/vllm/info.py b/llm_server/llm/vllm/info.py index e873a30..996c614 100644 --- a/llm_server/llm/vllm/info.py +++ b/llm_server/llm/vllm/info.py @@ -1,8 +1,10 @@ -vllm_info = """

Important: This endpoint is running vllm-gptq and not all Oobabooga parameters are supported.

+vllm_info = """

Important: This endpoint is running vllm and not all Oobabooga parameters are supported.

Supported Parameters: """ \ No newline at end of file diff --git a/llm_server/llm/vllm/tokenize.py b/llm_server/llm/vllm/tokenize.py index 9913603..dd9553b 100644 --- a/llm_server/llm/vllm/tokenize.py +++ b/llm_server/llm/vllm/tokenize.py @@ -14,6 +14,5 @@ def tokenize(prompt: str) -> int: j = r.json() return j['length'] except: - # Fall back to whatever the superclass is doing. print(traceback.format_exc()) return len(tokenizer.encode(prompt)) + 10