diff --git a/llm_server/llm/vllm/info.py b/llm_server/llm/vllm/info.py
index e873a30..996c614 100644
--- a/llm_server/llm/vllm/info.py
+++ b/llm_server/llm/vllm/info.py
@@ -1,8 +1,10 @@
-vllm_info = """
Important: This endpoint is running vllm-gptq and not all Oobabooga parameters are supported.
+vllm_info = """Important: This endpoint is running vllm and not all Oobabooga parameters are supported.
Supported Parameters:
- temperature
- top_p
- top_k
- max_new_tokens
+- num_beams (setting to greater than 1 enables beam search)
+- ban_eos_token
"""
\ No newline at end of file
diff --git a/llm_server/llm/vllm/tokenize.py b/llm_server/llm/vllm/tokenize.py
index 9913603..dd9553b 100644
--- a/llm_server/llm/vllm/tokenize.py
+++ b/llm_server/llm/vllm/tokenize.py
@@ -14,6 +14,5 @@ def tokenize(prompt: str) -> int:
j = r.json()
return j['length']
except:
- # Fall back to whatever the superclass is doing.
print(traceback.format_exc())
return len(tokenizer.encode(prompt)) + 10