2023-09-20 20:30:31 -06:00
|
|
|
import traceback
|
|
|
|
|
|
|
|
import requests
|
|
|
|
import tiktoken
|
|
|
|
|
|
|
|
from llm_server import opts
|
|
|
|
|
|
|
|
tokenizer = tiktoken.get_encoding("cl100k_base")
|
|
|
|
|
|
|
|
|
|
|
|
def tokenize(prompt: str) -> int:
|
2023-09-25 22:32:48 -06:00
|
|
|
if not prompt:
|
|
|
|
# The tokenizers have issues when the prompt is None.
|
|
|
|
return 0
|
2023-09-20 20:30:31 -06:00
|
|
|
try:
|
|
|
|
r = requests.post(f'{opts.backend_url}/tokenize', json={'input': prompt}, verify=opts.verify_ssl, timeout=opts.backend_generate_request_timeout)
|
|
|
|
j = r.json()
|
|
|
|
return j['length']
|
|
|
|
except:
|
2023-09-25 18:18:29 -06:00
|
|
|
traceback.print_exc()
|
2023-09-20 20:30:31 -06:00
|
|
|
return len(tokenizer.encode(prompt)) + 10
|