from llm_server.cluster.cluster_config import cluster_config from llm_server.llm import oobabooga, vllm def get_token_count(prompt: str, backend_url: str): backend_url = cluster_config.validate_backend(backend_url) backend_mode = cluster_config.get_backend(backend_url)['mode'] if backend_mode == 'vllm': return vllm.tokenize(prompt, backend_url) elif backend_mode == 'ooba': return oobabooga.tokenize(prompt) else: raise Exception(backend_mode)