local-llm-server/llm_server/llm/vllm
Cyberes 347a82b7e1 avoid sending to backend to tokenize if it's greater than our specified context size 2023-09-28 03:54:20 -06:00
..
__init__.py port to mysql, use vllm tokenizer endpoint 2023-09-20 20:30:31 -06:00
generate.py fix error handling 2023-09-27 14:36:49 -06:00
info.py adjust vllm info 2023-09-21 20:13:29 -06:00
tokenize.py avoid sending to backend to tokenize if it's greater than our specified context size 2023-09-28 03:54:20 -06:00
vllm_backend.py fix background log not doing anything 2023-09-25 18:18:29 -06:00