This repository has been archived on 2024-10-27. You can view files and clone it, but cannot push or open issues or pull requests.
local-llm-server/llm_server/llm/vllm
Cyberes 347a82b7e1 avoid sending to backend to tokenize if it's greater than our specified context size 2023-09-28 03:54:20 -06:00
..
__init__.py port to mysql, use vllm tokenizer endpoint 2023-09-20 20:30:31 -06:00
generate.py fix error handling 2023-09-27 14:36:49 -06:00
info.py adjust vllm info 2023-09-21 20:13:29 -06:00
tokenize.py avoid sending to backend to tokenize if it's greater than our specified context size 2023-09-28 03:54:20 -06:00
vllm_backend.py fix background log not doing anything 2023-09-25 18:18:29 -06:00