Commit Graph

24 Commits

Author SHA1 Message Date
Cyberes 0771c2325c fix inference workers quitting when a backend is offline, start adding logging, improve tokenizer error handling 2023-10-23 17:24:20 -06:00
Cyberes 70cf6843e5 update requiorements 2023-10-16 23:12:41 -06:00
Cyberes 806e522d16 don't pickle streaming 2023-10-16 18:35:10 -06:00
Cyberes 94141b8ecf fix processing not being decremented on streaming, fix confusion over queue, adjust stop sequences 2023-10-02 20:53:08 -06:00
Cyberes f7e9687527 finish openai endpoints 2023-10-01 16:04:53 -06:00
Cyberes d203973e80 fix routes 2023-10-01 01:13:13 -06:00
Cyberes e6267a7d46 remove vllm from requirements.txt 2023-09-30 19:45:04 -06:00
Cyberes 114f36e709 functional 2023-09-30 19:41:50 -06:00
Cyberes e7b57cad7b set up cluster config and basic background workers 2023-09-28 18:40:24 -06:00
Cyberes e86a5182eb redo background processes, reorganize server.py 2023-09-27 23:36:44 -06:00
Cyberes d9bbcc42e6 more work on openai endpoint 2023-09-26 22:09:11 -06:00
Cyberes d2651756df update requirements.txt 2023-09-24 21:46:48 -06:00
Cyberes 84ea2f8891 handle when auth token is not enabled 2023-09-24 15:57:39 -06:00
Cyberes 8d6b2ce49c minor changes, add admin token auth system, add route to get backend info 2023-09-24 15:54:35 -06:00
Cyberes 03e3ec5490 port to mysql, use vllm tokenizer endpoint 2023-09-20 20:30:31 -06:00
Cyberes 40ac84aa9a actually we don't want to emulate openai 2023-09-12 01:04:11 -06:00
Cyberes 4c9d543eab implement vllm backend 2023-09-11 20:47:19 -06:00
Cyberes e45eafd286 update requirements.txt 2023-08-29 17:57:06 -06:00
Cyberes bf648f605f implement streaming for hf-textgen 2023-08-29 17:56:12 -06:00
Cyberes 1a4cb5f786 reorganize stats page again 2023-08-27 22:24:44 -06:00
Cyberes 33190e3cfe fix stats for real 2023-08-23 01:14:19 -06:00
Cyberes ad9a91f1b5 concurrent gens setting, online status 2023-08-22 00:26:46 -06:00
Cyberes 6e3ddab42e fix relative paths for db path 2023-08-21 23:07:12 -06:00
Cyberes 8cbf643fd3 MVP 2023-08-21 21:28:52 -06:00