This repository has been archived on 2024-10-27. You can view files and clone it, but cannot push or open issues or pull requests.
2023-09-14 01:32:49 -06:00
|
|
|
[Unit]
|
|
|
|
Description=VLLM Backend
|
|
|
|
Wants=basic.target
|
|
|
|
After=basic.target network.target
|
|
|
|
|
|
|
|
[Service]
|
2024-03-18 12:42:44 -06:00
|
|
|
User=vllm
|
|
|
|
ExecStart=/storage/vllm/vllm-venv/bin/python3.11 /storage/vllm/api_server.py --model /storage/models/awq/MythoMax-L2-13B-AWQ --quantization awq --host 0.0.0.0 --port 7000 --gpu-memory-utilization 0.95 --max-log-len 100
|
2023-09-14 01:32:49 -06:00
|
|
|
Restart=always
|
|
|
|
RestartSec=2
|
|
|
|
|
|
|
|
[Install]
|
|
|
|
WantedBy=multi-user.target
|
2024-03-18 12:42:44 -06:00
|
|
|
|