import requests from llm_server import opts vllm_info = """
Important: This endpoint is running vllm and not all Oobabooga parameters are supported.