Skip to content

vllm chat

Options

--url

url of the running OpenAI-Compatible RESTful API server

Default: http://localhost:8000/v1

--model-name

The model name used in prompt completion, default to the first model in list models API call.

Default: None

--api-key

API key for OpenAI services. If provided, this api key will overwrite the api key obtained through environment variables.

Default: None

--system-prompt

The system prompt to be added to the chat template, used for models that support system prompts.

Default: None

-q, --quick

Send a single prompt as MESSAGE and print the response, then exit.

Default: None