Skip to main content
Version: v0.4.0

remote::vllm

Description​

Remote vLLM inference provider for connecting to vLLM servers.

Configuration​

FieldTypeRequiredDefaultDescription
allowed_modelslist[str] | NoneNoList of models that should be registered with the model registry. If None, all models are allowed.
refresh_modelsboolNoFalseWhether to refresh models periodically from the provider
api_tokenSecretStr | NoneNoThe API token
base_urlHttpUrl | NoneNoThe URL for the vLLM model serving endpoint
max_tokensintNo4096Maximum number of tokens to generate.
tls_verifybool | strNoTrueWhether to verify TLS certificates. Can be a boolean or a path to a CA certificate file.

Sample Configuration​

base_url: ${env.VLLM_URL:=}
max_tokens: ${env.VLLM_MAX_TOKENS:=4096}
api_token: ${env.VLLM_API_TOKEN:=fake}
tls_verify: ${env.VLLM_TLS_VERIFY:=true}