This section contains documentation for all available providers for the inference API.
inline::meta-reference
inline::sentence-transformers
inline::vllm
remote::anthropic
remote::bedrock
remote::cerebras
remote::cerebras-openai-compat
remote::databricks
remote::fireworks
remote::fireworks-openai-compat
remote::gemini
remote::groq
remote::groq-openai-compat
remote::hf::endpoint
remote::hf::serverless
remote::llama-openai-compat
remote::nvidia
remote::ollama
remote::openai
remote::passthrough
remote::runpod
remote::sambanova
remote::sambanova-openai-compat
remote::tgi
remote::together
remote::together-openai-compat
remote::vllm
remote::watsonx