Skip to main content
For models accessed via OpenAI-compatible APIs, such as:
  • Third-party API proxies: Use a unified API Base to call multiple models
  • Local models: Models deployed locally via Ollama, vLLM, LocalAI, etc.
  • Private deployments: Self-hosted model services within your organization
Unlike the openai provider, switching models under the Custom provider will not auto-switch the provider type. Your custom API address is always preserved.

Configuration

Third-party API Proxy

{
  "bot_type": "custom",
  "model": "deepseek-chat",
  "custom_api_key": "YOUR_API_KEY",
  "custom_api_base": "https://{your-proxy.com}/v1"
}
ParameterDescription
bot_typeMust be set to custom
modelModel name, any model supported by your proxy service
custom_api_keyAPI key provided by your proxy service
custom_api_baseAPI base URL, must be OpenAI-compatible

Local Models

Local models typically don’t require an API key — just set the API base:
{
  "bot_type": "custom",
  "model": "qwen3.5:27b",
  "custom_api_base": "http://localhost:11434/v1"
}
Common local deployment tools and their default addresses:
ToolDefault API Base
Ollamahttp://localhost:11434/v1
vLLMhttp://localhost:8000/v1
LocalAIhttp://localhost:8080/v1

Switching Models

Under the Custom provider, switching models only changes model without affecting bot_type or the API address:
/config model qwen3.5:27b