Skip to main content

AI Providers

Actyze supports 100+ AI providers through LiteLLM integration, giving you the flexibility to use any LLM service - from major cloud providers to local open-source models - with a consistent, simple configuration.

Overview

Powered by LiteLLM

Actyze uses LiteLLM, an open-source library that provides a unified interface to 100+ LLM providers. This means:

  • No Vendor Lock-in - Switch providers anytime with just 2 lines of config
  • Consistent Interface - Same configuration pattern for all providers
  • Automatic Handling - Authentication, rate limits, and retries handled automatically
  • Future-Proof - New providers supported as LiteLLM adds them
  • Local & Cloud - Use cloud APIs or run models locally

Configuration Pattern:

All providers follow this simple 2-line pattern:

PROVIDER_API_KEY=your-api-key
EXTERNAL_LLM_MODEL=provider-model-name

Supported Providers

Industry-leading AI services with production-ready APIs.

ProviderConfigurationModelsBest For
Anthropic ClaudeANTHROPIC_API_KEYclaude-sonnet-4-20250514, claude-opus-4-20250514Best SQL accuracy, complex queries
OpenAIOPENAI_API_KEYgpt-4o, gpt-4-turbo, gpt-3.5-turboGeneral purpose, widely supported
Google GeminiGEMINI_API_KEYgemini/gemini-pro, gemini/gemini-2.0-flashFast, cost-effective
PerplexityPERPLEXITY_API_KEYperplexity/sonar-reasoning-proReal-time data, research
CohereCOHERE_API_KEYcohere/command-r-plusEnterprise, fine-tuning

Example - Anthropic Claude:

ANTHROPIC_API_KEY=sk-ant-xxxxx
EXTERNAL_LLM_MODEL=claude-sonnet-4-20250514

Enterprise Cloud Providers

Enterprise-grade AI services with SLAs, compliance, and dedicated support.

ProviderConfigurationModelsBest For
Azure OpenAIAZURE_API_KEY, AZURE_API_BASEazure/gpt-4o, azure/your-deploymentMicrosoft Azure customers
AWS BedrockAWS_ACCESS_KEY_ID, AWS_SECRET_ACCESS_KEYbedrock/us.anthropic.claude-sonnet-4, bedrock/anthropic.claude-v2AWS infrastructure
Google Vertex AIVERTEXAI_PROJECT, VERTEXAI_LOCATIONvertex_ai/gemini-proGoogle Cloud Platform
IBM watsonx.aiWATSONX_API_KEYwatsonx/meta-llamaIBM enterprise stack

Example - Azure OpenAI:

AZURE_API_KEY=your-azure-key
AZURE_API_BASE=https://your-resource.openai.azure.com
EXTERNAL_LLM_MODEL=azure/your-deployment-name

Fast & Free Tier Providers

High-performance models with generous free tiers or pay-per-use pricing.

ProviderConfigurationModelsBest For
GroqGROQ_API_KEYgroq/llama-3.3-70b-versatile, groq/mixtral-8x7bUltra-fast inference, testing
Together AITOGETHERAI_API_KEYtogether_ai/meta-llama/Meta-Llama-3.1Open models, low cost
Fireworks AIFIREWORKS_API_KEYfireworks_ai/llama-v3-70bFast, open source models
ReplicateREPLICATE_API_KEYreplicate/meta/llama-2-70bModel marketplace

Example - Groq:

GROQ_API_KEY=gsk_xxxxx
EXTERNAL_LLM_MODEL=groq/llama-3.3-70b-versatile

Open Source & Local Providers

Run models locally or on your infrastructure - no API costs, complete privacy.

ProviderConfigurationModelsBest For
OllamaOLLAMA_API_BASEollama/llama3, ollama/mistralLocal deployment, privacy
LM StudioLM_STUDIO_API_BASElm_studio/model-nameDesktop, development
vLLMVLLM_API_BASEvllm/model-nameSelf-hosted, GPU clusters
OpenRouterOPENROUTER_API_KEYopenrouter/autoMultiple models, routing
Hugging FaceHUGGINGFACE_API_KEYhuggingface/model-nameResearch, custom models

Example - Ollama (Local):

OLLAMA_API_BASE=http://localhost:11434
EXTERNAL_LLM_MODEL=ollama/llama3
# No API key needed for local Ollama

Specialized Providers

Purpose-built AI services for specific use cases.

ProviderConfigurationUse Case
Anthropic via BedrockAWS BedrockEnterprise AWS deployments
OpenAI via AzureAzure OpenAIMicrosoft compliance requirements
Custom OpenAI-compatibleEnterprise gatewaysIT-managed AI access
Mistral AIMISTRAL_API_KEYEuropean AI sovereignty
AI21 LabsAI21_API_KEYMultilingual, long context

Configuration

Standard Providers

For cloud providers (Anthropic, OpenAI, Gemini, etc.), use this 2-line pattern:

Docker (.env):

# Choose your provider
ANTHROPIC_API_KEY=sk-ant-xxxxx
EXTERNAL_LLM_MODEL=claude-sonnet-4-20250514

# Or use OpenAI
# OPENAI_API_KEY=sk-xxxxx
# EXTERNAL_LLM_MODEL=gpt-4o

# Or use Gemini
# GEMINI_API_KEY=your-key
# EXTERNAL_LLM_MODEL=gemini/gemini-pro

Helm (values.yaml + values-secrets.yaml):

# values.yaml
modelStrategy:
externalLLM:
enabled: true
model: "claude-sonnet-4-20250514"

# values-secrets.yaml
secrets:
externalLLM:
apiKey: "sk-ant-xxxxx"

Local/Self-Hosted Models

For local models (Ollama, LM Studio, vLLM), set the base URL:

Docker (.env):

# Ollama running locally
OLLAMA_API_BASE=http://localhost:11434
EXTERNAL_LLM_MODEL=ollama/llama3

# Or vLLM on your server
# VLLM_API_BASE=http://your-server:8000
# EXTERNAL_LLM_MODEL=vllm/mistral-7b

Helm (values.yaml):

modelStrategy:
externalLLM:
enabled: true
model: "ollama/llama3"
baseUrl: "http://ollama-service:11434"

Enterprise Gateway

For IT-managed AI gateways that proxy multiple providers:

Docker (.env):

# Enterprise gateway configuration
EXTERNAL_LLM_MODE=openai-compatible
EXTERNAL_LLM_BASE_URL=https://llm-gateway.yourcompany.com/v1/chat/completions
EXTERNAL_LLM_API_KEY=your-enterprise-token
EXTERNAL_LLM_MODEL=your-internal-model-name

# Optional: Add custom headers for tracking
EXTERNAL_LLM_AUTH_TYPE=bearer
EXTERNAL_LLM_EXTRA_HEADERS={"X-Department": "engineering", "X-Cost-Center": "analytics"}

Use Cases:

  • Centralized billing and cost tracking
  • Compliance and audit logging
  • Rate limiting and access control
  • Multi-model routing and fallback
  • Custom authentication requirements

Model Selection Guide

By Use Case

Production SQL Generation:

  • Best: claude-sonnet-4-20250514 - Highest accuracy
  • Fast: gpt-4o - Good balance of speed and accuracy
  • Budget: gemini/gemini-pro - Cost-effective

Development & Testing:

  • Local: ollama/llama3 - No API costs
  • Cloud: groq/llama-3.3-70b-versatile - Free tier, very fast

Enterprise Compliance:

  • AWS: bedrock/us.anthropic.claude-sonnet-4-20250514 (inference profiles supported)
  • Azure: azure/gpt-4o
  • GCP: vertex_ai/gemini-pro

By Performance

Fastest (< 1s response):

  • Groq (any model)
  • Gemini Flash
  • Together AI

Most Accurate:

  • Claude Opus 4
  • GPT-4o
  • Claude Sonnet 4

Best Value:

  • Gemini Pro (cheapest)
  • Groq (free tier)
  • Ollama (self-hosted, no API costs)

Need Another Provider?

Check LiteLLM Documentation

Actyze supports all providers that LiteLLM supports. For the complete, up-to-date list:

📚 LiteLLM Providers Documentation:

This includes:

  • 100+ cloud and local providers
  • Configuration examples for each
  • Model names and capabilities
  • Authentication requirements

Configuration Pattern

All LiteLLM providers follow this pattern:

# 1. Set the provider's API key (if needed)
PROVIDER_API_KEY=your-key

# 2. Set the model name (format: provider/model-name)
EXTERNAL_LLM_MODEL=provider/model-name

# 3. Optional: Set base URL for self-hosted
PROVIDER_API_BASE=http://your-server:port

Example - Using Mistral AI:

  1. Go to: https://docs.litellm.ai/docs/providers/mistral
  2. Get your Mistral API key
  3. Configure:
    MISTRAL_API_KEY=your-mistral-key
    EXTERNAL_LLM_MODEL=mistral/mistral-large-latest

Request a Provider

If you need a provider not yet supported by LiteLLM:

  1. Check LiteLLM GitHub: https://github.com/BerriAI/litellm/issues
  2. Open an issue requesting the provider
  3. Contact Actyze Support with your use case

We'll work with the LiteLLM team to add support.

Troubleshooting

Provider Not Working

Check configuration:

# Verify environment variables
docker-compose exec nexus env | grep -i "api_key\|llm"

# Or in Kubernetes
kubectl exec -n actyze deployment/dashboard-nexus -- env | grep -i "api_key\|llm"

Common issues:

  • ❌ Wrong model name format (should be provider/model-name)
  • ❌ API key not set or incorrect
  • ❌ Model name doesn't exist for that provider
  • ❌ Rate limit exceeded (check provider dashboard)

LiteLLM Version & Compatibility

Current Version: Actyze uses LiteLLM v1.81.12 (February 2026)

Key Features:

  • ✅ AWS Bedrock inference profiles support (global.*, us.* prefixes)
  • ✅ Bedrock Batches API & Knowledge Bases
  • ✅ Claude Code web search across all providers
  • ✅ 100+ LLM providers with unified interface

Check Compatibility: https://docs.litellm.ai/docs/providers

⚠️ Bedrock Users: If using inference profiles like bedrock/us.anthropic.claude-sonnet-4-20250514, ensure you're on v1.81.12+ for proper support.

Test Provider Directly

Test your API key directly with the provider's API:

# Example: Test Anthropic API key
curl https://api.anthropic.com/v1/messages \
-H "x-api-key: $ANTHROPIC_API_KEY" \
-H "anthropic-version: 2023-06-01" \
-H "content-type: application/json" \
-d '{"model":"claude-sonnet-4-20250514","max_tokens":100,"messages":[{"role":"user","content":"test"}]}'

Best Practices

Production Deployments

  1. Use Enterprise Providers for SLAs and support

    • Azure OpenAI, AWS Bedrock, or Vertex AI
  2. Configure Fallbacks in your enterprise gateway

    • Primary: Claude Sonnet 4
    • Fallback: GPT-4o
  3. Monitor Costs via provider dashboards

  4. Set Rate Limits in .env:

    EXTERNAL_LLM_MAX_TOKENS=4096
    EXTERNAL_LLM_TEMPERATURE=0.1

Development & Testing

  1. Use Free Tiers for testing:

    • Groq (fastest, free tier)
    • Ollama (local, no costs)
  2. Switch Easily between providers:

    # Test with free Groq
    GROQ_API_KEY=gsk_xxxxx
    EXTERNAL_LLM_MODEL=groq/llama-3.3-70b-versatile

    # Then production with Claude
    ANTHROPIC_API_KEY=sk-ant-xxxxx
    EXTERNAL_LLM_MODEL=claude-sonnet-4-20250514

Cost Optimization

  1. Start with cheaper models: Gemini Pro or Groq
  2. Use local models for development: Ollama
  3. Cache responses (enabled by default in Actyze)
  4. Monitor token usage via provider dashboards

Next Steps

Additional Resources