AI Providers

Nine providers.
Swappable per tool.

Anthropic Claude, OpenAI, Mistral, Groq, DeepSeek, xAI Grok, Google Gemini, and Ollama (local) — all eight ship as native task types in every edition. Hugging Face rounds out the set as a 9th, job-only provider. Each is a first-class tool — not an abstraction layer papering over differences. Use Claude's extended thinking where it helps, OpenAI's structured outputs where they help, Gemini's multi-modal where relevant, Groq for sub-second LPU inference, DeepSeek/Mistral for OpenAI-compatible cost-down options, xAI Grok for the Grok roster, and Ollama for anything that needs to stay inside your network.

Each Provider, With Its Distinctive Surface

Anthropic Claude (5 tool variants)

  • Claude Agent — multi-turn with tool use
  • Claude Batch — async batch API for high-volume
  • Claude Document — PDF and document reading
  • Claude Tools — direct tool-use without agent loop
  • Claude (base) — simple prompt-completion

Valid model IDs: claude-opus-4-6, claude-sonnet-4-6, claude-haiku-4-5-20251001.

OpenAI

  • Chat completions with function calling
  • Structured outputs / JSON mode
  • Embeddings (for the RAG pipeline)
  • Fine-tuning management (for fine-tuned models)
  • Image generation (DALL-E)
  • Whisper speech-to-text, TTS text-to-speech

Valid model IDs: gpt-4o, gpt-4o-mini.

Google Gemini

  • Text and multi-modal (image, document)
  • Function calling
  • Context caching for long-context prompts

Valid model IDs: gemini-2.5-flash, gemini-2.5-pro.

Ollama (Local)

  • Any model the operator installs locally (llama3.2, mistral, gemma2, your own fine-tunes)
  • Zero restrictions, zero API cost, no data leaving the network
  • Ideal for regulated domains (finance, legal, medical) and air-gapped deployments

Model names are operator-set — no fixed dropdown.

Mistral, Groq, DeepSeek, xAI

  • Mistralmistral-large-latest and the Mistral roster via api.mistral.ai/v1
  • Groq — sub-second LPU inference; llama-3.3-70b-versatile and the Groq-hosted catalogue via api.groq.com/openai/v1
  • DeepSeekdeepseek-chat via api.deepseek.com/v1
  • xAI — Grok via api.x.ai/v1 (default grok-2)

All four use the OpenAI-compatible chat completions surface. Same task definition, swap the tool name and credential to switch provider.

Hugging Face (Job-only)

Hugging Face Inference is wired in as a job task type for batch — not as an Assistant chat backend. Useful when you want to call a specific HF model from a scheduled job (classification, embeddings, summarization) without standing up an inference server.

The Defaults Are the Right Defaults

Safety Preamble

Every AI call prepends a safety preamble resisting prompt injection and enforcing output format. Override per tool when you need bare prompts; on by default.

Rate-Limit Handling

429s retried with exponential backoff and jitter. No hand-rolled retry loops. Invisible when it's working, diagnostic when it's not.

Audit Log

Every call logs provider, model, input tokens, output tokens, response. Budget tracking via SQL query on the audit table.

"If Anthropic Refuses, Swap to Ollama"

Anthropic (and OpenAI, and Gemini) have content policies. When a domain crosses a refusal line — financial advice, legal synthesis, medical reasoning — swap the tool reference to Ollama with a local model. Zero restrictions, zero API cost, no data leaving the network. The job definition doesn't change; only the tool name and credential do.

Pick the Right AI for the Job

Not the one your SaaS vendor chose for you. Free Personal edition, all 8 native providers included (plus Hugging Face as a 9th, job-only).

Get Personal EditionAI Assistant Detail