Overview
IronClaw defaults to NEAR AI for model access, but supports any OpenAI-compatible endpoint as well as direct Anthropic, OpenAI, and Ollama integrations.Supported Providers
| Provider | Backend Value | API Key Required | Notes |
|---|---|---|---|
| NEAR AI | nearai | OAuth (browser) or API key | Default; multi-model access |
| Anthropic | anthropic | ANTHROPIC_API_KEY | Claude models |
| OpenAI | openai | OPENAI_API_KEY | GPT models |
| Ollama | ollama | No | Local inference |
| OpenRouter | openai_compatible | LLM_API_KEY | 300+ models via one API |
| Together AI | openai_compatible | LLM_API_KEY | Fast open-source inference |
| Fireworks AI | openai_compatible | LLM_API_KEY | Fast inference with compound AI |
| vLLM / LiteLLM | openai_compatible | Optional | Self-hosted |
| LM Studio | openai_compatible | No | Local GUI |
NEAR AI (Default)
No additional configuration required. On first run,ironclaw onboard opens a browser for OAuth authentication.
Session Token Auth (Default)
Best for: Local development, personal use~/.ironclaw/session.json.
Environment variables:
API Key Auth
Best for: CI/CD, hosting providers, VPS without browser accessNEARAI_API_KEY is set, IronClaw automatically uses the Chat Completions API at cloud-api.near.ai instead of session-based auth.
Popular Models
| Model | ID |
|---|---|
| GLM Latest (default) | zai-org/GLM-latest |
| Claude Sonnet 4 | anthropic::claude-sonnet-4-20250514 |
| GPT-5.3 Codex | openai::gpt-5.3-codex |
| GPT-5.2 | openai::gpt-5.2 |
| GPT-4o | openai::gpt-4o |
Anthropic (Claude)
Direct access to Claude models via Anthropic API.Popular Models
| Model | ID |
|---|---|
| Claude Sonnet 4 | claude-sonnet-4-20250514 |
| Claude 3.5 Sonnet | claude-3-5-sonnet-20241022 |
| Claude 3.5 Haiku | claude-3-5-haiku-20241022 |
Optional Base URL Override
OpenAI (GPT)
Direct access to OpenAI models.Popular Models
| Model | ID |
|---|---|
| GPT-4o | gpt-4o |
| GPT-4o Mini | gpt-4o-mini |
| o3-mini | o3-mini |
Optional Base URL Override
Ollama (Local)
Run models locally with Ollama.Installation
- Install Ollama from ollama.com
- Pull a model:
- Configure IronClaw:
Popular Models
| Model | Command | Notes |
|---|---|---|
| Llama 3.2 | ollama pull llama3.2 | 3B, fast |
| Mistral | ollama pull mistral | 7B, good quality |
| Qwen 2.5 | ollama pull qwen2.5 | Multilingual |
OpenAI-Compatible Providers
All providers below useLLM_BACKEND=openai_compatible. Set LLM_BASE_URL to the provider’s endpoint and LLM_API_KEY if required.
OpenRouter
OpenRouter routes to 300+ models from a single API key.Popular OpenRouter Models
| Model | ID |
|---|---|
| Claude Sonnet 4 | anthropic/claude-sonnet-4 |
| GPT-4o | openai/gpt-4o |
| Llama 4 Maverick | meta-llama/llama-4-maverick |
| Gemini 2.0 Flash | google/gemini-2.0-flash-001 |
| Mistral Small | mistralai/mistral-small-3.1-24b-instruct |
Optional HTTP Headers
OpenRouter supports custom headers for attribution:Together AI
Together AI provides fast inference for open-source models.Popular Together AI Models
| Model | ID |
|---|---|
| Llama 3.3 70B | meta-llama/Llama-3.3-70B-Instruct-Turbo |
| DeepSeek R1 | deepseek-ai/DeepSeek-R1 |
| Qwen 2.5 72B | Qwen/Qwen2.5-72B-Instruct-Turbo |
Fireworks AI
Fireworks AI offers fast inference with compound AI system support.vLLM / LiteLLM (Self-Hosted)
For self-hosted inference servers:LM Studio (Local GUI)
Start LM Studio’s local server, then:Extra HTTP Headers
For OpenAI-compatible providers that require custom headers:Key:Value pairs. Values can contain colons (e.g., URLs).
Using the Setup Wizard
Instead of editing.env manually, run:
Advanced Configuration
Fallback Models
For NEAR AI, configure automatic failover:Circuit Breaker
Prevent cascading failures:Response Caching
Cache LLM responses in memory (saves tokens on repeated prompts):Retries
Embeddings Providers
For semantic search in workspace memory:text-embedding-3-small.