Skip to main content
Siclaw needs an LLM to power its investigation engine. Configure via ~/.siclaw/config/settings.json or environment variable overrides.
{
  "providers": {
    "default": {
      "baseUrl": "https://api.anthropic.com/v1",
      "apiKey": "sk-ant-...",
      "api": "anthropic",
      "authHeader": true,
      "models": [{
        "id": "claude-sonnet-4-20250514",
        "name": "Claude Sonnet 4",
        "contextWindow": 200000,
        "maxTokens": 16000
      }]
    }
  }
}
Recommended models: claude-sonnet-4-20250514 (best balance) or claude-opus-4-20250514 (highest quality).

OpenAI

{
  "providers": {
    "default": {
      "baseUrl": "https://api.openai.com/v1",
      "apiKey": "sk-...",
      "api": "openai-completions",
      "authHeader": true,
      "models": [{
        "id": "gpt-4o",
        "name": "GPT-4o",
        "contextWindow": 128000,
        "maxTokens": 16384
      }]
    }
  }
}

OpenAI-Compatible Providers

Any API that implements the OpenAI chat completions format works with Siclaw — Ollama, vLLM, LiteLLM, Azure OpenAI, Moonshot, DeepSeek, and many others.
{
  "providers": {
    "default": {
      "baseUrl": "http://localhost:11434/v1",
      "apiKey": "ollama",
      "api": "openai-completions",
      "authHeader": true,
      "models": [{
        "id": "llama3.1:70b",
        "name": "Llama 3.1 70B",
        "contextWindow": 131072,
        "maxTokens": 8192
      }]
    }
  }
}

Common Providers

ProviderbaseUrlNotes
Ollamahttp://localhost:11434/v1Local, free. Use 70B+ for best results.
vLLMhttp://localhost:8000/v1Self-hosted GPU inference
Moonshot (Kimi)https://api.moonshot.cn/v1moonshot-v1-128k
DeepSeekhttps://api.deepseek.com/v1deepseek-chat
Qwen (DashScope)https://dashscope.aliyuncs.com/compatible-mode/v1qwen-plus
See settings.example.json for a complete example with all fields.

Environment Variable Overrides

These override the default provider’s settings at runtime (highest priority):
SICLAW_LLM_API_KEY=sk-...       # Override default provider's API key
SICLAW_LLM_BASE_URL=https://... # Override default provider's base URL
SICLAW_LLM_MODEL=gpt-4o         # Override default model ID
API keys also support $VAR / ${VAR} references in settings.json:
{
  "providers": {
    "default": {
      "apiKey": "${MY_API_KEY}",
      "baseUrl": "https://api.openai.com/v1",
      "api": "openai-completions",
      "models": [{ "id": "gpt-4o", "name": "GPT-4o" }]
    }
  }
}

Embedding Provider

Without an embedding provider, Investigation Memory semantic search is disabled. All other features work normally.
Embedding is used for memory search — matching current symptoms against past investigation records. Any OpenAI-compatible embedding API works:
{
  "embedding": {
    "baseUrl": "https://api.example.com/v1",
    "apiKey": "sk-...",
    "model": "bge-m3",
    "dimensions": 1024
  }
}
ProviderModelDimensionsNotes
BGE-M3 (recommended)bge-m31024Multilingual, good for technical content
OpenAItext-embedding-3-small1536Easy setup if you already have an OpenAI key
Ollamanomic-embed-text768Local, free

Model Recommendations

Use CaseRecommendedNotes
Production investigationsClaude Sonnet 4 / GPT-4oBest quality-to-speed ratio
Complex root cause analysisClaude Opus 4Highest reasoning capability
Cost-sensitive / air-gappedLlama 3.1 70B+ via OllamaLocal, no API costs
Testing / developmentAny available modelSmaller models work for basic checks