Skip to main content
Siclaw needs an LLM to power its investigation engine.
  • TUI mode reads .siclaw/config/settings.json
  • Local Server / Gateway is typically configured through the Models page in the Web UI
{
  "providers": {
    "default": {
      "baseUrl": "https://api.anthropic.com/v1",
      "apiKey": "sk-ant-...",
      "api": "anthropic",
      "authHeader": true,
      "models": [{
        "id": "claude-sonnet-4-20250514",
        "name": "Claude Sonnet 4",
        "contextWindow": 200000,
        "maxTokens": 16000
      }]
    }
  }
}
Recommended models: claude-sonnet-4-20250514 (best balance) or claude-opus-4-20250514 (highest quality).

OpenAI

{
  "providers": {
    "default": {
      "baseUrl": "https://api.openai.com/v1",
      "apiKey": "sk-...",
      "api": "openai-completions",
      "authHeader": true,
      "models": [{
        "id": "gpt-4o",
        "name": "GPT-4o",
        "contextWindow": 128000,
        "maxTokens": 16384
      }]
    }
  }
}

OpenAI-Compatible Providers

Any API that implements the OpenAI chat completions format works with Siclaw — Ollama, vLLM, LiteLLM, Azure OpenAI, Moonshot, DeepSeek, and many others.
{
  "providers": {
    "default": {
      "baseUrl": "http://localhost:11434/v1",
      "apiKey": "ollama",
      "api": "openai-completions",
      "authHeader": true,
      "models": [{
        "id": "llama3.1:70b",
        "name": "Llama 3.1 70B",
        "contextWindow": 131072,
        "maxTokens": 8192
      }]
    }
  }
}

Common Providers

ProviderbaseUrlNotes
Ollamahttp://localhost:11434/v1Local, free. Use 70B+ for best results.
vLLMhttp://localhost:8000/v1Self-hosted GPU inference
Moonshot (Kimi)https://api.moonshot.cn/v1moonshot-v1-128k
DeepSeekhttps://api.deepseek.com/v1deepseek-chat
Qwen (DashScope)https://dashscope.aliyuncs.com/compatible-mode/v1qwen-plus
See settings.example.json for a complete example with all fields.

Configuration Methods

  • TUI mode: First-run wizard or /setup command inside a session
  • Local Server / Gateway: Configure providers and models in the Web UI

Embedding Provider

Without an embedding provider, Investigation Memory semantic search is disabled. All other features work normally.
Embedding is used for memory search — matching current symptoms against past investigation records. Any OpenAI-compatible embedding API works:
{
  "embedding": {
    "baseUrl": "https://api.example.com/v1",
    "apiKey": "sk-...",
    "model": "bge-m3",
    "dimensions": 1024
  }
}
If embedding.apiKey is omitted, Siclaw falls back to the default provider API key.
ProviderModelDimensionsNotes
BGE-M3 (recommended)bge-m31024Multilingual, good for technical content
OpenAItext-embedding-3-small1536Easy setup if you already have an OpenAI key
Ollamanomic-embed-text768Local, free

Model Recommendations

Use CaseRecommendedNotes
Production investigationsClaude Sonnet 4 / GPT-4oBest quality-to-speed ratio
Complex root cause analysisClaude Opus 4Highest reasoning capability
Cost-sensitive / air-gappedLlama 3.1 70B+ via OllamaLocal, no API costs
Testing / developmentAny available modelSmaller models work for basic checks