{
  "version": 1,
  "llm": {
    "kind": "ollama_openai_compatible",
    "openai_compatible_base_path": "/llm/v1",
    "notes": "Same-origin paths only. nginx proxies /llm/ to the Ollama (or compatible) backend configured on the server. Change GPU/Ollama host via /etc/nginx/snippets/lisa-ollama-upstream.conf (or HTTPS location per lisa-llm-https-backend.example.md), not this JSON."
  },
  "gpu_bridge": {
    "api_base_path": "/gpu-bridge/v1"
  },
  "discovery": {
    "gpu_bridge_public_config_path": "/gpu-bridge/v1/public-config"
  }
}
