{% extends "base.html" %} {% block title %}LLM Settings - AUTARCH{% endblock %} {% block content %}
Configured backend: {{ llm_backend }} — select a tab, fill in settings, and click Save & Activate, then Initialize LLM to initialise.
Local GGUF models may take 10–60 s to load depending on size. The page will wait — check the Debug Log for live output.
Scans for .gguf, .ggml, .bin files and SafeTensors model directories.
Requires an Anthropic API key. AUTARCH calls Claude directly for chat, agent, and analysis tasks.
Save stores settings + API key to encrypted vault. Reload re-tests the connection without changing settings. Activate saves + loads Claude as the active backend.
Also compatible with any OpenAI-format endpoint: LiteLLM, Ollama (/v1), vLLM, LocalAI, etc.
Just set the Base URL to your local server.
Configure the AI agent backends used by Agent Hal and Autonomy. Agents use the THOUGHT/ACTION/PARAMS loop to accomplish tasks with tools.
Uses the currently loaded LLM backend (configured above) for agent operations. Best for offline or privacy-sensitive work.