Installation
Huginn ships as a single static binary. No runtime dependencies required — it manages its own inference engine and models.
macOS
Homebrew (recommended):
brew install huginn
curl:
curl -fsSL https://huginn.sh/install.sh | sh
Linux
curl -fsSL https://huginn.sh/install.sh | sh
Manual:
wget https://github.com/scrypster/huginn/releases/latest/download/huginn-linux-amd64
chmod +x huginn-linux-amd64
sudo mv huginn-linux-amd64 /usr/local/bin/huginn
Windows
Scoop:
scoop install huginn
Manual: Download huginn-windows-amd64.exe from GitHub Releases and add to your PATH.
Verify
huginn --version
# huginn v0.2.0
What Huginn creates on first run
When you run huginn for the first time, it sets up everything in ~/.huginn/:
| Path | What |
|---|---|
~/.huginn/config.json | Your configuration (auto-created with defaults) |
~/.huginn/agents.json | Named agent personas |
~/.huginn/bin/llama-server | Inference runtime — managed mode only |
~/.huginn/models/*.gguf | Downloaded models — managed mode only |
Managed mode (no Ollama required)
Huginn can manage its own inference runtime (llama.cpp) and models with no Ollama dependency. Enable managed mode in ~/.huginn/config.json:
{
"backend": {
"type": "managed"
}
}
On first run with managed mode, Huginn downloads the llama-server runtime and walks you through picking and downloading a model.
Using Ollama (default)
The default backend points to a local Ollama instance:
ollama pull qwen2.5-coder:14b
huginn
Or point at any OpenAI-compatible endpoint:
{
"backend": {
"type": "external",
"endpoint": "http://localhost:11434"
}
}
Huginn works with any OpenAI-compatible endpoint — Ollama, LM Studio, vLLM, or a remote server.
Cloud providers
Huginn supports Anthropic, OpenRouter, and OpenAI directly. Set the backend in ~/.huginn/config.json:
Anthropic:
{ "backend": { "type": "anthropic", "api_key": "$ANTHROPIC_API_KEY" } }
OpenRouter (200+ models):
{ "backend": { "type": "openrouter", "api_key": "$OPENROUTER_API_KEY" } }
OpenAI:
{ "backend": { "type": "openai", "endpoint": "https://api.openai.com/v1", "api_key": "$OPENAI_API_KEY" } }
API keys starting with $ are resolved from environment variables automatically.
Hardware requirements
| RAM | Recommended model |
|---|---|
| 8 GB | qwen2.5-coder:7b (4.7 GB) |
| 16 GB | qwen2.5-coder:14b (8.1 GB) |
| 32 GB+ | qwen2.5-coder:32b (18 GB) |
GPU acceleration (Metal on macOS, CUDA on Linux) is used automatically when available.