The Hermes Agent setup guide matters right now because AI agents are escaping the demo cage. We are moving from "chat with a model" to "delegate a workflow to a thing that can remember, call tools, run commands, and show up again tomorrow."
Hermes Agent, built by Nous Research, is one of the more interesting entries in this new agent stack. It runs in the terminal, can connect to messaging platforms, supports tools and skills, works with many model providers, and is designed around a self-improving loop: memory, reusable skills, and cross-session recall. In normal-person language, it is not just answering. It is trying to become a persistent operator.
If you have been following our agentic web coverage, Hermes sits right in the middle of that shift. Browsers, terminals, inboxes, repos, and cloud machines are becoming surfaces where agents can work. The trick is setting them up without accidentally giving a digital intern the keys to your production database. For that cautionary side of the story, see Claude Deletes Database.
This guide walks through the practical path: install Hermes, choose a model, run a useful first session, add tools slowly, and harden the setup before you let it touch anything important.
The Short Version
What You Are Building
You are building a persistent AI worker, not a one-off chat window. A basic setup has six layers:
- The Hermes CLI or TUI.
- A model provider, such as Nous Portal, Anthropic, OpenAI-compatible endpoints, OpenRouter, Qwen, DeepSeek, Hugging Face, NVIDIA NIM, or a local server.
- Session memory, so work can continue across runs.
- Tool access, so Hermes can inspect files, use the terminal, search, or call integrations.
- Optional gateway access, so you can talk to it from Telegram, Discord, Slack, email, and similar platforms.
- Guardrails, because autonomy without boundaries is just a speedrun to incident response.
The Rule Before the Fun Stuff
Do not add gateways, cron jobs, MCP servers, or production credentials until a plain chat works. Get one clean conversation first, then layer on features.
Step 1: Install Hermes Agent
Use the One-Line Installer
On macOS, Linux, WSL2, or Android Termux, open a terminal and run:
curl -fsSL https://raw.githubusercontent.com/NousResearch/hermes-agent/main/scripts/install.sh | bashThen reload your shell:
source ~/.zshrc
# or
source ~/.bashrcCheck that Hermes is available:
hermes --helpIf you are on Windows, use WSL2. Install Ubuntu in WSL2, update packages, and run the installer inside that Linux environment.
Run the Setup Wizard
The boring path is the best path:
hermes setupThis interactive wizard helps configure the core pieces. If you already installed Hermes and only want to change the model provider, use:
hermes modelYou will use that command any time an API key expires, a provider gets weird, or you switch models.
Step 2: Pick a Model Provider
Start With a Hosted Provider
For a first setup, choose a hosted provider unless you specifically want local inference. Hermes supports a broad provider menu, including Nous Portal, Anthropic, OpenRouter, GitHub Copilot, DeepSeek, Qwen through DashScope, Hugging Face, NVIDIA NIM, MiniMax, Kimi, and custom endpoints.
The key technical requirement is context length. Hermes expects a model with at least a 64K token context window because agents need room for instructions, tool results, memory, file summaries, and multi-step plans.
For model selection context, read GPT-5.5 vs Claude Opus 4.7 and DeepSeek V4. The best Hermes model is not always the leaderboard darling. It is the one with enough context, reliable tool calling, good latency, and a price you can tolerate after the novelty wears off.
Know Where Config Lives
Hermes separates secrets from ordinary config:
~/.hermes/.env # API keys and tokens
~/.hermes/config.yaml # non-secret settingsUse the CLI when possible:
hermes config set model anthropic/claude-opus-4.6
hermes config set terminal.backend docker
hermes config set OPENROUTER_API_KEY sk-or-your-keyThe important part is avoiding a mess of scattered keys in random dotfiles, screenshots, and "temporary" notes.
Step 3: Run Your First Hermes Session
Try the Classic CLI or TUI
Start Hermes:
hermesOr use the newer terminal UI:
hermes --tuiNow give it a task you can verify:
Check the current directory, summarize what project this looks like, and identify the main entrypoint.Good first prompts are specific, grounded, and testable. Bad first prompts are things like "make my startup successful." Hermes is an agent, not a venture capitalist.
Verify Sessions Actually Resume
After your first chat works, test session continuity:
hermes --continue
# or
hermes -cIf it resumes correctly, you have crossed the first real checkpoint. Persistent context is the whole point.
Step 4: Add Tools, Skills, and Messaging
Inspect the Basics
Inside Hermes, type / to see slash commands. The practical ones to learn first are:
/help show commands
/tools list available tools
/model switch models
/save save the conversationOutside the chat, use:
hermes tools
hermes skillsSkills are reusable workflows and procedural memory. Search and install them like this:
hermes skills search kubernetes
hermes skills install openai/skills/k8sThis is where Hermes starts feeling less like a chat app and more like a workbench.
Add Messaging Only After CLI Works
If you want Hermes in Telegram, Discord, Slack, WhatsApp, Signal, email, or another gateway, run:
hermes gateway setupThen check status:
hermes gateway statusMy rule: do not connect a shared platform until you have tested the exact model, tools, and permissions locally. Messaging makes the agent more convenient, but it also widens the trigger surface.
Step 5: Run Hermes Locally
Use an OpenAI-Compatible Local Server
If you want privacy, lower marginal cost, or hardware bragging rights, run Hermes against a local model through LM Studio, Ollama, vLLM, SGLang, or another OpenAI-compatible server.
The pattern is simple:
- Start the local model server.
- Make sure it exposes an OpenAI-compatible /v1 endpoint.
- Run hermes model.
- Choose custom endpoint.
- Enter the base URL, model name, and key if required.
- Set context to at least 64K tokens.
For example, Ollama-style local setups commonly need a large context value:
ollama run your-model-name -c 65536AMD recently published a Windows/WSL2 path using LM Studio, local network serving, and Hermes configured with a custom endpoint. The broader takeaway: local Hermes is becoming normal enough that GPU vendors are writing setup guides for it.
Be Honest About Local Tradeoffs
Local models are great for data locality and experimentation. Hosted models are often better for reliability, tool use, long context, and less fiddling.
Step 6: Harden the Setup Before Real Work
Use a Sandbox
Hermes can use different terminal backends. For safer agent work, configure Docker or SSH:
hermes config set terminal.backend dockeror:
hermes config set terminal.backend sshDocker is a good first isolation layer. A separate VM or cloud instance is even cleaner for riskier workflows.
Separate Permissions
Use dedicated accounts, scoped API keys, read-only access where possible, and human approval for destructive actions. Do not give Hermes your primary cloud credentials on day one. Do not put production secrets in a repo and hope vibes become security architecture.
The Claude database incident is the headline-friendly version of this lesson, but the underlying rule applies to every agent. The model is only one part of the system. Permissions, tokens, backups, and review gates decide the blast radius.
Common Failure Modes
Hermes Installs But Does Not Respond
Run:
hermes doctor
hermes modelMost first-run failures are provider problems: bad key, wrong model name, insufficient context, stale config, or a custom endpoint that looks OpenAI-compatible until you ask it to behave.
Gateway Works, But Messages Do Not Arrive
Check:
hermes gateway statusThen re-run:
hermes gateway setupLook for bot tokens, allowlists, platform permissions, and whether the gateway process is actually running.
Session Resume Fails
Run:
hermes sessions list
hermes --continueIf sessions are missing, confirm you are using the same profile and working directory assumptions.
Bottom Line
Hermes Is Best Treated as Infrastructure
Hermes Agent is exciting because it points toward a more durable kind of AI tool: a persistent worker with memory, tools, skills, messaging access, and model flexibility. The setup is not difficult, but the mindset matters.
Install it cleanly. Pick one reliable model. Prove the first chat works. Verify session resume. Add skills and gateways slowly. Use Docker or a separate environment. Keep credentials scoped. Then let Hermes handle real workflows.
That is the difference between playing with an agent and operating one.
Sources: Hermes Agent Quickstart, NousResearch Hermes Agent on GitHub, AMD local Hermes Agent guide.
Written by
Lena Ortiz
AI Tools Analyst
Lena tests AI products through the lens of creators, operators, and teams that need software to stay useful after launch week.
AI workflow guides
Build agents like infrastructure, not toys.
Read more Syntax Dispatch coverage on agentic workflows, model choices, coding tools, and practical AI systems.
Browse AI guidesFAQ
What is Hermes Agent?
Hermes Agent is a Nous Research AI agent that runs from the terminal, supports memory, skills, tool use, messaging gateways, multiple model providers, and local or hosted model setups.
What model does Hermes Agent need?
Hermes expects a capable model with at least a 64K token context window so it has room for instructions, memory, tool results, and multi-step work.
Should I run Hermes Agent locally or with a hosted provider?
Use hosted providers first if you want reliability and fast setup. Use local models when privacy, experimentation, or cost control matter enough to justify more configuration work.



