This guide shows how to connect Hermes Agent to Lunos using a custom endpoint. Since Lunos is fully OpenAI-compatible, it slots in with zero friction — giving you access to 100+ models through a single API key.
curl -fsSL https://raw.githubusercontent.com/NousResearch/hermes-agent/main/scripts/install.sh | bash
https://api.lunos.tech/v1Run the model selector:
hermes model
Select "Custom endpoint (self-hosted / VLLM / etc.)" from the menu, then enter the three values it asks for:
| Field | Value |
|---|---|
| Base URL | https://api.lunos.tech/v1 |
| API Key | Your Lunos API key |
| Model | e.g. openai/gpt-4o, anthropic/claude-sonnet-4-6, google/gemini-2-0-flash |
Hermes saves the selection and uses it for every subsequent run. To switch models later, just run hermes model again.
You can also set it directly in ~/.hermes/config.yaml:
provider: custom
base_url: https://api.lunos.tech/v1
api_key: YOUR_LUNOS_KEY
model: openai/gpt-4o # or any Lunos-supported model ID
Or using hermes config set:
hermes config set base_url https://api.lunos.tech/v1
hermes config set model openai/gpt-4o
Hermes supports a chain of fallback providers it tries in order when the primary model fails (rate limits, server errors, auth failures):
fallback_providers:
- provider: custom
base_url: https://api.lunos.tech/v1
model: anthropic/claude-sonnet-4-6
hermes --tui
You'll see a welcome banner with your model, available tools, and skills. If Hermes can't complete a normal chat, get one clean conversation working first before layering on gateway, cron, skills, or voice.
hermes model and re-enter your Lunos key carefully; don't use console/session tokens.HERMES_STREAM_READ_TIMEOUT=120 (or higher) before launching Hermes.hermes doctor.No headings found on this page.
