Adding Providers
Hermes can already talk to any OpenAI-compatible endpoint through the custom provider path. Do not add a built-in provider unless you want first-class UX for that service:
- provider-specific auth or token refresh
- a curated model catalog
- setup /
hermes modelmenu entries - provider aliases for
provider:modelsyntax - a non-OpenAI API shape that needs an adapter
If the provider is just "another OpenAI-compatible base URL and API key", a named custom provider may be enough.
The mental model
A built-in provider has to line up across a few layers:
hermes_cli/auth.pydecides how credentials are found.hermes_cli/runtime_provider.pyturns that into runtime data:providerapi_modebase_urlapi_keysource
run_agent.pyusesapi_modeto decide how requests are built and sent.hermes_cli/models.py,hermes_cli/main.py, andhermes_cli/setup.pymake the provider show up in the CLI.agent/auxiliary_client.pyandagent/model_metadata.pykeep side tasks and token budgeting working.
The important abstraction is api_mode.
- Most providers use
chat_completions. - Codex uses
codex_responses. - Anthropic uses
anthropic_messages. - A new non-OpenAI protocol usually means adding a new adapter and a new
api_modebranch.
Choose the implementation path first
Path A — OpenAI-compatible provider
Use this when the provider accepts standard chat-completions style requests.
Typical work:
- add auth metadata
- add model catalog / aliases
- add runtime resolution
- add CLI menu wiring
- add aux-model defaults
- add tests and user docs
You usually do not need a new adapter or a new api_mode.
Path B — Native provider
Use this when the provider does not behave like OpenAI chat completions.
Examples in-tree today:
codex_responsesanthropic_messages
This path includes everything from Path A plus:
- a provider adapter in
agent/ run_agent.pybranches for request building, dispatch, usage extraction, interrupt handling, and response normalization- adapter tests
File checklist
Required for every built-in provider
hermes_cli/auth.pyhermes_cli/models.pyhermes_cli/runtime_provider.pyhermes_cli/main.pyhermes_cli/setup.pyagent/auxiliary_client.pyagent/model_metadata.py- tests
- user-facing docs under
website/docs/
Additional for native / non-OpenAI providers
agent/<provider>_adapter.pyrun_agent.pypyproject.tomlif a provider SDK is required
Step 1: Pick one canonical provider id
Choose a single provider id and use it everywhere.
Examples from the repo:
openai-codexkimi-codingminimax-cn
That same id should appear in:
PROVIDER_REGISTRYinhermes_cli/auth.py_PROVIDER_LABELSinhermes_cli/models.py_PROVIDER_ALIASESin bothhermes_cli/auth.pyandhermes_cli/models.py- CLI
--providerchoices inhermes_cli/main.py - setup / model selection branches
- auxiliary-model defaults
- tests
If the id differs between those files, the provider will feel half-wired: auth may work while /model, setup, or runtime resolution silently misses it.
Step 2: Add auth metadata in hermes_cli/auth.py
For API-key providers, add a ProviderConfig entry to PROVIDER_REGISTRY with:
idnameauth_type="api_key"inference_base_urlapi_key_env_vars- optional
base_url_env_var
Also add aliases to _PROVIDER_ALIASES.
Use the existing providers as templates:
- simple API-key path: Z.AI, MiniMax
- API-key path with endpoint detection: Kimi, Z.AI
- native token resolution: Anthropic
- OAuth / auth-store path: Nous, OpenAI Codex
Questions to answer here:
- What env vars should Hermes check, and in what priority order?
- Does the provider need base-URL overrides?
- Does it need endpoint probing or token refresh?
- What should the auth error say when credentials are missing?
If the provider needs something more than "look up an API key", add a dedicated credential resolver instead of shoving logic into unrelated branches.
Step 3: Add model catalog and aliases in hermes_cli/models.py
Update the provider catalog so the provider works in menus and in provider:model syntax.
Typical edits:
_PROVIDER_MODELS_PROVIDER_LABELS_PROVIDER_ALIASES- provider display order inside
list_available_providers() provider_model_ids()if the provider supports a live/modelsfetch
If the provider exposes a live model list, prefer that first and keep _PROVIDER_MODELS as the static fallback.
This file is also what makes inputs like these work:
anthropic:claude-sonnet-4-6
kimi:model-name
If aliases are missing here, the provider may authenticate correctly but still fail in /model parsing.
Step 4: Resolve runtime data in hermes_cli/runtime_provider.py
resolve_runtime_provider() is the shared path used by CLI, gateway, cron, ACP, and helper clients.
Add a branch that returns a dict with at least:
{
"provider": "your-provider",
"api_mode": "chat_completions", # or your native mode
"base_url": "https://...",
"api_key": "...",
"source": "env|portal|auth-store|explicit",
"requested_provider": requested_provider,
}
If the provider is OpenAI-compatible, api_mode should usually stay chat_completions.
Be careful with API-key precedence. Hermes already contains logic to avoid leaking an OpenRouter key to unrelated endpoints. A new provider should be equally explicit about which key goes to which base URL.
Step 5: Wire the CLI in hermes_cli/main.py and hermes_cli/setup.py
A provider is not discoverable until it shows up in the interactive flows.
Update:
hermes_cli/main.py
provider_labels- provider dispatch inside the
modelcommand --providerargument choices- login/logout choices if the provider supports those flows
- a
_model_flow_<provider>()function, or reuse_model_flow_api_key_provider()if it fits
hermes_cli/setup.py
provider_choices- auth branch for the provider
- model-selection branch
- any provider-specific explanatory text
- any place where a provider should be excluded from OpenRouter-only prompts or routing settings
If you only update one of these files, hermes model and hermes setup will drift.
Step 6: Keep auxiliary calls working
Two files matter here:
agent/auxiliary_client.py
Add a cheap / fast default aux model to _API_KEY_PROVIDER_AUX_MODELS if this is a direct API-key provider.
Auxiliary tasks include things like:
- vision summarization
- web extraction summarization
- context compression summaries
- session-search summaries
- memory flushes
If the provider has no sensible aux default, side tasks may fall back badly or use an expensive main model unexpectedly.
agent/model_metadata.py
Add context lengths for the provider's models so token budgeting, compression thresholds, and limits stay sane.
Step 7: If the provider is native, add an adapter and run_agent.py support
If the provider is not plain chat completions, isolate the provider-specific logic in agent/<provider>_adapter.py.
Keep run_agent.py focused on orchestration. It should call adapter helpers, not hand-build provider payloads inline all over the file.
A native provider usually needs work in these places:
New adapter file
Typical responsibilities:
- build the SDK / HTTP client
- resolve tokens
- convert OpenAI-style conversation messages to the provider's request format
- convert tool schemas if needed
- normalize provider responses back into what
run_agent.pyexpects - extract usage and finish-reason data
run_agent.py
Search for api_mode and audit every switch point. At minimum, verify:
__init__chooses the newapi_mode- client construction works for the provider
_build_api_kwargs()knows how to format requests_api_call_with_interrupt()dispatches to the right client call- interrupt / client rebuild paths work
- response validation accepts the provider's shape
- finish-reason extraction is correct
- token-usage extraction is correct
- fallback-model activation can switch into the new provider cleanly
- summary-generation and memory-flush paths still work
Also search run_agent.py for self.client.. Any code path that assumes the standard OpenAI client exists can break when a native provider uses a different client object or self.client = None.
Prompt caching and provider-specific request fields
Prompt caching and provider-specific knobs are easy to regress.
Examples already in-tree:
- Anthropic has a native prompt-caching path
- OpenRouter gets provider-routing fields
- not every provider should receive every request-side option
When you add a native provider, double-check that Hermes is only sending fields that provider actually understands.
Step 8: Tests
At minimum, touch the tests that guard provider wiring.
Common places:
tests/test_runtime_provider_resolution.pytests/test_cli_provider_resolution.pytests/test_cli_model_command.pytests/test_setup_model_selection.pytests/test_provider_parity.pytests/test_run_agent.pytests/test_<provider>_adapter.pyfor a native provider
For docs-only examples, the exact file set may differ. The point is to cover:
- auth resolution
- CLI menu / provider selection
- runtime provider resolution
- agent execution path
- provider:model parsing
- any adapter-specific message conversion
Run tests with xdist disabled:
source .venv/bin/activate
python -m pytest tests/test_runtime_provider_resolution.py tests/test_cli_provider_resolution.py tests/test_cli_model_command.py tests/test_setup_model_selection.py -n0 -q
For deeper changes, run the full suite before pushing:
source .venv/bin/activate
python -m pytest tests/ -n0 -q
Step 9: Live verification
After tests, run a real smoke test.
source .venv/bin/activate
python -m hermes_cli.main chat -q "Say hello" --provider your-provider --model your-model
Also test the interactive flows if you changed menus:
source .venv/bin/activate
python -m hermes_cli.main model
python -m hermes_cli.main setup
For native providers, verify at least one tool call too, not just a plain text response.
Step 10: Update user-facing docs
If the provider is meant to ship as a first-class option, update the user docs too:
website/docs/getting-started/quickstart.mdwebsite/docs/user-guide/configuration.mdwebsite/docs/reference/environment-variables.md
A developer can wire the provider perfectly and still leave users unable to discover the required env vars or setup flow.
OpenAI-compatible provider checklist
Use this if the provider is standard chat completions.
-
ProviderConfigadded inhermes_cli/auth.py - aliases added in
hermes_cli/auth.pyandhermes_cli/models.py - model catalog added in
hermes_cli/models.py - runtime branch added in
hermes_cli/runtime_provider.py - CLI wiring added in
hermes_cli/main.py - setup wiring added in
hermes_cli/setup.py - aux model added in
agent/auxiliary_client.py - context lengths added in
agent/model_metadata.py - runtime / CLI tests updated
- user docs updated
Native provider checklist
Use this when the provider needs a new protocol path.
- everything in the OpenAI-compatible checklist
- adapter added in
agent/<provider>_adapter.py - new
api_modesupported inrun_agent.py - interrupt / rebuild path works
- usage and finish-reason extraction works
- fallback path works
- adapter tests added
- live smoke test passes
Common pitfalls
1. Adding the provider to auth but not to model parsing
That makes credentials resolve correctly while /model and provider:model inputs fail.
2. Forgetting that config["model"] can be a string or a dict
A lot of provider-selection code has to normalize both forms.
3. Assuming a built-in provider is required
If the service is just OpenAI-compatible, a custom provider may already solve the user problem with less maintenance.
4. Forgetting auxiliary paths
The main chat path can work while summarization, memory flushes, or vision helpers fail because aux routing was never updated.
5. Native-provider branches hiding in run_agent.py
Search for api_mode and self.client.. Do not assume the obvious request path is the only one.
6. Sending OpenRouter-only knobs to other providers
Fields like provider routing belong only on the providers that support them.
7. Updating hermes model but not hermes setup
Both flows need to know about the provider.
Good search targets while implementing
If you are hunting for all the places a provider touches, search these symbols:
PROVIDER_REGISTRY_PROVIDER_ALIASES_PROVIDER_MODELSresolve_runtime_provider_model_flow_provider_choicesapi_mode_API_KEY_PROVIDER_AUX_MODELSself.client.