Quick start
Pull a model
Before running the setup wizard, make sure you have a model available. Hermes will auto-detect models downloaded through Ollama.Install
Set up
After installation, Hermes launches the setup wizard automatically. Choose Quick setup:Connect to Ollama
- Select More providers…
- Select Custom endpoint (enter URL manually)
-
Set the API base URL to the Ollama OpenAI-compatible endpoint:
-
Leave the API key blank (not required for local Ollama):
-
Hermes auto-detects downloaded models, confirm the one you want:
-
Leave context length blank to auto-detect:
Connect messaging
Optionally connect a messaging platform during setup:Launch
Recommended models
Cloud models:kimi-k2.5:cloud— Multimodal reasoning with subagentsqwen3.5:cloud— Reasoning, coding, and agentic tool use with visionglm-5.1:cloud— Reasoning and code generationminimax-m2.7:cloud— Fast, efficient coding and real-world productivity
gemma4— Reasoning and code generation locally (~16 GB VRAM)qwen3.5— Reasoning, coding, and visual understanding locally (~11 GB VRAM)

