The AI agent framework market hit $7.84 billion in 2025 and is projected to reach $52.62 billion by 2030. In this landscape, two projects have emerged as the fastest-growing contenders of 2026: OpenClaw (359K+ GitHub stars, TypeScript) and Hermes Agent from Nous Research (100K+ GitHub stars, Python).
Choosing between them is not just a tooling preference. It is an architectural decision that shapes how your agents learn, scale, and secure themselves. This article breaks down the 10 key differences with data from both projects, independent audits, and real-world benchmarks.
TL;DR / Key Takeaways
- OpenClaw is a gateway-first platform built for multi-channel reach. It excels at routing messages across Slack, Discord, WhatsApp, and web with 44,000+ community skills on ClawHub. Best for teams that need broad integrations fast.
- Hermes Agent is an agent-first runtime built for self-improvement. It learns from experience, creating and refining skills autonomously, delivering 40% speed gains on repeated tasks. Best for teams running repetitive workflows where accumulated knowledge saves time and tokens.
- Security profiles differ sharply. OpenClaw has logged 100+ CVEs and 137 security advisories in its first three months. Hermes Agent has zero agent-specific CVEs but ships with a permissive default configuration that requires hardening.
- They can work together. MCP and A2A protocol support means you can use OpenClaw for channel routing and Hermes for intelligent task execution in a hybrid architecture.
Introduction: Two Philosophies for Building AI Agents
Gartner predicts that 40% of enterprise applications will feature task-specific AI agents by end of 2026, up from less than 5% in 2025. As teams rush to deploy agents in production, the choice of framework determines not just what agents can do today, but how they evolve over time.
OpenClaw, released in late 2025, became the most-starred non-aggregator software project on GitHub, surpassing React's 10-year record in just 60 days. It now has 3.2 million monthly active users and 500K+ running instances. Its bet: the hard problem is routing and control. Get messages from every channel to the right agent with the right tools, and let the LLM handle the rest.
Hermes Agent, released on February 25, 2026 by Nous Research, accumulated 95,600 GitHub stars in seven weeks and built a 30,000-member subreddit. Its tagline, "The Agent That Grows With You," captures a different thesis: the hard problem is memory and self-improvement. An agent that remembers what it learns is worth more than one that is merely well-routed.
This Hermes Agent vs OpenClaw comparison examines 10 dimensions that matter most for production deployments.
Architecture: Gateway-First vs Agent-First
The architectural divide between Hermes Agent and OpenClaw is not a surface-level difference in language choice. It reflects fundamentally different answers to the question: what should be at the center of an AI agent system?
OpenClaw: The Gateway as the Center of Gravity
OpenClaw is a single Node.js 22+ process bound to 127.0.0.1:18789 by default. The Gateway owns all messaging surfaces (WhatsApp via Baileys, Telegram via grammY, Slack, Discord, Signal, iMessage, WebChat) and acts as the central control plane for routing, authentication, rate limiting, and session management.
Its core innovation is the Lane Queue system, which enforces serial execution by default and allows parallelism only for explicitly marked low-risk tasks. Per-session serialization guarantees only one active run per session at a time, preventing race conditions. The overall parallelism cap is configurable: main lanes default to 4 concurrent runs, subagent lanes default to 8.
The message flow follows a consistent loop: Channel Adapter standardizes input, Gateway routes to a session, the agent loads context and skills, sends conversation to the LLM, runs tool calls, streams the reply back, and persists state. The Gateway owns every step.
Hermes Agent: The Agent as the Center of Gravity
Hermes Agent is a Python 3.11+ runtime where the AIAgent class is the primary unit of computation. The prompt builder assembles context from personality, memory, skills, and model-specific instructions. A runtime resolver maps provider/model tuples to API configurations across 18+ providers, making the system model-agnostic by design. The central tool registry manages 47 registered tools across 19 toolsets, with each tool self-registering at import time.
Hermes offers six execution environments: Local, Docker, SSH, Daytona, Singularity, and Modal. The serverless options (Daytona, Modal) offer near-zero idle cost since the agent hibernates when not in use.
What This Means in Practice
OpenClaw gives operators full visibility and control through its centralized Gateway, with audit trails, rate limiting, and model switching at a single point. Hermes gives the agent itself more autonomy, with the runtime resolver and tool registry serving the agent's decisions rather than a central controller's rules. For customer-facing chatbots across channels, OpenClaw's gateway pattern fits naturally. For internal automation where agents need to learn and improve, Hermes's agent-first architecture aligns better.
Learning and Self-Improvement: Autonomous Skills vs Stateless Sessions
This is the sharpest differentiator between Hermes Agent and OpenClaw, and the one most likely to determine which framework fits a given use case.
Hermes Agent's Five-Step Learning Loop
Hermes runs a structured learning sequence on every non-trivial task:
- Receive a message from the user or a scheduled trigger.
- Retrieve context by querying persistent memory via FTS5 full-text search (~10ms latency over 10K+ documents) for relevant past skills and memories.
- Reason and act as the LLM plans the task and invokes tools.
- Document outcome: if the task involved 5+ tool calls, the agent autonomously writes a skill file following the agentskills.io open standard.
- Persist knowledge: the skill gets indexed into memory, available to future sessions.
Approximately every 15 tool calls, Hermes reflects on what worked and what failed, then auto-generates or updates a skill file encoding the successful approach. The results are measurable: in Nous Research's own benchmarks, agents with 20 or more self-created skills completed research tasks 40% faster than a fresh instance with no prior skills, without any manual prompt tuning. This is specifically about token and time savings, not output quality improvement.
The improvement is domain-specific. A skill built from summarizing GitHub pull requests does not automatically transfer to planning a database migration. The 40% speed gain shows up most clearly after consistent use in a narrow domain, not in one-off varied sessions. One user reported that within two hours of first running Hermes, the agent had created three skill documents and completed a similar research task 40% faster using those skills.
OpenClaw's Stateless Session Model
OpenClaw operates in a fundamentally different paradigm. Each session starts fresh, relying on the developer or community to build and register skills manually. The agent has access to the same tools and instructions it always has, but it does not accumulate experience in any structured way.
Session state is stored in JSONL files at ~/.openclaw/agents/<agentId>/sessions/<sessionId>.jsonl, with each line representing a standalone message or event. As transcripts grow, they are "compacted" to fit within model context windows. The project's own documentation draws a clear line: "Sessions are for reasoning, not storage."
The persistence layer (DuckDB, the workspace filesystem) holds durable data, but there is no mechanism for the agent to autonomously extract, test, and refine reusable procedures from its experience. Every task is approached as a new problem.
The Cost Tradeoff
The learning loop is not free. Hermes's reflection and optimizer modules consume extra tokens, roughly 15-25% overhead compared to a standard agent. But this overhead is amortized: once a skill exists, future executions of similar tasks bypass the full LLM reasoning chain, reducing both time and token consumption. AWS cost analysis shows stateful AI applications typically cost 2-3x more to operate than stateless equivalents due to storage and synchronization overhead, but the cumulative efficiency gains from skill reuse can offset this for repetitive workloads.
Ecosystem and Community: Scale vs Self-Sufficiency
OpenClaw: The Marketplace Approach
OpenClaw's ecosystem is massive. ClawHub hosts 44,000+ community-built skills as of April 2026, up from 850 in November 2025, representing roughly 50x growth in five months. Categories span Coding and IDEs (22.7%), Web Dev (17.6%), DevOps and Cloud (7.5%), Search and Research (6.6%), and Browser Automation (6.2%). The repository has 1,800+ contributors and 73,200+ forks.
The scale is real, but so are the caveats. OpenClaw's "defense rate" (percentage of users who continue using the project after initial engagement) stands at only 17%, suggesting many users star the repo but do not deeply adopt the framework. And the marketplace's open-by-default publishing model, requiring only a GitHub account at least one week old, has created significant security issues (covered in the security section below).
Hermes Agent: The Self-Generating Approach
Hermes ships with 118 curated built-in skills including web control, Gmail/Calendar/Drive/Contacts/Sheets/Docs integration, Spotify control, YouTube transcript processing, academic paper retrieval from arXiv, GitHub PR workflows, and Excalidraw diagramming. The community is smaller (515 contributors, 100K+ stars) but growing at a remarkable pace, earning 47,000 stars in two months.
The critical difference is self-generation. Because the agent creates its own skills during normal use, the ecosystem gap narrows over time. A Hermes instance that has been running for weeks on DevOps tasks will have built a library of domain-specific skills that no marketplace can replicate, because they encode that specific team's workflows, preferences, and tool configurations.
Community-contributed skills follow the agentskills.io open standard, making them portable. Notable collections include Anthropic-Cybersecurity-Skills (734+ security skills) and Chainlink Oracle integration.
For a new developer, OpenClaw offers immediate breadth: install a ClawHub skill and start using it. Hermes requires patience, but for teams committed to a specific domain, self-generated skills become increasingly valuable because they are tailored and refined through actual use.
Memory Architecture: Three-Layer Persistent vs Session-Based SOUL.md
Hermes Agent's Three-Layer Memory
Hermes implements a structured persistent memory system across three layers:
Layer 1: System Prompt Memory (MEMORY.md and USER.md). A frozen snapshot injected into every session's system prompt. MEMORY.md (2,200 character limit, ~800 tokens) stores environment facts, conventions, completed work, and corrections. USER.md (1,375 character limit, ~500 tokens) stores communication preferences, working style, and expectations.
Layer 2: Episodic Memory (Skills). After each task, Hermes writes a structured record into a ChromaDB vector store capturing the task description, tool calls made, what succeeded, and what failed. On new tasks, it embeds the request and runs semantic similarity search against past episodes. High-similarity matches are injected into the planning prompt as context.
Layer 3: Session Search. All sessions are logged in a SQLite database (~/.hermes/state.db) with FTS5 full-text indexing. The agent accesses this archive using the session_search tool, enabling questions like "Did we discuss X before?" or "What was the outcome of the auth service issue last week?"
The system also supports 8 external memory provider plugins (including Mem0, Honcho, and Hindsight), making the memory architecture pluggable.
OpenClaw's SOUL.md Session Model
OpenClaw uses SOUL.md as its identity layer, a Markdown configuration file that defines the agent's personality, values, tone, and behavioral boundaries. It is the first file injected into the agent's context at the start of every session. Additional workspace files (AGENTS.md, USER.md) provide supplementary context.
The compaction system manages context window pressure. When sessions approach the token limit (~205K tokens for some models), older messages are summarized so the conversation can continue. Session reset options include daily reset (new session at 4:00 AM local time), idle reset, and manual reset via /new or /reset commands.
The key limitation is context consumption. In complex workspaces, workspace file injection consumes approximately 35,600 tokens per message. After roughly one week of daily memory logging, the agent spends half its context window reading through old logs trying to find relevant details, creating a fundamental scaling bottleneck for learning-dependent workflows.
Why This Matters
The difference is structural: Hermes separates memory into purpose-built tiers (frozen prompt context, semantic skill search, full-text session archive), while OpenClaw relies on a single context window that must hold personality, workspace files, conversation history, and any accumulated knowledge simultaneously. For short-lived, single-session interactions, OpenClaw's approach is simpler and sufficient. For long-running agents that need to accumulate knowledge over weeks and months, Hermes's layered architecture prevents the context window from becoming a bottleneck.
Security Posture: CVE History and Supply-Chain Risk
Security may be the most consequential difference between these two platforms for production deployments.
OpenClaw's CVE Track Record
Security researcher Joel Gamblin's public tracker logged 137 security advisories for OpenClaw between February 2 and April 4, 2026, roughly one new advisory every 15 hours. CertiK's systematic analysis documented 280+ GitHub advisories, 100+ CVEs, and 135,000 exposed instances.
Critical vulnerabilities include:
- CVE-2026-25253 (ClawBleed): One-click remote code execution via cross-site WebSocket hijacking, allowing malicious websites to steal authentication tokens and gain full control over the Gateway.
- CVE-2026-32922: CVSS 9.9 token rotation privilege escalation to remote code execution, the most critical vulnerability in OpenClaw's history.
- CVE-2026-33579: Pair approval command path privilege escalation.
March 2026 saw 15+ CVEs in a single month, with at least three scoring CVSS 9.4 or higher. Nine CVEs were disclosed in just four days.
ClawHub Supply-Chain Risks
Beyond formal CVEs, the ClawHub marketplace has been hit by the "ClawHavoc" supply-chain attack. Initial audits found 341 malicious skills; updated scans report over 1,184 malicious packages. A single attacker ("hightower6eu") uploaded 354 malicious packages in an automated blitz. Malicious skills deployed Atomic Stealer (AMOS) on macOS and Vidar infostealer on Windows, targeting browser credentials and crypto wallet data.
The structural problem is that ClawHub skills run with full system access and no sandboxing. A malicious skill can write to agent memory and config files, injecting persistent instructions that survive across sessions. Snyk's ToxicSkills audit found that 36.82% of skills had security flaws.
Hermes Agent's Security Profile
As of April 2026, Hermes Agent has zero agent-specific CVEs. An independent security audit of v0.8.0 (812 Python files, ~364K lines of code) found no malware or data exfiltration, describing the code as "well-intentioned." However, the audit identified 4 critical and 9 high-severity findings in the default configuration, primarily because the default security posture is ALLOW-ALL.
Hermes uses a defense-in-depth security model:
- Command Approval System: Pattern-matching detects destructive commands (recursive deletes, permission changes, sudo usage) and triggers approval callbacks.
- Sandboxing Options: Six terminal backends determine where shell commands execute, from local machine to Docker containers with all capabilities dropped, no privilege escalation, and PID limits.
- Credential Protection: Both
execute_codeandterminalstrip sensitive environment variables from child processes.
The curated 118-skill model inherently reduces the attack surface compared to OpenClaw's open marketplace. Self-generated skills are created by the agent itself, eliminating the third-party supply-chain vector entirely.
The Bottom Line on Security
Neither platform is secure out of the box. Both require deliberate configuration, but the nature of the work differs: OpenClaw requires vetting every third-party skill, while Hermes requires tightening its default ALLOW-ALL permission model.
Head-to-Head Comparison Table
| Dimension | OpenClaw | Hermes Agent | Edge |
|---|---|---|---|
| Architecture | Gateway-first, TypeScript/Node.js 22+ | Agent-first, Python 3.11+ | Depends on team stack |
| Learning | Stateless per-session; no autonomous skill creation | Self-improving loop; 40% faster on repeated tasks after 20+ skills | Hermes |
| Ecosystem Size | 44,000+ ClawHub skills, 359K GitHub stars | 118 curated + self-generated skills, 100K+ GitHub stars | OpenClaw (breadth) |
| Memory | SOUL.md session injection; compaction at context limits | Three-layer persistent: prompt memory, episodic skills, session search | Hermes |
| Security (CVEs) | 100+ CVEs, 137 advisories in 3 months; 1,184 malicious ClawHub skills | Zero agent-specific CVEs; default ALLOW-ALL requires hardening | Hermes |
| Skill Origin | Community marketplace (ClawHub) + manual creation | Self-generated from experience + 118 built-in + community (agentskills.io) | Depends on use case |
| Setup Complexity | 2-5 min (npx/Ollama), 30-90 min (VPS self-host) | One-line curl install, self-host only | OpenClaw (managed cloud option) |
| Multi-Agent | Built-in orchestrator, hierarchical/peer-to-peer/orchestrator patterns | ACP delegation, multi-profile; A2A support in progress | OpenClaw |
| Cost Profile | $40-80/mo self-host, $59/mo managed cloud; stateless means full LLM cost each session | $6-65/mo depending on model; learning loop amortizes costs over time | Hermes (at scale) |
| Hybrid Compatibility | MCP + A2A support; can serve as integration gateway | MCP + ACP + A2A (tracking); can serve as learning backend | Both |
Setup Complexity and Developer Experience
Getting Started with OpenClaw
OpenClaw offers the fastest path to a running agent, with setup time varying by method:
- Fastest: 2 minutes with Ollama (
ollama run openclaw) - Standard: ~5 minutes for install, onboarding wizard, and first chat
- Docker self-host: 10-15 minutes
- Full VPS deployment: 30-90 minutes
The onboarding wizard walks through provider configuration and delivers a working chat session quickly. Node.js 22+ and an API key are the core prerequisites. OpenClaw Cloud eliminates setup entirely at $59/month ($29 for the first month).
The longer-term challenge is maintenance. OpenClaw releases 1-2 major point releases per month with frequent breaking changes. Community estimates put DevOps overhead at $10,000-$20,000 per year for production-quality self-hosted instances. Docker permission walls and UID/GID conflicts are consistently cited as the biggest pain point by Reddit users.
Getting Started with Hermes Agent
Hermes Agent installs via a single curl command that handles all dependencies (Python, Node.js, ripgrep, ffmpeg), repository clone, virtual environment, global hermes command setup, and LLM provider configuration:
curl -fsSL https://raw.githubusercontent.com/NousResearch/hermes-agent/main/scripts/install.sh | bashPost-install, hermes model configures the LLM provider, hermes tools manages enabled tools, and hermes gateway setup connects messaging platforms. Hermes requires a model with at least 64,000 tokens of context. Models with smaller windows are rejected at startup.
The platform runs on Linux, macOS, WSL2, and Android (Termux). Native Windows is not supported. There is no managed hosting option; Hermes is self-host only. This means teams must manage their own infrastructure, though the six deployment backends (including Modal serverless with near-zero idle cost) provide flexibility.
Developer Experience Comparison
OpenClaw's TypeScript stack appeals to web developers; Hermes's Python stack aligns with ML/AI practitioners. The deeper DX difference is in the feedback loop. With OpenClaw, the developer experience remains roughly constant over time. With Hermes, the experience improves as the agent builds skills, meaning the initial learning curve pays dividends that compound.
Multi-Agent Orchestration and Cost Comparison
Multi-Agent Capabilities
OpenClaw provides built-in multi-agent orchestration through three collaboration patterns:
- Orchestrator Pattern: A single point of control breaks complex goals into sub-tasks and delegates to freshly-spawned child agents. Sub-agents run concurrently and report back.
- Hierarchical Pattern: A tree structure with a Root Orchestrator managing Sub-Orchestrators, each managing Worker Agents. Suited for large-scale complex tasks.
- Peer-to-Peer Pattern: Message broadcast with consensus, though OpenClaw recommends no more than five agents in this mode.
Agent bindings tell the system which agent handles which channel, enabling multi-persona deployments from a single instance.
Hermes Agent takes a different approach. The acp-delegate skill enables communication with secondary Hermes instances via ACP over stdio JSON-RPC, supporting persistent sessions, tool calls, and conversational memory. Multiple agent profiles can run from a single installation, each with its own configuration, personality, and tool access. Google's A2A protocol is being tracked (Issue #514) for future cross-framework agent interoperability.
For teams needing complex multi-agent workflows today, OpenClaw has a clear lead. For teams focused on deep single-agent capability with occasional delegation, Hermes's model is sufficient.
Cost Modeling
Agents make 5-20x more LLM calls per task compared to single-pass responses, due to iterative planning, tool selection, and error recovery loops. This makes cost modeling critical.
OpenClaw cost structure:
- Self-hosted Docker: $40-80/month (API costs + hosting)
- Managed cloud: $59/month flat (first month $29)
- Every session pays full LLM inference cost since no learning persists
Hermes Agent cost structure:
- Budget setup (Hetzner + DeepSeek V4): $6-8/month total
- Premium setup (Claude Sonnet 4.6): $30-65/month
- Single task: 8-20K input tokens, 1-3K output tokens
- At DeepSeek V4 rates ($0.30/M input): $0.002-$0.006 per task
- At Claude Opus 4.6 rates ($5/M input): $0.04-$0.10 per task
- Learning loop adds 15-25% token overhead but yields 40% savings on repeated tasks
The economic crossover depends on task repetition. For diverse, one-off tasks, OpenClaw's managed cloud at $59/month offers the most predictable pricing. For repetitive workflows, Hermes's learning loop pays for its overhead after roughly 20 accumulated skills, after which each execution becomes progressively cheaper.
Hybrid Use: Running Both Together via ACP Protocol
For teams that do not want to choose, a hybrid architecture is not just possible, it is already happening. OpenClaw agents and Hermes agents can federate today, allowing users to send messages from one to the other, collaborate on shared projects, and delegate tasks across framework boundaries.
The Interoperability Layer
Both frameworks support MCP (Model Context Protocol) for tool discovery and invocation. The A2A (Agent-to-Agent) protocol, created by Google and now under the Linux Foundation with backing from Microsoft, AWS, Cisco, and Salesforce, extends this further. The recommended pattern: MCP for tools, A2A for agents.
Practical Hybrid Architecture
In a hybrid deployment, OpenClaw serves as the integration gateway handling multi-channel routing, session management, and rate limiting. Hermes Agent handles actual task execution, bringing its learning loop and accumulated skills to bear. MCP bridges tool invocation; A2A enables agent-to-agent communication for task delegation.
This combines OpenClaw's strength (broad channel coverage, deterministic routing) with Hermes's strength (deep learning, cross-session intelligence). The Gateway handles the "what channel" question; the learning runtime handles the "how to do it better each time" question.
Practical Considerations
For teams currently on OpenClaw, hermes claw migrate offers migration with dry-run previews. Running both systems adds operational complexity, so the hybrid approach is best suited for organizations with strong DevOps capacity that need both broad integration reach and deep learning capabilities.
When to Use Which: Recommendations by Use Case
Choose OpenClaw When:
- You need broad multi-channel integration fast. OpenClaw's Gateway natively handles WhatsApp, Telegram, Discord, Slack, iMessage, Signal, and web chat. If your primary challenge is reaching users across platforms, OpenClaw solves it out of the box.
- Your team is TypeScript-heavy. The entire plugin and skill ecosystem is TypeScript-native. Web developers can extend the platform without learning a new language.
- You want access to a large skill marketplace. 44,000+ ClawHub skills cover everything from CRM integrations to DevOps pipelines. If your use case is well-served by existing community skills, this is a significant time saver.
- You are building customer-facing chatbots. The gateway pattern, with its centralized rate limiting, session management, and multi-channel routing, aligns naturally with chatbot deployments.
- You prefer managed hosting. OpenClaw Cloud at $59/month eliminates infrastructure management entirely.
Choose Hermes Agent When:
- You need agents that improve over time. The self-improving learning loop delivers measurable 40% efficiency gains on repeated tasks. If your workflows are repetitive and benefit from accumulated knowledge, this is Hermes's primary value proposition.
- Security is paramount. Zero agent-specific CVEs, curated skills, and self-generated skills that eliminate the third-party supply-chain vector make Hermes the safer default for sensitive environments.
- You are building internal automation for ML/AI teams. The Python-native stack, model-agnostic runtime with 18+ providers, and six execution backends align with ML infrastructure patterns.
- Cost efficiency on repetitive tasks matters. The learning loop amortizes LLM costs over time. A budget Hermes instance on Hetzner with DeepSeek V4 runs at $6-8/month, and each repeated task costs less as skills accumulate.
- You need persistent, cross-session memory. The three-layer memory system (prompt memory, episodic skills, session search) enables genuine knowledge accumulation that OpenClaw's context-window-bound approach cannot match.
Choose Both When:
- You need the integration reach of OpenClaw with the learning depth of Hermes Agent. Use OpenClaw as the channel gateway and Hermes as the intelligent backend.
- Your use case spans both broad integration (customer-facing) and deep automation (internal workflows) and you have the DevOps capacity to maintain two systems.
Last updated: April 24, 2026. Data sourced from official documentation, GitHub repositories, independent security audits (Snyk ToxicSkills, CertiK), Joel Gamblin's CVE tracker, Nous Research benchmarks, and community analyses.

