🇨🇳 中文

Hermes Agent v0.10 Review: How 113K Stars in 7 Weeks Hides an Economic Innovation, Not a Technical One

Hermes Agent hit 113K GitHub stars 8 days after shipping v0.10 — the fastest open-source agent framework of 2026. We dissect the three-tier memory, 118 skills, and Tool Gateway to separate marketing ("self-improving") from what Hermes actually does well: bundling the tax of running an agent into one subscription.

Bruce

Hermes AgentNous ResearchAI AgentHarness EngineeringOpen Source Agent

3165  Words

2026-04-24


Hermes Agent v0.10 deep review: the 113K star open-source agent

Seven weeks ago, Hermes Agent did not exist on GitHub. Today, April 24, 2026, it has 113,000 stars. Version 0.10.0 shipped 8 days ago with 118 skills, three-tier memory, Tool Gateway, and six message gateways. Two versions ago it was 27,000 stars. The growth curve from 0 to 113K in 7 weeks makes Hermes the fastest-growing open-source agent framework of 2026, by a margin of 3-4x over OpenClaw’s 7-month climb to the same milestone.

I spent four days running v0.10 end-to-end — source read, deploy, 72-hour live task, memory audit. My conclusion diverges from the official pitch.

Hermes v0.10 is not winning on “self-improvement.” That claim is marketing glue over a prompt-template generator, not a technical moat. Hermes is winning because it is the first open-source agent project to package the economic problem of running an agent — API key management, subscription sprawl, tool configuration overhead — into a single bundled subscription. Tool Gateway is an economic innovation, not a technical one, and that is what MIT-licensed open source has historically been bad at. When Nous Research bundled Firecrawl, FAL, OpenAI TTS, and Browser Use into a $20/month Nous Portal, they quietly ate the indie-developer segment that OpenClaw and LangChain could not monetize. For solo devs who want “one subscription, one agent, everything works” — Hermes just made that product real.

This review separates the 113K-star hype from the mechanics. Three questions drive the rest of the article: what are the three aces in v0.10, does “self-improvement” hold up under audit, and where does Hermes sit relative to Managed Agents and OpenClaw?

0 to 113K Stars in 7 Weeks: What Did Hermes Actually Do Right

Open-source agent frameworks have a known growth shape. OpenClaw took roughly seven months to reach 100K stars. LangChain took over a year. Hermes compressed that to seven weeks. A curve this steep is never pure product quality — it is product quality multiplied by timing multiplied by release cadence. I mapped the release timeline:

timeline
    title Hermes Agent 0 to 113K stars (2026-02-25 to 2026-04-24)
    section Launch
        Feb 25 v0.1.0 : Nous Research open-sources the first release : stars 0
        Mar 5 HN front page : Community discovers three-tier memory : stars ~8K
        Mar 18 v0.5 : MCP support, multi-model switching : stars ~25K
    section Acceleration
        Apr 8 v0.8.0 : 209 PRs, 82 issues closed, background notifications : stars ~68K
        Apr 13 Mobile/Android : Termux, iMessage, WeChat, Fast Mode : stars ~92K
        Apr 16 v0.10.0 + Tool Gateway : 118 skills, three-tier memory, 6 gateways : stars ~108K
    section Breakout
        Apr 20 wshobson comparison post : Community benchmarks against Managed Agents : stars ~112K
        Apr 24 this review : 8 days post-v0.10 : stars 113K+

Three compound factors explain the shape, in order of importance.

Timing against the Harness-layer enclosure. March-April 2026 is the exact window when the agent ecosystem is reconfiguring: April 4 Anthropic banned OpenClaw from Claude Code subscriptions, April 16 Managed Agents launched in beta, April 17 Claude Design shipped, April 20 wshobson/agents 79-plugin breakdown went viral. Every one of those events expanded the population of developers who now believe they need an autonomous agent. Hermes happens to be the most complete, most open, most affordable option in that expanded demand pool. Market attention accumulated to a threshold and Hermes was the vessel that caught it.

Release cadence synchronized with platform events. Hermes shipped v0.8 (April 8), Mobile support (April 13), and v0.10 (April 16) on dates clustered around HN peak hours and, more importantly, on the same day as Anthropic’s Managed Agents beta launch. That is not coincidence. When mainstream coverage lit up around Managed Agents, secondary search demand (“what’s the open-source alternative”) appeared automatically, and Hermes placed itself directly in the answer slot. Traffic piggybacking on a competitor’s launch is an advanced move for a 2-month-old project.

Product decision: no choice paralysis. OpenClaw’s skill ecosystem is larger (1200+ vs Hermes’s 118), but every new OpenClaw skill demands that you provision its API keys, read its docs, handle its rate limits. Hermes v0.10 does the opposite — Tool Gateway collapses the four most-used tools (Firecrawl search, FAL FLUX 2 Pro image gen, OpenAI TTS, Browser Use browser automation) into one Nous Portal subscription. One bill, zero keys, works out of the box. This is product thinking, not engineering thinking. OpenClaw hands you a parts bin; Hermes hands you a car.

The 113K stars is not evidence that Hermes is technically ahead. It is evidence that in a month of ecosystem chaos, Hermes gave ordinary users a “pay the fee, it works” path. That positioning determines the ceiling and the failure modes, which I will come back to.

The Three Aces in v0.10

v0.10.0 shipped a 400-line release notes document. Not all changes matter. Three do — these three carry the weight of Hermes’s current valuation.

Ace 1: 118 Skills, and the Distribution That Matters

The number 118 appears on the website, in release notes, in every Hacker News comment. I spent an afternoon clicking through every entry in the NousResearch/hermes-agent/skills/ directory. The count is accurate. The distribution is not uniform.

CategoryCountExamplesVerdict
Production-grade, high-frequency~60file-ops, git, web-search, browser-automation, markdown-edit, python-exec, shell, memory-opsOxygen for daily agent work, maturity solid
Integration, mid-frequency~30notion-api, linear, slack, github-issues, google-calendar, gmail, trelloUseful if you use the corresponding SaaS
Experimental or redundant~28Three YouTube transcript variants, two RSS parsers, undocumented PoCsCommunity PR surplus, candidates for v0.11 cleanup

My take: 118 is an anchor number. What actually determines your Hermes experience is the 20-30 high-frequency skills you load. Practical advice for new users — disable skills/auto_discover: true on day one and manually enable only the skills you need. Every agent startup pays 3-5K tokens to load the skill manifest; loading the full 118 costs noticeable dollars over a month.

Ace 2: Three-Tier Memory Is the Hardest Engineering in Hermes

Three-tier memory is not Hermes’s invention — Letta and MemGPT predated it. But Hermes’s implementation is the first open-source agent memory system I would call production-deployable. The architecture:

flowchart LR
    U[User Request] --> W[Working Memory
current turn context
~2K token] W -->|every turn| S[Session Memory
task/day history
local SQLite
~20K token] S -->|nightly compress| L[Long-term Memory
cross-session persistent
vector + relational
unbounded] L -->|semantic retrieval| W S -->|hot query| W style W fill:#dbeafe,stroke:#3b82f6,color:#1e3a8a style S fill:#fef3c7,stroke:#f59e0b,color:#78350f style L fill:#dcfce7,stroke:#10b981,color:#064e3b

The roles, as I verified by running the system:

  • Working Memory: single-turn scope, ephemeral, discarded at end of agent.run().
  • Session Memory: task or day scope, stored in SQLite at ~/.hermes/memory.db. You can open it with sqlite3 and inspect rows directly. Schema includes timestamp, task_id, importance_score.
  • Long-term Memory: persistent, backed by Qdrant or Chroma for vector storage plus SQLite for relational metadata. A nightly cron at 00:00 local time compresses Session Memory entries with importance_score > 0.6 into long-term records.

The thing that matters about this design is not the memory itself — it is that the data is visible, auditable, and deletable. You can open the SQLite and see exactly what the agent remembered. You can hermes memory forget <keyword> to delete specific entries. You can rm -rf ~/.hermes/memory.db to reset. Compare that to OpenAI’s ChatGPT Memory (you see a vague summary in Settings, no direct data access), or Claude Code sessions (cleared on /clear, no persistence) — Hermes memory is data the user owns, not a vendor-hosted black box.

This is a production requirement, not a nice-to-have. A legal-tech team I spoke with last week migrated from OpenAI Assistants to Hermes in March, not because Hermes is smarter but because their legal review required “demonstrable control over what the AI system remembers.” The SQLite architecture passed the audit; the Assistants API did not.

Ace 3: Tool Gateway Is an Economic Innovation

Tool Gateway is the piece that the technical press underweights because it is not technically novel. It is a subscription bundle. But for an open-source project’s long-term survival, business model matters more than technical novelty.

Mechanism: subscribe to Nous Portal ($20/month entry tier), and the four highest-frequency tools that normally require separate API keys become automatically available, with no configuration:

ToolUnderlying VendorSelf-serve costNous Portal equivalent
Web SearchFirecrawl$20-50/month (medium)Bundled in $20/month
Image GenFAL FLUX 2 Pro$0.05/image × frequencyBundled in $20/month
TTSOpenAI TTS$0.015/1K charsBundled in $20/month
Browser AutomationBrowser Use$30-60/month (medium)Bundled in $20/month

If your combined spend on those four exceeds $30/month, Nous Portal pays for itself immediately. But the economic calculation is half the story. The other half is cognitive — self-provisioning four vendor keys means four vendor accounts, four quotas, four billing cycles, four credential rotation schedules. Nous Portal collapses that into one monthly charge. For developers who do not want to be DevOps engineers, that value far exceeds $20.

The embedded cost is vendor lock-in. Once you rely on Tool Gateway, migrating to another agent framework means re-provisioning four API keys and re-architecting four rate limits. Nous clearly understands this — Tool Gateway is the smoothest possible onboarding from free-tier Hermes to paid Nous Portal, much better than the typical “free tier feature-crippled” pattern.

This is the first sustainable business model an open-source agent framework has landed. LangChain monetizes via enterprise consulting. OpenClaw monetized via clawdhub enterprise subscriptions (and got enclosed by Anthropic for it). Hermes picked the smartest path available — do not compete with model vendors for model revenue, do not compete with consultancies for enterprise contracts, just capture the “indie developers who do not want to manage API keys” segment. Whether Nous Research can keep the lights on with that alone is the key variable for Hermes’s survival past Q3 2026.

Auditing the Self-Improvement Claim

Hermes’s largest marketing claim is on the landing page: “An autonomous agent that lives on your server, remembers what it learns, and gets more capable the longer it runs.” The feature backing that claim is Auto-Generated Skills.

I ran a 72-hour experiment. Task: scrape 5 technical blogs daily, deduplicate, summarize in English, send to my Telegram. Check every 24 hours what the agent “learned.”

The Mechanism

The Auto-Generated Skills flow is in hermes/core/skill_generation.py:

  1. Problem detection: on task failure or negative user feedback, append failure context + eventual solution to Session Memory.
  2. Pattern extraction: nightly batch at 02:00 scans 24-hour failure records, finds patterns that repeated 3+ times.
  3. Skill generation: for each pattern, call the LLM to produce a new skill markdown template, save to ~/.hermes/skills/auto/.
  4. Skill application: the next similar task loads the new skill.

This is not reinforcement learning. This is prompt-template assembly. Actual self-improvement would update model weights. Hermes never touches weights — it just captures the mapping “when X problem appears, use Y prompt template” and persists it as a markdown file. Semantically equivalent to a hand-written CLAUDE.md entry, except written by the agent instead of by you.

What 72 Hours Produced

Three auto-skills in ~/.hermes/skills/auto/:

  1. blog-dedup-improved.md: detected two blogs cross-publishing the same article. Added URL normalization + title-similarity deduplication. Useful.
  2. telegram-retry.md: Telegram API occasionally returns 429. Added exponential backoff. Useful but basic.
  3. en-summary-length.md: detected that I consistently preferred summaries under 80 words. Tightened the summary prompt’s length constraint. Subtle — adapts to user preference, but is still a prompt tweak.

Three auto-skills after 72 hours. Two engineering patches, one user-preference adaptation. Not a single instance of “acquired a new capability.” Every output was “materialized something that would have been hand-coded into a prompt.”

My take: Auto-Generated Skills is not self-improvement. It is self-documenting CLAUDE.md. The context rules you would have hand-written, Hermes drafts after observing your usage. That has genuine value — it saves you the documentation labor — but it cannot make the agent do things it was structurally incapable of on day one. If your task exceeds Hermes’s ability on day one, it will still exceed it six months later.

Is the “gets more capable the longer it runs” claim a lie? Not a lie, but a definitional bait-and-switch. Hermes gets more aligned to your preferences and more detail-complete over time. It does not acquire capabilities. The two are genuinely different axes of improvement, conflated by marketing.

Operational advice: treat auto-generated skills as “draft CLAUDE.md entries that Hermes wrote for you” — review them weekly, keep the useful ones, delete the rest. Do not trust them as a black box. A weekly 5-minute review of ~/.hermes/skills/auto/ is the ritual that separates getting value from Hermes versus accumulating drift.

Hermes vs Managed Agents vs OpenClaw: A Comparison That Actually Helps

Five-dimension comparison matrix. This is about which one fits which scenario, not which one is “best”:

DimensionHermes v0.10Managed Agents (Anthropic)OpenClaw
DeploymentSelf-hosted ($5 VPS), Nous Portal optionalAnthropic-hosted sandbox, not self-deployableSelf-hosted or clawdhub-hosted
PricingMIT free + API tokens + optional $20/month PortalAPI tokens + managed runtime surcharge (20-40% of token cost)MIT free + API tokens + optional clawdhub enterprise
MemoryThree-tier (working/session/long-term), local SQLite + vectorOfficial memory tool, Anthropic-hostedSingle-layer context + user-managed memory
Tool ecosystem118 official skills + Tool Gateway bundle of 45-6 Anthropic built-in tools + custom tools1200+ clawdhub skills (largest)
ControlFully open source, every layer editable and auditableBlack-box runtime, API config onlyOpen-source harness + optional hosting

Scenario Recommendations

Choose Hermes if you are:

  • An indie dev or small team (< 5 people)
  • Willing to spend $20-50/month on AI tools
  • Seeking “one subscription covers all agent tasks”
  • Not subject to SOC2 / enterprise SLA requirements
  • Comfortable tinkering but unwilling to be full-time DevOps

Choose Managed Agents if you are:

Choose OpenClaw if you:

  • Need the largest skill ecosystem (1200+ vs Hermes’s 118)
  • Already have significant OpenClaw deployment, high migration cost
  • Run enterprise consulting or sell agent runtime (but note Anthropic enclosure risk — see OpenClaw multi-agent setup guide)

Do not choose any of the three if you only need interactive coding with no background autonomous runs and no cross-session memory. Use Claude Code or Cursor directly. Picking up a harness framework is self-inflicted overhead.

Cost Analysis: Two Realistic Usage Profiles

Hermes itself is MIT-free. Your three cost centers are VPS, LLM API, and optional Nous Portal. Two concrete monthly budgets:

Profile A: Light usage (1-2 hours/day of agent work)

Line itemVendorMonthly
VPSHetzner CX22$4
LLMOpenRouter → Claude Haiku or DeepSeek V3$10-15
SearchFirecrawl self-serve key$10 (light)
Total$25-30/month

Skip Nous Portal at this usage level — you will not saturate the subscription.

Profile B: Medium-heavy usage (4-8 hours/day of agent work, including image, voice, browser automation)

Line itemVendorMonthly
VPSHetzner CPX31 or DigitalOcean $20 tier$20
LLMOpenRouter → Claude Sonnet 4.5 + Opus 4.7 mix$60-100
Nous PortalFirecrawl + FAL + TTS + Browser Use bundled$20
Total$100-140/month

Comparison: Managed Agents at equivalent usage lands around $80-120/month (tokens + runtime surcharge), but you still need Browser Use and FAL separately — add roughly $50 to match capability. Hermes is the cheapest autonomous-agent path at medium-heavy usage levels.

The largest cost lever is LLM routing. Hermes supports OpenRouter, which means dynamic routing — simple tasks to DeepSeek V3 (10x cheaper), complex tasks to Claude Sonnet. My tests show 40-60% LLM cost reduction with this routing. This is Hermes’s biggest economic advantage over Managed Agents — Managed Agents locks you to Claude models, no routing option.

Six-Month Outlook: Does Hermes Survive 2026 Q3?

113K stars is exciting. The failure rate for open-source projects only starts to manifest at 6 months. Three concerns, in order of severity:

Concern 1: Can Nous Portal monetize sustainably. Tool Gateway’s economic model is bundling four vendor API costs. Nous needs enough subscribers to cover the bundled cost (Firecrawl, FAL, OpenAI TTS, Browser Use prices are not Nous-controllable). Historical open-source subscription conversion rates sit at 0.5-2%. Optimistic case: Hermes converts 1000-2000 subscribers, $20K-40K/month recurring. Enough to fund a small team. Not enough to fund aggressive hiring or long-term R&D. If conversion falls short, options are price increases, tool cuts, or bleeding the project. 113K stars ≠ 113K subscribers.

Concern 2: Anthropic’s posture. The April 4 OpenClaw enclosure is still fresh. Hermes does not currently threaten Anthropic directly (no enterprise runtime sale), but if Tool Gateway succeeds, user scale grows, and Hermes starts capturing the “indie devs who do not want to manage DevOps” segment that Managed Agents also targets — will Anthropic repeat the enclosure? Probability is lower than for OpenClaw (Hermes uses direct API keys rather than subscription auth, harder to block at protocol layer), but not zero. Risk assessment: medium.

Concern 3: Marketing-reality gap on self-improvement. The “gets more capable” pitch attracted both AI researchers and investors. The moment someone writes a rigorous HN post showing Auto-Generated Skills is prompt-template assembly, not RL, there will be a sentiment correction. It has not happened yet, but as user base and expectations both grow, the gap eventually gets written about. This will not kill the project, but will bend the growth curve from exponential to linear.

My six-month call: Hermes survives past Q3 but does not maintain April’s explosive cadence. The transition from “hot project” to “stable tool” happens around June, when star growth decelerates, early adopters churn back to Managed Agents or wait for Hermes v0.15, and the remaining user base is actual paying Nous Portal subscribers. This is the normal lifecycle for open-source agent projects.

Specific guidance:

  • If you are considering adopting Hermes today: now is the window. v0.10 is capable enough, ecosystem is mature enough, Nous Portal pricing has not gone up yet. Lowest friction moment.
  • If you want “more stable version”: wait for v0.12 (expected mid-to-late May), after a bug-shaking pass on v0.10.
  • If you are an enterprise decision-maker: do not adopt Hermes now — wait until Q3 to see if it survives. Managed Agents is safer.
  • If you are evaluating as an investor: Nous Research’s moat is “experience design + Tool Gateway subscription,” not technology. Whether it is defensible depends on your belief that the business model can lock in users before Anthropic or OpenAI ship a similar subscription.

Further Reading

Related depth:

Primary sources:

Comments

Join the discussion — requires a GitHub account