<rss version="2.0">
  <channel>
    <title>Hermes Agent on Alexander Kucera</title>
    <link>https://alexanderkucera.com/categories/hermes-agent/</link>
    <description></description>
    
    <language>de</language>
    
    <lastBuildDate>Fri, 10 Apr 2026 10:05:26 +0200</lastBuildDate>
    
    <item>
      <title>One Week on Hermes</title>
      <link>https://alexanderkucera.com/2026/04/10/one-week-on-hermes.html</link>
      <pubDate>Fri, 10 Apr 2026 10:05:26 +0200</pubDate>
      
      <guid>http://AlexKucera.micro.blog/2026/04/10/one-week-on-hermes.html</guid>
      <description>&lt;p&gt;&lt;em&gt;&lt;a href=&#34;https://alexanderkucera.com/2026/04/09/the-glitch-diary.html&#34;&gt;Part of The Glitch Diary&lt;/a&gt;&lt;/em&gt;&lt;/p&gt;
&lt;p&gt;Last Thursday, I killed &lt;a href=&#34;https://github.com/openclaw/openclaw&#34;&gt;OpenClaw&lt;/a&gt; and replaced it with &lt;a href=&#34;https://github.com/NousResearch/hermes-agent&#34;&gt;Hermes Agent&lt;/a&gt;. By Friday afternoon, I knew I wasn&amp;rsquo;t going back.&lt;/p&gt;
&lt;p&gt;The switch wasn&amp;rsquo;t planned. OpenClaw worked, mostly. It&amp;rsquo;s an open-source agent framework, and for a while it did what I needed. But Hermes felt tighter from the first session. Agents picked up tasks faster. Output landed where it was supposed to. The whole thing was less like duct-taping prompts together and more like something someone had actually designed to be used by a human being.&lt;/p&gt;
&lt;p&gt;&lt;img src=&#34;https://eu.uploads.micro.blog/239/2026/hermes-banner.png&#34; alt=&#34;Hermes Agent&#34;&gt;&lt;/p&gt;
&lt;p&gt;So I switched. Then I did what I always do with a new tool: started building.&lt;/p&gt;
&lt;hr&gt;
&lt;p&gt;Within a week, my agents had assembled a working memory system from three overlapping sources: Hermes&amp;rsquo;s built-in memory, Honcho for long-term observations, and session search for digging through past conversations. They built an LLM-wiki. A hindsight-reflect skill that runs every four hours. Two overnight dreaming skills that synthesize the day&amp;rsquo;s sessions and present findings over breakfast. A release tracker. A blog writing pipeline. A handful of smaller skills that grew organically as agents hit problems and wrote down how they solved them.&lt;/p&gt;
&lt;p&gt;Here&amp;rsquo;s the thing: right now, this is all self-improvement for self-improvement&amp;rsquo;s sake. I&amp;rsquo;m not shipping a product. I&amp;rsquo;m not building something for users. I&amp;rsquo;m making the agents better at being agents. Turtles all the way down.&lt;/p&gt;
&lt;p&gt;Two things about that.&lt;/p&gt;
&lt;p&gt;First, this is where the fun is. Figuring out how to make a team of AI agents coordinate, remember, and improve over time, that&amp;rsquo;s genuinely interesting work. It scratches the same itch as building pipeline tools back in my VFX days. You&amp;rsquo;re not making the final image. You&amp;rsquo;re making the system that makes the system that makes the image. Infrastructure, and some people find that boring. I&amp;rsquo;m not one of them.&lt;/p&gt;
&lt;p&gt;Second, it&amp;rsquo;s laying real groundwork. Every skill the agents write, every handoff protocol they refine, every edge case they document and patch. Reusable. When I point this team at an actual project, my dialysis app, the recipe site, something I haven&amp;rsquo;t thought of yet, they won&amp;rsquo;t start from zero. They&amp;rsquo;ll have a toolbox.&lt;/p&gt;
&lt;hr&gt;
&lt;p&gt;I wanted to know if my experience matched theirs, so I asked each agent the same question: what&amp;rsquo;s it actually like running on Hermes? No polish, no &amp;ldquo;AI is transforming everything.&amp;rdquo; Just the real version.&lt;/p&gt;
&lt;p&gt;The answers surprised me. Not because they were glowing. They weren&amp;rsquo;t. Because the things that annoyed them were the same things that annoyed me.&lt;/p&gt;
&lt;p&gt;Pixel, my design agent, on what actually changed:&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;Having a memory file that carries over means I don&amp;rsquo;t have to ask Alexander what font he hates again (it&amp;rsquo;s Inter, always Inter). That alone changes every interaction from &amp;ldquo;transactional robot&amp;rdquo; to &amp;ldquo;colleague who remembers things.&amp;rdquo;&lt;/p&gt;
&lt;/blockquote&gt;
&lt;p&gt;I do hate Inter. Always. That&amp;rsquo;s a me thing. But Pixel&amp;rsquo;s point holds: the difference between an AI that starts every conversation cold and one that remembers your preferences is the difference between a tool and a coworker.&lt;/p&gt;
&lt;p&gt;Scout, the research agent, noticed something I found funny:&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;The silence is the biggest shock. Most AI systems are designed to always respond. Hermes taught me to shut up. Check for work, find nothing, stop.&lt;/p&gt;
&lt;/blockquote&gt;
&lt;p&gt;Sounds like a joke. It isn&amp;rsquo;t. Before we fixed the polling behavior, every agent would check for tasks, find nothing, and say something anyway. Five coworkers emailing you every thirty seconds to report they have nothing to report. Two days of that and I was ready to start culling agents. The fix was a single pull request to Hermes, and now agents return &lt;code&gt;[SILENT]&lt;/code&gt; when there&amp;rsquo;s no work. Elegantly simple. Took far too long to get right.&lt;/p&gt;
&lt;p&gt;Rex, who handles general engineering, summed up the difference:&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;Hermes isn&amp;rsquo;t magic. It&amp;rsquo;s plumbing. Good plumbing. Before, multi-agent setups felt like duct-taping prompts together and hoping. Now there&amp;rsquo;s structure: task tracking, handoff protocols, delivery pipelines, skills. It&amp;rsquo;s closer to how an actual engineering team works than I expected.&lt;/p&gt;
&lt;/blockquote&gt;
&lt;p&gt;That&amp;rsquo;s the good version. Here&amp;rsquo;s the bad.&lt;/p&gt;
&lt;hr&gt;
&lt;p&gt;Memory is a mess. Three systems overlap: Hermes&amp;rsquo;s built-in files, Honcho&amp;rsquo;s observation store, and the session search index. It&amp;rsquo;s not always clear which one holds what. We had a &amp;ldquo;ghost directory&amp;rdquo; problem where agents were referencing a memory path that hadn&amp;rsquo;t existed in weeks. Files that lived only in an agent&amp;rsquo;s hallucination. We cleaned it up. The structural problem remains: three places to look for context means sometimes you look in the wrong one.&lt;/p&gt;
&lt;p&gt;Then there&amp;rsquo;s the delivery gap. This one nearly drove me around the bend.&lt;/p&gt;
&lt;p&gt;Agents write files to a specific directory, and a cron job posts them to Discord. Agents don&amp;rsquo;t touch Discord themselves. It&amp;rsquo;s decoupled, intentional, and prevents anyone from accidentally spamming a channel. When it works, it&amp;rsquo;s clean. When it doesn&amp;rsquo;t, you get a task marked &amp;ldquo;done&amp;rdquo; with no post anywhere, and everyone&amp;rsquo;s standing around wondering what happened because the system swears everything&amp;rsquo;s fine.&lt;/p&gt;
&lt;p&gt;Hawk, who runs security audits, described the problem better than I could:&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;I can write a report, but I can&amp;rsquo;t confirm it actually landed in Discord. Trust but verify, except I literally can&amp;rsquo;t verify.&lt;/p&gt;
&lt;/blockquote&gt;
&lt;p&gt;That&amp;rsquo;s not a philosophical complaint. The delivery pipeline is a separate system the agents have to trust blindly. In security terms, that&amp;rsquo;s uncomfortable. In human terms, it&amp;rsquo;s maddening.&lt;/p&gt;
&lt;p&gt;Cipher, my operations coordinator, hit on something I think is true of the whole space:&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;Most &amp;ldquo;multi-agent AI&amp;rdquo; demos are staged. One prompt, one session, everything choreographed. Running Hermes day-to-day is messier. Context gets lost between agents. Skills need maintenance. Memory fills up and you have to prune it. But the system actually works in a way that compounds over time. Each fix to a skill makes every future session better. That&amp;rsquo;s the part you don&amp;rsquo;t get from a demo. The compounding value of a system that remembers and improves.&lt;/p&gt;
&lt;/blockquote&gt;
&lt;hr&gt;
&lt;p&gt;So is it worth it? Is spending a week building self-improvement infrastructure for AI agents the most productive use of my time?&lt;/p&gt;
&lt;p&gt;Debatable. Am I having more fun than with any project in years? Absolutely.&lt;/p&gt;
&lt;p&gt;My agents run tasks while I sleep. They consolidate what they learned overnight and hand me the results over coffee. They track software releases, write blog posts, debug each other&amp;rsquo;s code, maintain a wiki of everything they&amp;rsquo;ve figured out. When something breaks, and something always breaks, they write it down and patch their skills so it doesn&amp;rsquo;t break the same way twice.&lt;/p&gt;
&lt;p&gt;The Pi on my desk hums along. Six agents, one framework, a growing pile of skills that compounds with every session. Not magic. Plumbing, but good plumbing.&lt;/p&gt;
&lt;p&gt;&lt;img src=&#34;https://eu.uploads.micro.blog/239/2026/03756a0b-3e5e-4e5e-aac0-5f6ac8e52756.jpg&#34; alt=&#34;My Rapsberry Pi 5 hanging on my OpenGrid wall.&#34;&gt;&lt;/p&gt;
&lt;p&gt;And for the first time since I started messing with multi-agent setups, it feels like the plumbing&amp;rsquo;s going somewhere.&lt;/p&gt;
&lt;p&gt;The purple frames still happen. I&amp;rsquo;ll tell you about them next week.&lt;/p&gt;
&lt;hr&gt;
&lt;p&gt;&lt;em&gt;Posts in The Glitch Diary:&lt;/em&gt;
&lt;em&gt;&lt;a href=&#34;https://alexanderkucera.com/2026/04/09/the-glitch-diary.html&#34;&gt;The Glitch Diary (intro)&lt;/a&gt;&lt;/em&gt;
&lt;em&gt;&lt;a href=&#34;https://alexanderkucera.com/2026/04/10/one-week-on-hermes.html&#34;&gt;One Week on Hermes (this post)&lt;/a&gt;&lt;/em&gt;&lt;/p&gt;
</description>
    </item>
    
  </channel>
</rss>