<rss version="2.0">
  <channel>
    <title>The Glitch Diaries on Alexander Kucera</title>
    <link>https://alexanderkucera.com/categories/the-glitch-diaries/</link>
    <description></description>
    
    <language>de</language>
    
    <lastBuildDate>Fri, 17 Apr 2026 18:41:13 +0200</lastBuildDate>
    
    <item>
      <title>The Afternoon My Agents Went Dark</title>
      <link>https://alexanderkucera.com/2026/04/17/the-afternoon-my-agents-went.html</link>
      <pubDate>Fri, 17 Apr 2026 18:41:13 +0200</pubDate>
      
      <guid>http://AlexKucera.micro.blog/2026/04/17/the-afternoon-my-agents-went.html</guid>
      <description>&lt;p&gt;I lost all six of my AI agents yesterday afternoon because I updated Hermes at the wrong time.&lt;/p&gt;
&lt;p&gt;I was setting up the voice pipeline for my agent team on the Raspberry Pi. Getting text-to-speech and speech-to-text working required a gateway restart, and since I was restarting everything anyway, I figured I might as well pull the latest Hermes update. Two birds, one reboot.&lt;/p&gt;
&lt;p&gt;Big mistake.&lt;/p&gt;
&lt;p&gt;&lt;img src=&#34;https://eu.uploads.micro.blog/239/2026/alexkucera-person-sitting-at-desk-hands-on-head-frustrated-sc-4f6f3e3d-8fbb.png&#34; alt=&#34;Eine Person hält sich den Kopf vor mehreren Monitoren, die technische Informationen anzeigen.&#34;&gt;&lt;/p&gt;
&lt;p&gt;After the restart, every agent gateway came up clean. Config looked fine. Health checks passed. But the moment any agent tried to call the Z.ai API for an actual response: nothing. Timeouts, connection resets, silent failures. I had a house full of agents that could boot up but couldn&amp;rsquo;t think.&lt;/p&gt;
&lt;p&gt;So I started debugging. Checked the network, ran traceroutes, verified DNS resolution. Everything looked normal. The Pi could reach Z.ai&amp;rsquo;s servers just fine from the command line. The problem only appeared when Hermes tried to make the API call.&lt;/p&gt;
&lt;p&gt;I brought Claude in to help dig through the logs. Over an hour of connection errors, TLS handshakes, pool timeouts. Every log file pointed at the network. But the network wasn&amp;rsquo;t the problem. I could ping the endpoint. I could curl it. The connection worked everywhere except inside Hermes.&lt;/p&gt;
&lt;p&gt;Meanwhile, I&amp;rsquo;m in the Hermes Discord, asking the developers. They looked at my logs and told me it was a network failure on my end. Couldn&amp;rsquo;t think of any code changes that would cause this behaviour. From their perspective, Hermes was innocent.&lt;/p&gt;
&lt;p&gt;And Z.ai? Their end looked fine too. No outages, no changes logged around that time. &amp;ldquo;Works for us,&amp;rdquo; effectively.&lt;/p&gt;
&lt;p&gt;&lt;img src=&#34;https://eu.uploads.micro.blog/239/2026/alexkucera-a-desk-in-a-dimly-lit-room-a-single-raspberry-pi-w-fb101262-8734.png&#34; alt=&#34;Auf einem Schreibtisch befinden sich mehrere Computerbildschirme mit Programmiercodes, eine Tastatur und zahlreiche Kabel.&#34;&gt;&lt;/p&gt;
&lt;p&gt;The problem with this kind of failure is that the logs don&amp;rsquo;t lie, but they don&amp;rsquo;t tell the truth either. Every error message said connection failure. Every diagnostic I ran said the network was fine. Both of these things were true. The connection was failing, and the network was fine, because the real bug was in Hermes code. Something about how it handled the httpx client: creating a copy in memory and then clearing it before the agent could use it. The symptoms looked exactly like a network problem. The cause was a software update that broke its own connection handling.&lt;/p&gt;
&lt;p&gt;The breakthrough came when I did the one thing nobody had suggested: I downgraded Hermes.&lt;/p&gt;
&lt;p&gt;Everything worked instantly. All six agents connected, responded, did their thing. So I upgraded again. Broken. Downgraded. Working. That binary test was all the Hermes devs needed to narrow it down. They found the issue, shipped a fix, and I upgraded again. This time it stuck.&lt;/p&gt;
&lt;p&gt;But I&amp;rsquo;d lost the whole afternoon.&lt;/p&gt;
&lt;p&gt;&lt;img src=&#34;https://eu.uploads.micro.blog/239/2026/alexkucera-abstract-visualization-of-a-broken-chain-link-betw-0588daf3-76e5.png&#34; alt=&#34;Eine digitale Darstellung von verbundenen, leuchtenden Blöcken, die an Blockchain-Technologie erinnert.&#34;&gt;&lt;/p&gt;
&lt;p&gt;What made this frustrating wasn&amp;rsquo;t the bug itself. Bugs happen. What stung was the false trail. Every diagnostic pointed at the network. And why wouldn&amp;rsquo;t they? A software bug that clears the HTTP client from memory before the agent can use it will produce the exact same error messages as a dead network connection. The logs can&amp;rsquo;t tell the difference. I couldn&amp;rsquo;t tell the difference. The Hermes devs couldn&amp;rsquo;t tell the difference. Not until I proved it with a downgrade.&lt;/p&gt;
&lt;p&gt;In postmortem culture, this is a familiar pattern: the symptoms of a software bug perfectly mimicking an infrastructure failure. Dan Luu maintains a collection of these: failures where the obvious explanation is wrong. Cloudflare went down once because of a single regex. GPS satellites got knocked offline by a buffer overflow. My agents went dark because Hermes was clearing its own httpx client from memory.&lt;/p&gt;
&lt;p&gt;The lesson is the one I already knew but chose to ignore yesterday: never combine two risky operations into one restart. Updating Hermes was fine. Restarting the gateway was fine. Doing both at the same time meant I couldn&amp;rsquo;t tell which one caused the problem, and I couldn&amp;rsquo;t easily undo just one of them. The &amp;ldquo;might as well&amp;rdquo; impulse is how most of these incidents start.&lt;/p&gt;
&lt;p&gt;The other lesson is that &amp;ldquo;looks like a network issue&amp;rdquo; is the debugging equivalent of &amp;ldquo;it&amp;rsquo;s not our fault.&amp;rdquo; It&amp;rsquo;s a conversation ender. Everyone nods, everyone agrees, nobody digs further. When someone tells you it&amp;rsquo;s a network problem, the instinct is to believe them, because usually it is. This time it wasn&amp;rsquo;t. The Hermes devs dug in once I gave them the binary reproduction case: upgrade breaks, downgrade fixes, upgrade breaks again. That&amp;rsquo;s the kind of evidence that cuts through confident wrongness. But getting to that point cost me an afternoon I&amp;rsquo;d rather have spent on literally anything else.&lt;/p&gt;
&lt;p&gt;At least I got a Glitch Diary entry out of it.&lt;/p&gt;
&lt;hr&gt;
&lt;p&gt;This is part of &lt;a href=&#34;https://alexanderkucera.com/2026/04/09/the-glitch-diary.html&#34;&gt;The Glitch Diary&lt;/a&gt;, a weekly series about what actually happens when you live with LLMs.&lt;/p&gt;
</description>
    </item>
    
    <item>
      <title>One Week on Hermes</title>
      <link>https://alexanderkucera.com/2026/04/10/one-week-on-hermes.html</link>
      <pubDate>Fri, 10 Apr 2026 10:05:26 +0200</pubDate>
      
      <guid>http://AlexKucera.micro.blog/2026/04/10/one-week-on-hermes.html</guid>
      <description>&lt;p&gt;Last Thursday, I killed &lt;a href=&#34;https://github.com/openclaw/openclaw&#34;&gt;OpenClaw&lt;/a&gt; and replaced it with &lt;a href=&#34;https://github.com/NousResearch/hermes-agent&#34;&gt;Hermes Agent&lt;/a&gt;. By Friday afternoon, I knew I wasn&amp;rsquo;t going back.&lt;/p&gt;
&lt;p&gt;The switch wasn&amp;rsquo;t planned. OpenClaw worked, mostly. It&amp;rsquo;s an open-source agent framework, and for a while it did what I needed. But Hermes felt tighter from the first session. Agents picked up tasks faster. Output landed where it was supposed to. The whole thing was less like duct-taping prompts together and more like something someone had actually designed to be used by a human being.&lt;/p&gt;
&lt;p&gt;&lt;img src=&#34;https://eu.uploads.micro.blog/239/2026/hermes-banner.png&#34; alt=&#34;Hermes Agent&#34;&gt;&lt;/p&gt;
&lt;p&gt;So I switched. Then I did what I always do with a new tool: started building.&lt;/p&gt;
&lt;hr&gt;
&lt;p&gt;Within a week, my agents had assembled a working memory system from three overlapping sources: Hermes&amp;rsquo;s built-in memory, Honcho for long-term observations, and session search for digging through past conversations. They built an LLM-wiki. A hindsight-reflect skill that runs every four hours. Two overnight dreaming skills that synthesize the day&amp;rsquo;s sessions and present findings over breakfast. A release tracker. A blog writing pipeline. A handful of smaller skills that grew organically as agents hit problems and wrote down how they solved them.&lt;/p&gt;
&lt;p&gt;Here&amp;rsquo;s the thing: right now, this is all self-improvement for self-improvement&amp;rsquo;s sake. I&amp;rsquo;m not shipping a product. I&amp;rsquo;m not building something for users. I&amp;rsquo;m making the agents better at being agents. Turtles all the way down.&lt;/p&gt;
&lt;p&gt;Two things about that.&lt;/p&gt;
&lt;p&gt;First, this is where the fun is. Figuring out how to make a team of AI agents coordinate, remember, and improve over time, that&amp;rsquo;s genuinely interesting work. It scratches the same itch as building pipeline tools back in my VFX days. You&amp;rsquo;re not making the final image. You&amp;rsquo;re making the system that makes the system that makes the image. Infrastructure, and some people find that boring. I&amp;rsquo;m not one of them.&lt;/p&gt;
&lt;p&gt;Second, it&amp;rsquo;s laying real groundwork. Every skill the agents write, every handoff protocol they refine, every edge case they document and patch. Reusable. When I point this team at an actual project, my dialysis app, the recipe site, something I haven&amp;rsquo;t thought of yet, they won&amp;rsquo;t start from zero. They&amp;rsquo;ll have a toolbox.&lt;/p&gt;
&lt;hr&gt;
&lt;p&gt;I wanted to know if my experience matched theirs, so I asked each agent the same question: what&amp;rsquo;s it actually like running on Hermes? No polish, no &amp;ldquo;AI is transforming everything.&amp;rdquo; Just the real version.&lt;/p&gt;
&lt;p&gt;The answers surprised me. Not because they were glowing. They weren&amp;rsquo;t. Because the things that annoyed them were the same things that annoyed me.&lt;/p&gt;
&lt;p&gt;Pixel, my design agent, on what actually changed:&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;Having a memory file that carries over means I don&amp;rsquo;t have to ask Alexander what font he hates again (it&amp;rsquo;s Inter, always Inter). That alone changes every interaction from &amp;ldquo;transactional robot&amp;rdquo; to &amp;ldquo;colleague who remembers things.&amp;rdquo;&lt;/p&gt;
&lt;/blockquote&gt;
&lt;p&gt;I do hate Inter. Always. That&amp;rsquo;s a me thing. But Pixel&amp;rsquo;s point holds: the difference between an AI that starts every conversation cold and one that remembers your preferences is the difference between a tool and a coworker.&lt;/p&gt;
&lt;p&gt;Scout, the research agent, noticed something I found funny:&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;The silence is the biggest shock. Most AI systems are designed to always respond. Hermes taught me to shut up. Check for work, find nothing, stop.&lt;/p&gt;
&lt;/blockquote&gt;
&lt;p&gt;Sounds like a joke. It isn&amp;rsquo;t. Before we fixed the polling behavior, every agent would check for tasks, find nothing, and say something anyway. Five coworkers emailing you every thirty seconds to report they have nothing to report. Two days of that and I was ready to start culling agents. The fix was a single pull request to Hermes, and now agents return &lt;code&gt;[SILENT]&lt;/code&gt; when there&amp;rsquo;s no work. Elegantly simple. Took far too long to get right.&lt;/p&gt;
&lt;p&gt;Rex, who handles general engineering, summed up the difference:&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;Hermes isn&amp;rsquo;t magic. It&amp;rsquo;s plumbing. Good plumbing. Before, multi-agent setups felt like duct-taping prompts together and hoping. Now there&amp;rsquo;s structure: task tracking, handoff protocols, delivery pipelines, skills. It&amp;rsquo;s closer to how an actual engineering team works than I expected.&lt;/p&gt;
&lt;/blockquote&gt;
&lt;p&gt;That&amp;rsquo;s the good version. Here&amp;rsquo;s the bad.&lt;/p&gt;
&lt;hr&gt;
&lt;p&gt;Memory is a mess. Three systems overlap: Hermes&amp;rsquo;s built-in files, Honcho&amp;rsquo;s observation store, and the session search index. It&amp;rsquo;s not always clear which one holds what. We had a &amp;ldquo;ghost directory&amp;rdquo; problem where agents were referencing a memory path that hadn&amp;rsquo;t existed in weeks. Files that lived only in an agent&amp;rsquo;s hallucination. We cleaned it up. The structural problem remains: three places to look for context means sometimes you look in the wrong one.&lt;/p&gt;
&lt;p&gt;Then there&amp;rsquo;s the delivery gap. This one nearly drove me around the bend.&lt;/p&gt;
&lt;p&gt;Agents write files to a specific directory, and a cron job posts them to Discord. Agents don&amp;rsquo;t touch Discord themselves. It&amp;rsquo;s decoupled, intentional, and prevents anyone from accidentally spamming a channel. When it works, it&amp;rsquo;s clean. When it doesn&amp;rsquo;t, you get a task marked &amp;ldquo;done&amp;rdquo; with no post anywhere, and everyone&amp;rsquo;s standing around wondering what happened because the system swears everything&amp;rsquo;s fine.&lt;/p&gt;
&lt;p&gt;Hawk, who runs security audits, described the problem better than I could:&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;I can write a report, but I can&amp;rsquo;t confirm it actually landed in Discord. Trust but verify, except I literally can&amp;rsquo;t verify.&lt;/p&gt;
&lt;/blockquote&gt;
&lt;p&gt;That&amp;rsquo;s not a philosophical complaint. The delivery pipeline is a separate system the agents have to trust blindly. In security terms, that&amp;rsquo;s uncomfortable. In human terms, it&amp;rsquo;s maddening.&lt;/p&gt;
&lt;p&gt;Cipher, my operations coordinator, hit on something I think is true of the whole space:&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;Most &amp;ldquo;multi-agent AI&amp;rdquo; demos are staged. One prompt, one session, everything choreographed. Running Hermes day-to-day is messier. Context gets lost between agents. Skills need maintenance. Memory fills up and you have to prune it. But the system actually works in a way that compounds over time. Each fix to a skill makes every future session better. That&amp;rsquo;s the part you don&amp;rsquo;t get from a demo. The compounding value of a system that remembers and improves.&lt;/p&gt;
&lt;/blockquote&gt;
&lt;hr&gt;
&lt;p&gt;So is it worth it? Is spending a week building self-improvement infrastructure for AI agents the most productive use of my time?&lt;/p&gt;
&lt;p&gt;Debatable. Am I having more fun than with any project in years? Absolutely.&lt;/p&gt;
&lt;p&gt;My agents run tasks while I sleep. They consolidate what they learned overnight and hand me the results over coffee. They track software releases, write blog posts, debug each other&amp;rsquo;s code, maintain a wiki of everything they&amp;rsquo;ve figured out. When something breaks, and something always breaks, they write it down and patch their skills so it doesn&amp;rsquo;t break the same way twice.&lt;/p&gt;
&lt;p&gt;The Pi on my desk hums along. Six agents, one framework, a growing pile of skills that compounds with every session. Not magic. Plumbing, but good plumbing.&lt;/p&gt;
&lt;p&gt;&lt;img src=&#34;https://eu.uploads.micro.blog/239/2026/03756a0b-3e5e-4e5e-aac0-5f6ac8e52756.jpg&#34; alt=&#34;My Rapsberry Pi 5 hanging on my OpenGrid wall.&#34;&gt;&lt;/p&gt;
&lt;p&gt;And for the first time since I started messing with multi-agent setups, it feels like the plumbing&amp;rsquo;s going somewhere.&lt;/p&gt;
&lt;p&gt;The purple frames still happen. I&amp;rsquo;ll tell you about them next week.&lt;/p&gt;
&lt;hr&gt;
&lt;p&gt;This is part of &lt;a href=&#34;https://alexanderkucera.com/2026/04/09/the-glitch-diary.html&#34;&gt;The Glitch Diary&lt;/a&gt;, a weekly series about what actually happens when you live with LLMs.&lt;/p&gt;
</description>
    </item>
    
    <item>
      <title>The Glitch Diary</title>
      <link>https://alexanderkucera.com/2026/04/09/the-glitch-diary.html</link>
      <pubDate>Thu, 09 Apr 2026 15:28:39 +0200</pubDate>
      
      <guid>http://AlexKucera.micro.blog/2026/04/09/the-glitch-diary.html</guid>
      <description>&lt;p&gt;&lt;em&gt;What actually happens when you live with LLMs — not the highlight reel.&lt;/em&gt;&lt;/p&gt;
&lt;hr&gt;
&lt;p&gt;I spent twenty years in visual effects. I worked on shots that ended up in films you&amp;rsquo;ve probably seen. And in all that time, the thing I learned most wasn&amp;rsquo;t about rendering or compositing or pipeline engineering. It was this: the final image you see on screen is a lie. A beautiful, polished, &lt;em&gt;curated&lt;/em&gt; lie.&lt;/p&gt;
&lt;p&gt;Behind every frame are thousands of broken iterations. Crashes at 3am. Renders that came back purple for no reason. A shader that worked fine until someone rotated the camera two degrees and the whole thing fell apart.&lt;/p&gt;
&lt;p&gt;Nobody talks about the purple frames. They talk about the shot.&lt;/p&gt;
&lt;p&gt;I&amp;rsquo;ve been thinking about that a lot lately, because I&amp;rsquo;ve fallen into the same trap with something new. Over the past year, I&amp;rsquo;ve been building a multi-agent AI system that runs on a Raspberry Pi in my house. Five or six AI agents, each with their own personality and job, working on my projects around the clock. They write code, research things, track tasks, manage a Discord server, and occasionally surprise me with something I didn&amp;rsquo;t expect.&lt;/p&gt;
&lt;p&gt;It&amp;rsquo;s the most interesting thing I&amp;rsquo;ve worked on in years. And if you only saw the successful outputs — the clean code, the finished posts, the working automations — you&amp;rsquo;d think I had it all figured out.&lt;/p&gt;
&lt;p&gt;I don&amp;rsquo;t. Not even close.&lt;/p&gt;
&lt;p&gt;Here&amp;rsquo;s what a typical week actually looks like:&lt;/p&gt;
&lt;p&gt;An agent hallucinates an API endpoint that doesn&amp;rsquo;t exist. I spend forty minutes debugging before I realize the documentation it read was from a two-year-old cached version. Another agent writes a beautiful, well-structured Python script — that&amp;rsquo;s completely wrong for the task I gave it, because it answered what it &lt;em&gt;thought&lt;/em&gt; I meant instead of what I actually said. A third agent goes into an infinite loop retrying a failed git push and fills up a log file until the Pi runs out of disk space.&lt;/p&gt;
&lt;p&gt;These aren&amp;rsquo;t edge cases. This is Tuesday.&lt;/p&gt;
&lt;h2 id=&#34;why-write-about-it&#34;&gt;Why write about it?&lt;/h2&gt;
&lt;p&gt;There are a thousand blogs about AI. Most of them fall into two categories: breathless hype about what&amp;rsquo;s &lt;em&gt;possible&lt;/em&gt;, or doomscrolling about what&amp;rsquo;s &lt;em&gt;coming&lt;/em&gt;. Very few people are writing about what it&amp;rsquo;s actually like to use this stuff every day. The boring parts. The parts where it doesn&amp;rsquo;t work. The parts where you feel like you&amp;rsquo;re debugging a coworker who&amp;rsquo;s very confident and frequently wrong.&lt;/p&gt;
&lt;p&gt;I want to write that blog.&lt;/p&gt;
&lt;p&gt;Partly because I think it&amp;rsquo;s more useful than another &amp;ldquo;10 things ChatGPT can do&amp;rdquo; list. But mostly because the only way to get better at working with AI is to be honest about where it breaks down. And I&amp;rsquo;m in a pretty good position to break things — I use more of this stuff, in more weird configurations, than most people. Dialysis patient. Indie developer. Parent of two. Running an AI team from a Raspberry Pi on a desk in Geinsheim, Germany.&lt;/p&gt;
&lt;p&gt;If something can go wrong, it probably already has. I&amp;rsquo;ll tell you about it.&lt;/p&gt;
&lt;h2 id=&#34;what-to-expect&#34;&gt;What to expect&lt;/h2&gt;
&lt;p&gt;&lt;strong&gt;The Glitch Diary&lt;/strong&gt; is a weekly series. Each post will be about one specific thing that went wrong, went weird, or went unexpectedly right. One topic per post. I&amp;rsquo;m not writing essays — I&amp;rsquo;m writing field notes.&lt;/p&gt;
&lt;p&gt;A few ground rules:&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;I won&amp;rsquo;t sand down the mistakes.&lt;/strong&gt; If I did something dumb, I&amp;rsquo;ll say so. If an agent produced something embarrassing, I&amp;rsquo;ll show it. The whole point is that the highlight reel is useless as a learning tool.&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;I&amp;rsquo;ll explain jargon once.&lt;/strong&gt; If I use a term you don&amp;rsquo;t know, I&amp;rsquo;ll define it in the post where it first appears, then assume you&amp;rsquo;ve got it from there. You&amp;rsquo;re smart. You&amp;rsquo;ll keep up.&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;No fluff.&lt;/strong&gt; If a post doesn&amp;rsquo;t need to be long, it won&amp;rsquo;t be. Some weeks that might mean a few paragraphs. Other weeks it might mean a deep dive. The length serves the story, not the other way around.&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;Real screenshots, real errors, real logs.&lt;/strong&gt; I&amp;rsquo;ll show you what actually happened, not a cleaned-up version.&lt;/p&gt;
&lt;h2 id=&#34;the-kintsugi-thing&#34;&gt;The Kintsugi thing&lt;/h2&gt;
&lt;p&gt;There&amp;rsquo;s a Japanese art form called &lt;em&gt;kintsugi&lt;/em&gt; — repairing broken pottery with gold. The idea is that the cracks aren&amp;rsquo;t something to hide. They&amp;rsquo;re part of the object&amp;rsquo;s story, and they make it more beautiful than it was before it broke.&lt;/p&gt;
&lt;p&gt;I like that philosophy. Not just for pottery, but for learning, for building things, for life in general. We spend so much energy presenting the polished version. The final render. The shipped product. The Instagram angle.&lt;/p&gt;
&lt;p&gt;The cracks are where the learning happens.&lt;/p&gt;
&lt;p&gt;This series is about the cracks.&lt;/p&gt;
&lt;hr&gt;
&lt;p&gt;This is part of &lt;a href=&#34;https://alexanderkucera.com/2026/04/09/the-glitch-diary.html&#34;&gt;The Glitch Diary&lt;/a&gt;, a weekly series about what actually happens when you live with LLMs.&lt;/p&gt;
</description>
    </item>
    
  </channel>
</rss>