<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0"
     xmlns:atom="http://www.w3.org/2005/Atom"
     xmlns:content="http://purl.org/rss/1.0/modules/content/"
     xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>Shifting Ground — Dispatches</title>
    <link>https://chapsoft.com/dispatches</link>
    <description>Long-form dispatches on AI transformation, career disruption, and ground truth. By Russ Kohn.</description>
    <language>en-us</language>
    <managingEditor>info@chapsoft.com (Russ Kohn)</managingEditor>
    <webMaster>info@chapsoft.com (Russ Kohn)</webMaster>
    <lastBuildDate>Wed, 08 Apr 2026 10:51:00 -0700</lastBuildDate>
    <atom:link href="https://chapsoft.com/dispatches/feed.xml" rel="self" type="application/rss+xml"/>

    <item>
      <title>Why AI Projects Fail — What the Evidence Says When You Trace the Citations</title>
      <link>https://chapsoft.com/dispatches/why-ai-projects-fail</link>
      <description>Everyone cites the failure rates. 80%. 85%. 95%. When you actually trace the citations, a different picture emerges. Only 5-12% of organizations achieve significant financial impact from AI.</description>
      <content:encoded><![CDATA[<p>I'm sitting in my backyard, phone in hand, dictating instructions to an AI running on a server in the next room. The sky is the kind of clear you get in the desert when the wind dies down. My dog is asleep next to me, twitching at something in a dream.</p>

        <p>Over the past couple of hours, a structured research process has produced higher-quality data than I could have gotten any other way — even a few months ago. Fifty-plus sources, structured evidence tiers, convergence analysis across independent studies. The kind of work that would have taken a research team weeks.</p>

        <p>And the first thing that research told me was that the statistics I'd been drafting for my own website wouldn't survive scrutiny.</p>

        <p>If you lead an AI initiative, evaluate one for investment, or just need to know whether the claims being made to you are real — this is what the evidence says when you actually trace the citations.</p>

        <h2>The Numbers Everyone Cites</h2>

        <p>If you've read anything about AI adoption in the past two years, you've seen the numbers. <em>More than 80% of AI projects fail.</em> That one's attributed to RAND. <em>55% of employers regret their AI-related layoffs.</em> That's Forrester.</p>

        <p>I had both in my draft. They supported my argument. They felt authoritative. RAND is RAND. Forrester is Forrester.</p>

        <p>When I built the site, I used AI to research those statistics. The citations looked solid. I moved on — there was always something more pressing to build. Later, before publishing, I ran an adversarial review: a different AI model doing a deliberately oppositional assessment of every claim on the site. It flagged the stats. <em>Are these properly sourced? Can you trace them to methodology?</em></p>

        <p>The process worked — it caught the problem. But my first instinct wasn't to question the numbers. It was to tighten the citations and move on. The numbers felt so well-grounded that resisting the feedback felt rational. It's always easier to move forward than to stop and verify.</p>

        <p>It wasn't until I decided to run a formal assessment — the same structured pipeline I built for client work — that the citation chain actually unraveled.</p>

        <p>The 80% doesn't come from RAND's research. RAND's 2024 report uses the phrase "by some estimates" and cites the number without generating it. The trail leads back to a Gartner <em>forecast</em> from 2018 — a prediction about what might happen, not a measurement of what did. From there it spread: a VentureBeat article reported it as "87%" based on a conference talk. Other publications rounded it to 85%. By 2024 it had become received wisdom, sourced to institutions that never produced it.</p>

        <p>The 55% linked to a blog post summarizing a Forrester report that sits behind a paywall with undisclosed methodology. The actual sample size for that specific figure has never been publicly confirmed.</p>

        <p>A scoping review published on SSRN in August 2025 examined the major failure-rate studies and concluded: "None of the sources employs probability sampling or standardized outcome definitions suitable for population-level prevalence claims."<a href="https://papers.ssrn.com/sol3/papers.cfm?abstract_id=5459054" target="_blank" rel="noopener" class="cite-link"><sup>[1]</sup></a></p>

        <p>The adversarial review caught the problem before anything went live. The formal pipeline produced better numbers. The process worked. But the resistance I felt — the pull to trust numbers that <em>felt</em> authoritative — that's not a personal failing. It's a calibration problem that everyone using these tools faces. I'm not immune. Nobody is.</p>

        <h2>What the Evidence Actually Says</h2>

        <p>When you stop trusting the headline numbers and look at what the studies actually measured, a different picture emerges. It's less dramatic than "80% fail" and more troubling.</p>

        <p><strong>The concentration finding.</strong> Only 5–12% of organizations achieve significant enterprise-level financial impact from AI. McKinsey's 2025 survey of 1,993 respondents across 105 countries found 6%.<a href="https://www.mckinsey.com/capabilities/quantumblack/our-insights/the-state-of-ai" target="_blank" rel="noopener" class="cite-link"><sup>[2]</sup></a> BCG found 5% in a separate survey of 1,250 executives.<a href="https://www.bcg.com/publications/2025/are-you-generating-value-from-ai-the-widening-gap" target="_blank" rel="noopener" class="cite-link"><sup>[3]</sup></a> PwC's CEO survey of 4,454 leaders found 12% reporting both revenue gains and cost benefits.<a href="https://www.pwc.com/gx/en/issues/c-suite-insights/ceo-survey.html" target="_blank" rel="noopener" class="cite-link"><sup>[4]</sup></a> Meanwhile, a survey of 6,000 executives across four countries found that 90% reported no measurable impact from AI on their productivity or employment over the past three years.<a href="https://fortune.com/2026/02/17/ai-productivity-paradox-ceo-study-robert-solow-information-technology-age/" target="_blank" rel="noopener" class="cite-link"><sup>[5]</sup></a></p>

        <p>This convergence across independent methodologies is the most reliable finding in the entire literature. Not "80% fail" — rather, "5–12% succeed at scale, and the rest are in a messy middle of stalled pilots and incremental gains." AI tools adopted for email drafting and meeting summaries, but not for anything that changes how the business actually works.</p>

        <p><strong>It's not the technology.</strong> BCG surveyed 1,000 C-level executives across 59 countries and found that roughly 70% of AI implementation challenges stem from people and processes. Twenty percent from technology infrastructure. Ten percent from the algorithms themselves.<a href="https://www.bcg.com/press/24october2024-ai-adoption-in-2024-74-of-companies-struggle-to-achieve-and-scale-value" target="_blank" rel="noopener" class="cite-link"><sup>[6]</sup></a> A subsequent BCG survey of over 10,000 employees confirmed the ratio.<a href="https://www.bcg.com/publications/2025/ai-at-work-momentum-builds-but-gaps-remain" target="_blank" rel="noopener" class="cite-link"><sup>[7]</sup></a> RAND's interviews with 65 data scientists identified the top cause of failure as "misunderstanding the problem AI needs to solve" — not model limitations or data gaps, but picking the wrong problem in the first place.<a href="https://www.rand.org/pubs/research_reports/RRA2680-1.html" target="_blank" rel="noopener" class="cite-link"><sup>[8]</sup></a> McKinsey found that whether the organization had fundamentally redesigned its workflows was one of the strongest predictors of enterprise AI impact — more than what model it used or how much it spent.<a href="https://www.mckinsey.com/capabilities/quantumblack/our-insights/the-state-of-ai" target="_blank" rel="noopener" class="cite-link"><sup>[2]</sup></a></p>

        <p class="geo-quotable">The models work. The organizations don't adapt.</p>

        <p><strong>Companies acted on potential, not evidence.</strong> This is the finding that stopped me. A Harvard Business Review study of 1,006 global executives found that only 2% of AI-driven layoffs were based on demonstrated AI capability. Sixty percent were "anticipatory" — cuts made based on what AI <em>might</em> do, not what it had done.<a href="https://hbr.org/2026/01/companies-are-laying-off-workers-because-of-ais-potential-not-its-performance" target="_blank" rel="noopener" class="cite-link"><sup>[9]</sup></a> A separate survey of 600 HR professionals who had conducted AI-related layoffs found that 73% broke even or lost money on the cycle — the cost of rehiring, retraining, and recovering lost institutional knowledge consumed whatever savings the AI was supposed to deliver.<a href="https://careerminds.com/blog/cost-of-ai-layoffs" target="_blank" rel="noopener" class="cite-link"><sup>[10]</sup></a></p>

        <p>Klarna is the canonical case. The company announced its AI chatbot was doing the work of 700 customer service agents, handling 2.3 million conversations in its first month, and projected $40 million in annual savings. Within a year, customer satisfaction on complex interactions had declined and repeat contact rates increased. The CEO told Bloomberg that "investing in the quality of human support is the way of the future."<a href="https://www.entrepreneur.com/business-news/klarna-ceo-reverses-course-by-hiring-more-humans-not-ai/491396" target="_blank" rel="noopener" class="cite-link"><sup>[11]</sup></a> Months later, the company went public at a $19.6 billion valuation, citing AI-driven efficiency gains. The CEO simultaneously warned other tech leaders they were "sugarcoating" AI's impact on jobs.</p>

        <p><strong>The collaboration surprise.</strong> Here's one the AI optimists won't like. Across multiple studies of routine decision-making tasks, adding humans to AI systems actually <em>reduced</em> accuracy. In one medical diagnosis study, AI alone scored above 92%; adding a physician brought it down. In legal research, AI tools outperformed the average lawyer. In demand forecasting, human adjustments to AI predictions degraded accuracy.<a href="https://fortune.com/2025/12/09/ai-tools-outperform-human-professionals-law-advertising-ai-alone/" target="_blank" rel="noopener" class="cite-link"><sup>[12]</sup></a> The pattern isn't universal — humans still add value in creative work and domains where they're the initially stronger performer — but the default assumption that human oversight always improves AI output doesn't survive contact with the evidence. The best sequencing isn't human-plus-AI. It's AI first, human review — and the human's job is to catch what the AI misses, not to improve what the AI got right.</p>

        <h2>The Practitioner's Dilemma</h2>

        <p>There are three layers of verification in any AI initiative, and most organizations skip all of them.</p>

        <p><strong>Layer one: Are the goals right?</strong> Is this the right problem for AI to solve? RAND calls this the most common failure — bias toward the latest technology rather than solving a real problem. Organizations start with "we need an AI strategy" instead of "we have a problem that might benefit from AI."</p>

        <p><strong>Layer two: Are the requirements aligned?</strong> Do the success metrics, timelines, and resource plans actually serve the goals? This is where the process saved me with my own website. The statistics I'd drafted served my argument. They felt authoritative. I wouldn't have questioned them on my own — because verification felt like a detour from the work. The adversarial review forced the question. The formal pipeline answered it.</p>

        <p><strong>Layer three: Is the implementation proven?</strong> Does the system actually do what it claims? Not in a demo. Not in a pilot. In production, under real conditions, with real consequences.</p>

        <p>The organizations that succeed — that 5–12% — don't skip these layers. BCG found that the organizations generating the most value from AI focus on fewer initiatives — an average of 3.5 use cases versus 6.1 for those that struggle — and generate 2.1 times more ROI by going deeper on each one.<a href="https://www.bcg.com/publications/2025/closing-the-ai-impact-gap" target="_blank" rel="noopener" class="cite-link"><sup>[13]</sup></a> Sixty percent of organizations lack defined financial KPIs for their AI initiatives. The ones that define them perform measurably better. They redesign workflows instead of layering AI onto existing processes. And they are willing to stop when the evidence says stop.</p>

        <p>None of this is unique to AI. It's the same discipline that separated successful software projects from failed ones in the 1990s, successful ERP implementations from disastrous ones in the 2000s, and successful cloud migrations from expensive detours in the 2010s. Define the problem. Verify the claims. Prove it works before you commit.</p>

        <p class="geo-quotable">The tools just made it easier to forget.</p>

        <h2>The Quiet Hours</h2>

        <p>The formal assessment that surfaced all of this ran in my backyard over the course of a quiet evening. Eight research agents working in parallel, each blind to the others' findings, each searching for every credible study published in 2025 and 2026. A convergence analysis that identified which findings appeared across multiple independent sources — and which were single-source claims dressed up as consensus.</p>

        <p>Without the structured pipeline, the same work would have taken hours of manual prompting and reprompting and reprompting — the AI equivalent of asking the same question slightly differently until you get an answer that looks right. With the pipeline, the process forced citation tracing, evidence comparison, methodology assessment, and explicit uncertainty flagging. The AI couldn't take shortcuts because the pipeline didn't offer any.</p>

        <p>The irony isn't lost on me. The same technology that produced the unreliable statistics produced the better ones. The difference wasn't the model. It was the process — the structured requirement that every claim be traced to a primary source, every source be evaluated for methodology, and every conclusion be rated for confidence.</p>

        <p>The 5% that succeed aren't using better models. They're willing to stop and check — even when the story looks good enough, even when the sky is beautiful and the dog is sleeping and there's always something more pressing to build.</p>

        <p>That willingness is the whole game. It always has been.</p>

<footer class="dispatch-notes"><h2>Sources</h2>
        <ol>
          <li>Vallone, J. <a href="https://papers.ssrn.com/sol3/papers.cfm?abstract_id=5459054" target="_blank" rel="noopener">"Reassessing AI Pilot Failure Rates: A Scoping Review and Managerial Implications"</a> (SSRN, August 2025)</li>
          <li>McKinsey &amp; Company, <a href="https://www.mckinsey.com/capabilities/quantumblack/our-insights/the-state-of-ai" target="_blank" rel="noopener">"The State of AI in 2025"</a> (n=1,993 respondents, 105 countries)</li>
          <li>BCG, <a href="https://www.bcg.com/publications/2025/are-you-generating-value-from-ai-the-widening-gap" target="_blank" rel="noopener">"The Widening AI Value Gap"</a> (September 2025, n=1,250 executives)</li>
          <li>PwC, <a href="https://www.pwc.com/gx/en/issues/c-suite-insights/ceo-survey.html" target="_blank" rel="noopener">"29th Annual Global CEO Survey"</a> (January 2026, n=4,454 CEOs, 95 countries)</li>
          <li>NBER / Fortune, <a href="https://fortune.com/2026/02/17/ai-productivity-paradox-ceo-study-robert-solow-information-technology-age/" target="_blank" rel="noopener">"CEOs Admit AI Had No Impact on Productivity"</a> (February 2026, n=6,000 executives, 4 countries)</li>
          <li>BCG, <a href="https://www.bcg.com/press/24october2024-ai-adoption-in-2024-74-of-companies-struggle-to-achieve-and-scale-value" target="_blank" rel="noopener">"AI Adoption in 2024"</a> (October 2024, n=1,000 CxOs, 59 countries)</li>
          <li>BCG, <a href="https://www.bcg.com/publications/2025/ai-at-work-momentum-builds-but-gaps-remain" target="_blank" rel="noopener">"AI at Work 2025"</a> (June 2025, n=10,635 employees)</li>
          <li>RAND Corporation, <a href="https://www.rand.org/pubs/research_reports/RRA2680-1.html" target="_blank" rel="noopener">"The Root Causes of Failure for Artificial Intelligence Projects"</a> (2024, 65 practitioner interviews)</li>
          <li>Davenport &amp; Srinivasan, <a href="https://hbr.org/2026/01/companies-are-laying-off-workers-because-of-ais-potential-not-its-performance" target="_blank" rel="noopener">"Companies Are Laying Off Workers Because of AI's Potential, Not Its Performance"</a> (Harvard Business Review, January 2026, n=1,006)</li>
          <li>Careerminds, <a href="https://careerminds.com/blog/cost-of-ai-layoffs" target="_blank" rel="noopener">"The Cost of AI Layoffs"</a> (February 2026, n=600 HR professionals)</li>
          <li>Entrepreneur, <a href="https://www.entrepreneur.com/business-news/klarna-ceo-reverses-course-by-hiring-more-humans-not-ai/491396" target="_blank" rel="noopener">"Klarna CEO Reverses Course"</a> (May 2025)</li>
          <li>Fortune, <a href="https://fortune.com/2025/12/09/ai-tools-outperform-human-professionals-law-advertising-ai-alone/" target="_blank" rel="noopener">"AI Tools Outperform Human Professionals"</a> (December 2025, compilation of independent studies)</li>
          <li>BCG, <a href="https://www.bcg.com/publications/2025/closing-the-ai-impact-gap" target="_blank" rel="noopener">"From Potential to Profit: Closing the AI Impact Gap"</a> (January 2025, n=1,803 C-level executives, 19 markets)</li>
        </ol>
        <p><small>Sources include primary research where available, supplemented by secondary reporting and practitioner evidence. Sample sizes and methodology are noted where disclosed by the original source. A full evidence matrix with 50+ sources, convergence analysis, and confidence ratings is maintained as part of the assessment that produced these findings.</small></p></footer>]]></content:encoded>
      <pubDate>Mon, 06 Apr 2026 00:00:00 -0700</pubDate>
      <author>info@chapsoft.com (Russ Kohn)</author>
      <dc:creator>Russ Kohn</dc:creator>
      <guid isPermaLink="true">https://chapsoft.com/dispatches/why-ai-projects-fail</guid>
      <category>dispatch</category>
    </item>

    <item>
      <title>The Moat Is the Harness — What the Claude Code Leak Really Means</title>
      <link>https://chapsoft.com/dispatches/the-moat-is-the-harness</link>
      <description>512,000 lines of Claude Code leaked. Everyone asked what&#x27;s inside. The real question: what does it mean for building tools for AI agents?</description>
      <content:encoded><![CDATA[<p>The leak dropped on a Tuesday. I bookmarked it. The rest of the week was what it always is &mdash; client delivery across multiple projects, context-switching between frontend builds and backend gap analyses, fixing bugs in my own agent orchestration tooling, standing up a new publication in the margins. Four different codebases, five days, the usual cognitive overload of a practitioner who can't afford to stop building while the ground shifts underneath.</p>

        <p>The most interesting thing to happen in AI tooling all year finally got my attention on Saturday morning.</p>

        <p>If you're building right now &mdash; whether that's software, systems, or a team that depends on either &mdash; you know this feeling. Not the work itself, but the stack of contexts you're carrying. Client delivery in one window. Your own agent tooling in another. And somewhere underneath, the low hum of something important you haven't had time to think about yet.</p>

        <p>So when I finally sat down with 512,000 lines of Claude Code source, I wasn't asking the pundit's question. I was asking the practitioner's: what does this mean for the things I was building while everyone else was writing about it?</p>

        <hr>

        <h2>What the Code Actually Reveals</h2>

        <p>On March 31, 2026, Anthropic accidentally shipped the full source of Claude Code &mdash; their flagship agentic coding tool &mdash; inside an npm package. Within hours: 84,000 GitHub stars, coverage in Fortune, CNN, Bloomberg. We surveyed 47 public analyses. Every one asked the same question: <em>what's inside?</em></p>

        <p>Fair question. Here's what's inside: a streaming agent harness built on five architectural pillars.</p>

        <ol>
          <li><strong>Prompt-as-protocol</strong> &mdash; behavior via instructions, not runtime enforcement</li>
          <li><strong>Streaming generators</strong> &mdash; the async generator is the agent</li>
          <li><strong>Fail-closed defaults</strong> &mdash; conservative permissions with graduated autonomy</li>
          <li><strong>File-based state</strong> &mdash; git-friendly, human-readable, no database</li>
          <li><strong>Layered context</strong> &mdash; five-layer hierarchy, CSS-specificity override model</li>
        </ol>

        <p><strong>Prompt-as-protocol.</strong> Behavior is specified through system prompts, not runtime enforcement. The coordinator tells worker agents what to do via prompt instructions &mdash; there are no runtime compliance checks. This maximizes iteration speed. It also means that every behavioral guarantee is one prompt-edit away from disappearing.</p>

        <p><strong>Streaming-first generators.</strong> The async generator <em>is</em> the agent. Multiple consumers &mdash; the CLI, the SDK, the desktop app &mdash; tap the same stream. Tool execution is interleaved. Cancellation is clean. But there are seven distinct &ldquo;continue sites&rdquo; in the state machine, which is an audit and maintenance headache that will only grow.</p>

        <p><strong>Fail-closed with opt-in relaxation.</strong> New tools are restricted by default. ML classifiers learn safe patterns over time and auto-approve them. The permission model runs on a gradient: from &ldquo;always ask&rdquo; through &ldquo;ML-assisted auto-approve&rdquo; to an autonomous mode called KAIROS &mdash; more on that shortly.</p>

        <p><strong>File-based state.</strong> Everything is git-friendly, human-readable, no database. No migrations, no schema drift. The tradeoff: no indexing, no transactions, no query language.</p>

        <p><strong>Layered context assembly.</strong> A five-layer hierarchy &mdash; managed, user, project, local, auto-memory &mdash; where more specific layers override general ones. Think CSS specificity for AI context.</p>

        <p>These are real engineering choices. Defensible ones. The kind of architecture you'd build if you needed to iterate at frontier-model speed while shipping to millions of developers. And if that were the whole story, the 47 analysts would have covered it fine.</p>

        <p>But the code also reveals what they didn't intend to share.</p>

        <p>Buried in a comment in <code>src/constants/prompts.ts</code>:</p>

        <blockquote>
          <code>// @[MODEL LAUNCH]: False-claims mitigation for Capybara v8 (29-30% FC rate vs v4's 16.7%)</code>
        </blockquote>

        <p>A false-claim rate nearly double the current tier's, mitigated by a prompt patch &mdash; and initially applied only to internal employees.<sup><a href="#fn1">1</a></sup> Internal telemetry runs on two layers (Anthropic and Datadog) with no user opt-out and disk-persistent retry. Feature flags use obfuscated names &mdash; random word pairs &mdash; for concealment. An &ldquo;undercover mode&rdquo; strips AI attribution from employee contributions to open-source projects. And at least one function runs 3,167 lines long.</p>

        <p>The pundits focused on the pillars or the embarrassments. We're interested in what they mean together.</p>

        <hr>

        <h2>The Moat Isn't the Model</h2>

        <p>Here's the thesis, and it's the one the 47 analyses missed: <strong>the code isn't the product. The code is the product's governance layer. And the moat isn't the model &mdash; it's the harness.</strong></p>

        <p>Consider what Claude Code actually does. It takes a foundation model &mdash; a raw capability engine &mdash; and wraps it in a system of constraints, permissions, context management, and tool orchestration that turns capability into something you can hand to a developer and say &ldquo;this is safe to use.&rdquo; The model provides the intelligence. The harness provides the <em>judgment</em> about when and how to deploy it.</p>

        <p>Prompt-as-protocol is the clearest example. It's simultaneously the architecture's greatest strength and its most revealing weakness. Writing behavioral contracts as prompts means you can iterate at the speed of language &mdash; change a sentence, change the behavior. But it also means your safety guarantees are as durable as the model's willingness to follow instructions. There's no compiler. No type system. No enforcement layer between &ldquo;the prompt says don't do this&rdquo; and &ldquo;the model does it anyway.&rdquo;</p>

        <p>For anyone building tools that AI agents interact with &mdash; MCP servers, API endpoints, governance systems &mdash; this is the architectural reality you're designing for. The harness can request good behavior. It cannot guarantee it.</p>

        <hr>

        <h2>When the Model Becomes the Agent</h2>

        <p>The leaked code references something called KAIROS: an autonomous daemon mode with heartbeat monitoring, focus-state behavior switching between collaborative and autonomous modes, a tick engine, cost-aware sleep, and persistent execution. Over 150 references in the codebase. Alongside it, references to a model tier called Mythos &mdash; above current Opus &mdash; that appears designed for natively agentic operation.</p>

        <p>I want to be precise about what's confirmed and what's inferred here. The code references are real &mdash; they're in the leaked source. What they mean for shipping products is interpretation. But the direction is clear: Anthropic is building toward a model that doesn't need a harness to be an agent. The model <em>is</em> the agent.</p>

        <p>That's a phase transition, not an incremental improvement. Today's architecture is &ldquo;model + harness = agent.&rdquo; Tomorrow's may be &ldquo;model = agent, harness = governance.&rdquo; The harness doesn't go away &mdash; it transforms from execution environment to oversight layer.</p>

        <p>And here's where it gets uncomfortable for practitioners: <strong>better execution demands better oversight, not less.</strong></p>

        <p>The next generation of models can identify and correct their own errors recursively. Self-correction catches tactical errors &mdash; typos, logic bugs, off-by-one mistakes. The kind of thing a code review catches. But self-correction doesn't catch strategic errors &mdash; wrong requirements, missing edge cases, violated constraints, building the wrong thing beautifully. A model that can fix its own bugs but can't question its own premises is a more efficient way to build the wrong thing.</p>

        <p>If you've read <a href="https://chapsoft.com/dispatches/the-twenty-percent">The Twenty Percent</a>, you'll recognize this as the same pattern at a different scale. The 80% gets automated &mdash; McKinsey estimates over half of US work hours are already automatable as of late 2025.<sup><a href="#fn2">2</a></sup> The 20% &mdash; the judgment, the verification, the &ldquo;should we be building this at all?&rdquo; &mdash; becomes more important, not less. The organizations that treat self-correcting AI as a reason to reduce review are the ones that will ship well-tested software that solves the wrong problem.</p>

        <hr>

        <h2>Consider the Possibility</h2>

        <p>One question we can't avoid: was this leak accidental?</p>

        <p>I'm not asserting it was deliberate. But consider the information dynamics. A leak generates orders of magnitude more attention than a press release &mdash; 84,000 GitHub stars in hours,<sup><a href="#fn3">3</a></sup> 47 analyst pieces, mainstream coverage that no marketing budget could buy. The source code teaches agentic architecture to the developer community more effectively than any certification program. The Mythos references let developers anticipate the capability jump rather than be surprised by it.</p>

        <p>Whether accidental or strategic, the outcome is the same: Anthropic demonstrated that their engineering is sound enough to survive radical transparency. That's not a small thing when you're asking the world to trust your autonomous agents.</p>

        <hr>

        <h2>The Practitioner's Question</h2>

        <p>While 47 analysts cataloged features, the practitioner's question was always different: <em>what does this change about what I'm building?</em></p>

        <p>If you're building tools that AI agents interact with, the answer is: prepare for the governance layer. The harness will evolve from something that controls the model to something that the model operates within voluntarily. Your tool APIs, your permission models, your audit trails &mdash; these aren't features for today's agent. They're the foundation for tomorrow's, when the agent is autonomous and the question shifts from &ldquo;what can it do?&rdquo; to &ldquo;what should it do, and who decides?&rdquo;</p>

        <p>If you're a non-technical leader evaluating AI investments, the answer is: ask your vendors about their governance architecture, not their model benchmarks. The model will improve every quarter. The governance layer is what determines whether that improvement makes your organization more capable or more exposed.</p>

        <p>And if you're a mid-career professional wondering where you fit in all of this &mdash; the answer hasn't changed since <a href="https://chapsoft.com/dispatches/the-garden-timer">the garden timer</a>. The 80% is being automated. The 20% &mdash; the judgment, the oversight, the practical wisdom that comes from watching things fail &mdash; is where the value lives. The moat, for organizations and individuals alike, isn't the capability. It's the harness around the capability.</p>

        <p>The <a href="https://chapsoft.com/dispatches/fog">fog</a> hasn't lifted. But something large just crossed the road ahead of me &mdash; close enough to feel the draft, gone before the headlights caught it. Might have been a capybara. And in the swirl it left behind, I could see the edge of a cliff I'd been driving toward without knowing it. Not where the road goes. Just where it doesn't.</p>

        <p>The headlights are still on. The engine's still running. And now I know the terrain isn't what I assumed.</p>

        <p>Again.</p>]]></content:encoded>
      <pubDate>Sat, 04 Apr 2026 00:00:00 -0700</pubDate>
      <author>info@chapsoft.com (Russ Kohn)</author>
      <dc:creator>Russ Kohn</dc:creator>
      <guid isPermaLink="true">https://chapsoft.com/dispatches/the-moat-is-the-harness</guid>
      <category>dispatch</category>
    </item>

    <item>
      <title>Fog — Navigating AI Disruption Without a Map</title>
      <link>https://chapsoft.com/dispatches/fog</link>
      <description>From punch cards to frontier models in three short years. We&#x27;re moving faster than our senses can apprehend. A dispatch from the AI fog.</description>
      <content:encoded><![CDATA[<p>I woke up this morning lost in the fog. The rain had been welcome and refreshing, feeding a thin winter and thirsty ground — air still and cold, unwilling to hold the water. Old roads barely visible. My bright headlights that have served me so well for so long only reflect this small shallow bubble around me.</p>

        <p>I sat in the driveway for a minute, engine running, watching the beams bounce back. Forty years of building systems and I've never felt less able to see what's ahead.</p>

        <hr>

        <p>From the early days of punch cards and paper tape encoding bits traveling slowly over analogue phone lines with acoustic modems, through personal computers and improvements in network capacity and computing that would have staggered the imaginations of those early engineers, nothing compares to the quantum leap in capability that the latest generation of frontier models provide. In three short years — prompt engineering to context engineering to agentic orchestration — we're moving faster than our senses can apprehend.</p>

        <p>Even those of us engrossed in the tech, maybe especially us, are increasingly confronted with the inevitability of dislocating change and how to simultaneously embrace it, defend against it, and consider that the landscape we emerge into after this fog may look dramatically different than when the fog descended.</p>

        <p>Historically we have used celestial navigation to find our way, or trusted to cairns left by prior Finders to trace the safe route up and over the dangerous passes. But here we are in orbit around the Sun, hurtling toward an asteroid field of our own making, without a map.</p>

        <p>We seek explorers who have ventured into the fog and come back to report on what they found beyond the edge of the known — not with reassurance, but with honest reconnaissance.</p>

        <p>The Luddites will claim: repent, ignore, cling to past ways. The reckless will say: damn the torpedoes, full speed ahead. But for those of us of a more contemplative bent, all routes feel unstable, unsettled, and fraught.</p>

        <hr>

        <p>So what do you navigate by when the instruments fail? Not by certainty. Not by ideology. By the things that don't change when everything else does — the connections you've built, the judgment you've earned, the willingness to walk into the fog and report back what you find.</p>

        <p>The headlights are still on. The engine's still running. I can't see the road ahead, but I've been sitting in this driveway long enough.</p>

        <p>Hand on the wheel.</p>]]></content:encoded>
      <pubDate>Thu, 02 Apr 2026 00:00:00 -0700</pubDate>
      <author>info@chapsoft.com (Russ Kohn)</author>
      <dc:creator>Russ Kohn</dc:creator>
      <guid isPermaLink="true">https://chapsoft.com/dispatches/fog</guid>
      <category>dispatch</category>
    </item>

    <item>
      <title>The Twenty Percent — What AI Can&#x27;t Build (Yet)</title>
      <link>https://chapsoft.com/dispatches/the-twenty-percent</link>
      <description>My son built a beautiful React app with zero training. So what&#x27;s left? The 80% that looks easy vs the 20% that makes it real.</description>
      <content:encoded><![CDATA[<p>My son, with zero formal training, just built a beautifully rendered React pricing calculator. Responsive, clean typography, smooth animations. My brother, with some IT and project management background, has spun up three or four vertical market prototypes in the past month. These are things that in the past would have taken me weeks to build and wouldn't have looked as good.</p>

        <p>The tools democratized the craft. Anyone can build now.</p>

        <p>So what's left?</p>

        <p>This is the question that wakes me up at 3 AM, and I suspect I'm not alone. If you've spent decades building software — or managing projects, or designing systems, or writing code — you're watching people with a fraction of your experience produce output that looks indistinguishable from yours. Maybe better. Certainly faster.</p>

        <p>The Dunning-Kruger curve has a blind spot nobody talks about. We all know about Mount Stupid — the peak of inflated expectations where people with minimal experience think they've mastered the field. But there's a corresponding experience on the other side: the expert standing at the base of the mountain, watching Mount Stupid residents produce beautiful output, and wondering if the mountain was the illusion all along.</p>

        <p>It wasn't. The mountain is real. But the tools made the first 80% of the climb effortless, and from a distance that looks like the whole climb.</p>

        <p>Here's what the 80% looks like: a working demo. A responsive layout. A form that submits. An API that returns data. A dashboard with charts. It looks right. It runs. If you screenshot it, it's indistinguishable from a production application.</p>

        <p>Here's what the 20% looks like: nothing. It's invisible. It's the edge case that corrupts your database when two users submit the same form at the same instant. It's the authentication flow that works perfectly until someone's session expires mid-transaction. It's the data model that handles your first hundred customers but collapses at ten thousand. It's the compliance requirement that nobody mentioned because nobody knew to ask. It's the business model that assumes a conversion rate three times higher than any comparable product has ever achieved.</p>

        <p>The 20% is defined by absence. You can't see it in a demo. You can only see it when things break — and by then, you're in production, with real users, and real money, and a board asking why the thing that looked so good three months ago is now a crisis. Only 5–12% of organizations achieve significant financial impact from AI, according to convergent findings across McKinsey, BCG, and PwC.<sup><a href="#fn1">1</a></sup> The 20% is usually why.</p>

        <p>The Greeks had a word for the knowledge that lives in the 20%: phronesis. Practical wisdom. Not theoretical knowledge — not knowing that a database can have race conditions. Practical wisdom — knowing that THIS database, with THIS access pattern, under THIS load, WILL have race conditions, because you've seen it happen four times before in systems that looked exactly like this one.</p>

        <p>Phronesis can't be prompted. It can't be fine-tuned. It accumulates through the specific, unrepeatable experience of watching things fail and understanding why. Forty years of that builds a sense that no tool replicates — not because the tool is stupid, but because the knowledge is embodied. It lives in the flinch when you see a data model that's too normalized for its query patterns, in the pause when a timeline estimate feels too confident, in the question you ask that nobody else thought to ask because nobody else has watched that particular thing go wrong.</p>

        <p>Every mid-career professional is about to have the vertigo moment. The moment when someone with less experience and better tools produces something that looks like what you do — and you have to decide whether to defend your territory or redefine your value.</p>

        <p>Defending your territory doesn't work. The 80% is already gone. It's free now. Fighting that is like fighting the tide.</p>

        <p>Redefining your value means accepting that your job was never "making things." Your job was always knowing which things to make and why — and knowing which beautiful things will collapse when they meet reality.</p>

        <p>The wayfinder's value isn't "I can build things you can't." It's "I know which things will stand and which will fall, because I've watched hundreds of beautiful things collapse when they hit reality."</p>

        <p>That's the 20%. And if you've been doing this long enough to feel the vertigo, you're already standing in it.</p>]]></content:encoded>
      <pubDate>Wed, 01 Apr 2026 00:00:00 -0700</pubDate>
      <author>info@chapsoft.com (Russ Kohn)</author>
      <dc:creator>Russ Kohn</dc:creator>
      <guid isPermaLink="true">https://chapsoft.com/dispatches/the-twenty-percent</guid>
      <category>dispatch</category>
    </item>

    <item>
      <title>The Garden Timer — When Nothing Works Anymore</title>
      <link>https://chapsoft.com/dispatches/the-garden-timer</link>
      <description>A broken garden timer, a helpdesk queue, and the question every mid-career professional is asking: how will I continue to pay the bills in the age of AI?</description>
      <content:encoded><![CDATA[<p>There's a garden. I planted it. Or had it planted. Full of native grasses, fire resistant, local. In harmony I hope. I put in the drip drip drip system. Which broke day one. Poor fittings. Water flooding the zone.</p>

        <p>Like so many things these days.</p>

        <p>I bought a timer. Did some research, but not really enough. Cut corners — this one looks good enough. Amazoned it to my place. Installed it. Used it manually. Went to automate it. Bluetooth didn't work. WiFi didn't work. Pairing didn't work. Seems like just another Monday. Nothing ever just works anymore.</p>

        <p>Tried the app help section. Lots of terrible UX and crappy weak instructions. Of course I already unplugged it and replugged it. Of course I tried the 5g, the 2.4g, the this net and the that net. And the reset button. And unplugged it again.</p>

        <p>So like a peon I waited in the online queue for a poor bedraggled helpdesk clerk in god knows where dealing with god knows how many screens to navigate the terrible authentication loop the model serial number ceremony and the unbelievably bad chat interface to finally after 20 minutes get to the "unplug the gray ribbon cable and count to 20."</p>

        <p>Unplug the gray ribbon cable. Count to 20.</p>

        <p>So unplugging isn't enough. There must be a battery or capacitor or something buried in there, holding onto its state even when the power's gone. And nowhere — not in their online resources, not in their help system, not in any FAQ or troubleshooting page — do they bother to put that information where a customer could find it.</p>

        <p>And we tolerate this level of "customer service" because we will drink the water when we are thirsty.</p>

        <hr>

        <p>I keep thinking about that capacitor. The thing that holds charge after you think you've turned everything off. The hidden state that nobody documents because documenting it would mean admitting the system is more fragile than advertised.</p>

        <p>How many other places are we carrying hidden state? How many other systems — not just garden timers, but the ones that run our companies, our careers, our assumptions about what we're worth — are holding onto charge from a world that's already been unplugged?</p>

        <p>Walking all this way with tech my skill has been the distillation of need to code. Now code is increasingly ephemeral. My son, with no formal training, builds beautiful React applications. My brother spins up vertical market solutions in a weekend. The tools democratized the visible part of the craft — the part that used to take weeks and looked impressive. Anyone can make something that looks right now.</p>

        <p>So where is my value? What is my identity? How will I continue to pay the bills?</p>

        <p>These aren't theoretical questions. They're the ones I ask myself on a Tuesday afternoon when the garden timer finally connects and I'm standing in the yard realizing the easy part is automated and wondering what's left.</p>

        <hr>

        <p>Many of us are in this place this time. We see the storm clouds, we hear the hoofbeats. We don't trust our leaders — they're selling transformation while cutting headcount. We don't trust the pundits — they're monetizing our anxiety. And in this attention based economy it is very hard to know what to believe.</p>

        <p>The helpdesk clerk couldn't tell me about the capacitor because the system wasn't built to surface that knowledge. The company couldn't fix it because the incentives don't reward honesty about fragility. And here I am — same problem, different scale. Forty years of assumptions about what I'm worth, and I'm the one who needs to count to twenty.</p>

        <p>So that's what this is. This site, these dispatches — it's me unplugging the gray ribbon cable. Documenting the reboot in real time, because I don't think I'm the only one who needs one.</p>

        <p>Hello future. I'm Russ. Let's dance.</p>]]></content:encoded>
      <pubDate>Tue, 31 Mar 2026 00:00:00 -0700</pubDate>
      <author>info@chapsoft.com (Russ Kohn)</author>
      <dc:creator>Russ Kohn</dc:creator>
      <guid isPermaLink="true">https://chapsoft.com/dispatches/the-garden-timer</guid>
      <category>dispatch</category>
    </item>

  </channel>
</rss>
