<?xml version="1.0" encoding="utf-8" standalone="yes"?><rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom">
    <channel>
        <title>Governance on EXPLAIN ANALYZE</title>
        <link>https://explainanalyze.com/tags/governance/</link>
        <description>Recent content in Governance on EXPLAIN ANALYZE</description>
        <generator>Hugo -- gohugo.io</generator>
        <language>en-us</language>
        <lastBuildDate>Sun, 26 Apr 2026 00:00:00 +0000</lastBuildDate><atom:link href="https://explainanalyze.com/tags/governance/index.xml" rel="self" type="application/rss+xml" /><item>
            <title>The Hello-World Procurement Problem: Why LLM Tooling Gets Bought Wrong</title>
            <link>https://explainanalyze.com/p/the-hello-world-procurement-problem-why-llm-tooling-gets-bought-wrong/</link>
            <pubDate>Sun, 26 Apr 2026 00:00:00 +0000</pubDate>
            <guid>https://explainanalyze.com/p/the-hello-world-procurement-problem-why-llm-tooling-gets-bought-wrong/</guid>
            <description>&lt;img src=&#34;https://explainanalyze.com/&#34; alt=&#34;Featured image of post The Hello-World Procurement Problem: Why LLM Tooling Gets Bought Wrong&#34; /&gt;&lt;div class=&#34;tldr-box&#34;&gt;&#xA;    &lt;strong&gt;TL;DR&lt;/strong&gt;&#xA;    &lt;div&gt;A CTO declares &amp;ldquo;full agentic&amp;rdquo; off a demo. Without an SME watching the rollout, corruption ships and surfaces a year later when a customer reports a wrong number. With an SME, the job is information infrastructure first (so agents have enough context to make high-probability decisions) and guardrails for the cases where context isn&amp;rsquo;t enough.&lt;/div&gt;&#xA;&lt;/div&gt;&#xA;&#xA;&lt;p&gt;A CTO sits through a vendor demo. A sales engineer types &amp;ldquo;show me the top ten customers by revenue last quarter&amp;rdquo; into a prompt and a working SQL query materializes in 30 seconds, runs against a sample dataset, returns plausible numbers. The CTO declares the company is going full agentic. Procurement closes the contract by Friday.&lt;/p&gt;&#xA;&lt;p&gt;Two paths follow. Which one the company ends up on depends on whether anyone with veto authority over rollouts can measure how wrong the model will be on the company&amp;rsquo;s actual data.&lt;/p&gt;&#xA;&lt;h2 id=&#34;without-smes&#34;&gt;Without SMEs&#xA;&lt;/h2&gt;&lt;p&gt;If the agent-generated SQL looks like gold to everyone in the room, the Rounders rule applies: if you can&amp;rsquo;t spot the sucker in your first half hour at the table, you are the sucker. Without someone in the room who&amp;rsquo;d catch the polysemic &lt;code&gt;tier&lt;/code&gt; column or the undocumented soft-delete convention buried in three tables, the team is approving plausibility on a system optimized to produce it.&lt;/p&gt;&#xA;&lt;p&gt;If the produced code looks good to you, you&amp;rsquo;re probably not the SME.&lt;/p&gt;&#xA;&lt;p&gt;The corruption rate observed in the demo is a lower bound for what the tool produces against real data, &lt;a class=&#34;link&#34; href=&#34;https://explainanalyze.com/p/corruption-is-a-feature-not-a-bug-why-llms-corrupt-by-design/&#34; &gt;often by a multiple&lt;/a&gt;. The realities catalogued in &lt;a class=&#34;link&#34; href=&#34;https://explainanalyze.com/p/what-ai-gets-wrong-about-your-database/&#34; &gt;What AI Gets Wrong About Your Database&lt;/a&gt; (undocumented conventions, polysemic columns, business logic in tribal knowledge, ten-year-old codebases with three &amp;ldquo;current&amp;rdquo; patterns) are exactly the regions of input space where the model&amp;rsquo;s training distribution is sparse and contradictory. Demos run in the dense-distribution sweet spot. Production runs the inverse on every axis.&lt;/p&gt;&#xA;&lt;p&gt;With nobody positioned to measure the gap, nothing flags it. Corruption is silent by construction. It doesn&amp;rsquo;t surface as one identifiable bug; it surfaces as drift across many places at once, traced back to LLM-generated code or queries whose authors can&amp;rsquo;t reconstruct what the model meant. By the time the rate is visible, the corruption has been propagating for weeks or months. The team has too many simultaneous issues to triage one at a time. Backups have rolled past the worst of the window.&lt;/p&gt;&#xA;&lt;p&gt;The detection mode is external. A customer reports a number that doesn&amp;rsquo;t match what they expected. An analyst running LLM-powered queries on the company&amp;rsquo;s data publishes a report that contradicts internal numbers. A regulator asks a question and the answer doesn&amp;rsquo;t match the previous quarter&amp;rsquo;s filing. Whatever surfaces it, the failure is now a public one, and the team learning the failure mode is the same team trying to contain it.&lt;/p&gt;&#xA;&lt;h2 id=&#34;with-smes&#34;&gt;With SMEs&#xA;&lt;/h2&gt;&lt;p&gt;The CTO&amp;rsquo;s declaration doesn&amp;rsquo;t change. The job changes. With an SME watching the rollout, the work is infrastructure first.&lt;/p&gt;&#xA;&lt;p&gt;Agents make high-probability decisions when their inputs are dense. That means the schema is documented, polysemic columns are tagged, conventions are written down somewhere the model can reach, the dataset the agent runs against mirrors production rather than a curated subset. The realities that make a mature codebase mature (patterns evolved over years, decisions encoded in column names, exceptions buried in tribal knowledge) are exactly the inputs the agent doesn&amp;rsquo;t have unless someone puts them there. The SME&amp;rsquo;s first job is documenting what currently lives in heads. Without that, the agent operates in the sparse regions of its training distribution, and the floor on its corruption rate stays high regardless of how the harness is tuned.&lt;/p&gt;&#xA;&lt;p&gt;Guardrails are the second piece, for the cases where dense inputs still aren&amp;rsquo;t enough. Decompose work into chunks small enough to verify. Route checkpoints between chunks to the SME whose domain it is. Audits produce a failure-rate number against ground truth, not a yes/no. Recovery drills test rolling back six months of LLM-generated changes, because that&amp;rsquo;s the realistic detection horizon for silent corruption. The point is to catch the cases where the agent&amp;rsquo;s confidence and its accuracy are decoupled, which is where most of the corruption lives.&lt;/p&gt;&#xA;&lt;p&gt;Both pieces have to be in place before the deployment goes wide. Once the rate is visible from outside, the SME bench is already triaging incidents instead of building infrastructure, and the architecture won&amp;rsquo;t grow either piece on its own.&lt;/p&gt;&#xA;&lt;h2 id=&#34;when-this-doesnt-apply&#34;&gt;When this doesn&amp;rsquo;t apply&#xA;&lt;/h2&gt;&lt;ul&gt;&#xA;&lt;li&gt;&lt;strong&gt;Small teams.&lt;/strong&gt; The buyer is the SME, or one degree away. The infrastructure question gets answered by the same person making the rollout call.&lt;/li&gt;&#xA;&lt;li&gt;&lt;strong&gt;Bounded, low-stakes use cases.&lt;/strong&gt; Personal productivity tooling, draft generation, internal-only knowledge work where corruption is recoverable.&lt;/li&gt;&#xA;&lt;li&gt;&lt;strong&gt;Mature vendor categories.&lt;/strong&gt; Office suites, established CI/CD platforms, well-trodden CRM tooling. The failure modes are known and the buyer has reference points. New categories are where the asymmetry lives, and that&amp;rsquo;s exactly where LLM tooling sits in 2026.&lt;/li&gt;&#xA;&lt;/ul&gt;&#xA;&lt;h2 id=&#34;the-bigger-picture&#34;&gt;The bigger picture&#xA;&lt;/h2&gt;&lt;p&gt;If you can&amp;rsquo;t evaluate the output, it looks great. LLM corruption is silent by construction, and silence reads as correctness to anyone without the framework to see what&amp;rsquo;s wrong. The productivity dividend the CTO booked off the demo is real. The bill arrives in the quarter a customer surfaces a number that doesn&amp;rsquo;t match the books.&lt;/p&gt;&#xA;</description>
        </item></channel>
</rss>
