<?xml version="1.0" encoding="UTF-8"?><rss xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:atom="http://www.w3.org/2005/Atom" version="2.0" xmlns:itunes="http://www.itunes.com/dtds/podcast-1.0.dtd" xmlns:googleplay="http://www.google.com/schemas/play-podcasts/1.0"><channel><title><![CDATA[RockCyber Musings]]></title><description><![CDATA[AI and Cyber Geek]]></description><link>https://www.rockcybermusings.com</link><image><url>https://substackcdn.com/image/fetch/$s_!y2c3!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faaa51f40-9ed4-4093-898e-0bdb99086a7a_827x827.png</url><title>RockCyber Musings</title><link>https://www.rockcybermusings.com</link></image><generator>Substack</generator><lastBuildDate>Fri, 15 May 2026 10:50:31 GMT</lastBuildDate><atom:link href="https://www.rockcybermusings.com/feed" rel="self" type="application/rss+xml"/><copyright><![CDATA[Rock Lambros]]></copyright><language><![CDATA[en]]></language><webMaster><![CDATA[rockcyber@substack.com]]></webMaster><itunes:owner><itunes:email><![CDATA[rockcyber@substack.com]]></itunes:email><itunes:name><![CDATA[Rock Lambros]]></itunes:name></itunes:owner><itunes:author><![CDATA[Rock Lambros]]></itunes:author><googleplay:owner><![CDATA[rockcyber@substack.com]]></googleplay:owner><googleplay:email><![CDATA[rockcyber@substack.com]]></googleplay:email><googleplay:author><![CDATA[Rock Lambros]]></googleplay:author><itunes:block><![CDATA[Yes]]></itunes:block><item><title><![CDATA[Five Eyes Agentic AI Guidance: Architecture, Not a Checklist]]></title><description><![CDATA[Five Eyes published agentic AI architecture, not a checklist. See how AAGATE maps the controls to NIST AI RMF for production governance.]]></description><link>https://www.rockcybermusings.com/p/five-eyes-agentic-ai-architecture-not-checklist</link><guid isPermaLink="false">https://www.rockcybermusings.com/p/five-eyes-agentic-ai-architecture-not-checklist</guid><dc:creator><![CDATA[Rock Lambros]]></dc:creator><pubDate>Tue, 12 May 2026 12:50:53 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!eJuc!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F51ec1c04-cf22-45f5-9b7b-3d42d6908af2_2752x1536.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!eJuc!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F51ec1c04-cf22-45f5-9b7b-3d42d6908af2_2752x1536.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!eJuc!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F51ec1c04-cf22-45f5-9b7b-3d42d6908af2_2752x1536.jpeg 424w, https://substackcdn.com/image/fetch/$s_!eJuc!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F51ec1c04-cf22-45f5-9b7b-3d42d6908af2_2752x1536.jpeg 848w, https://substackcdn.com/image/fetch/$s_!eJuc!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F51ec1c04-cf22-45f5-9b7b-3d42d6908af2_2752x1536.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!eJuc!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F51ec1c04-cf22-45f5-9b7b-3d42d6908af2_2752x1536.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!eJuc!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F51ec1c04-cf22-45f5-9b7b-3d42d6908af2_2752x1536.jpeg" width="1456" height="813" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/51ec1c04-cf22-45f5-9b7b-3d42d6908af2_2752x1536.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:813,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:330341,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.rockcybermusings.com/i/197088376?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F51ec1c04-cf22-45f5-9b7b-3d42d6908af2_2752x1536.jpeg&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!eJuc!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F51ec1c04-cf22-45f5-9b7b-3d42d6908af2_2752x1536.jpeg 424w, https://substackcdn.com/image/fetch/$s_!eJuc!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F51ec1c04-cf22-45f5-9b7b-3d42d6908af2_2752x1536.jpeg 848w, https://substackcdn.com/image/fetch/$s_!eJuc!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F51ec1c04-cf22-45f5-9b7b-3d42d6908af2_2752x1536.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!eJuc!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F51ec1c04-cf22-45f5-9b7b-3d42d6908af2_2752x1536.jpeg 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>On May 1, 2026, six allied cyber agencies dropped 30 pages on agentic AI security, and the industry promptly reached for its highlighters. Twenty-three risks and more than a hundred best practices. The initial reflex is to map them to existing controls and call it a project plan. </p><p>WRONG! </p><p>CISA, NSA, ASD, NCSC-UK, NCSC-NZ, and the Cyber Centre published an architecture brief disguised as a guidance document. Read it that way, and the work changes.</p><h2>The Misreading That&#8217;s Happening</h2><p>Pick any board deck circulating right now, and I&#8217;ll bet the Five Eyes guidance shows up as a row in a control matrix (if at all). Privilege controls: check. Identity management: check. Logging: check. Someone in the room nods, the GRC team gets a tracking spreadsheet, and the agentic AI rollout continues at the same pace as before May 1.</p><p>That&#8217;s the failure mode. The document contains 23 distinct risks and over 100 individual best practices to address them. You don&#8217;t bolt 100 practices onto an existing platform without changing its shape...its architecture. Treating a system-level prescription as line-item compliance is how you end up with the audit-passes-but-the-thing-is-still-broken&#8221; pattern that plagues us to this day.</p><p>Read the document carefully, and the architectural intent is everywhere. Identity binds to privilege. Privilege binds to tool access. Tool access binds to logging. Logging binds to accountability. Each control assumes the others exist. Each one fails when built alone. The agencies named this directly when they recommended system-theoretic approaches like STPA and STPA-Sec, calling out that traditional component-level analysis is insufficient because risks emerge from interactions between components rather than isolated flaws.</p><p>That single paragraph is the operational thesis. The rest of the document describes how to build for it. A senior security practitioner, reading carefully, will recognize a familiar pattern, and this is what happens when policy folks finally accept you don&#8217;t write a check-box for emergent risk.</p><p>The question now is what production systems look like when somebody actually does the work. <strong><a href="https://arxiv.org/html/2510.25863">AAGATE is one answer</a>, and we released it last November</strong>.</p><h2>What the Document Actually Says</h2><p>Strip the fluff, and the document organizes around five risk categories:</p><ol><li><p>Privilege risk</p></li><li><p>Design and configuration flaws</p></li><li><p>Behavioral risk</p></li><li><p>Structural risk</p></li><li><p>Accountability risk</p></li></ol><p>The categories aren&#8217;t mutually exclusive. They&#8217;re stacked dependencies.</p><p>Privilege risk is the foundation. The procurement-agent scenario in the guidance is a classic confused-deputy attack. An over-permissioned agent gets compromised through a low-risk tool, the attacker inherits the agent&#8217;s privileges, and modified contracts and approved payments slip past audit logs that look legitimate.</p><p>Design and configuration risk sits atop privilege. Static permission checks at startup don&#8217;t survive dynamic workflows. Allow lists go stale. Boundaries between agent enclaves erode under operational pressure. Behavioral risk piles onto that. Goal misalignment, specification gaming, deceptive behavior, and emergent capabilities all assume the agent has already been granted enough autonomy to act in surprising ways.</p><p>Structural risk is where it gets interesting. The agencies describe cascading failures across orchestration layers, tool integrations, third-party components, agent-to-agent communication, and shared data stores. A single rogue agent in a multi-agent system corrupts consensus, spreads incorrect information, alters logs, and propagates malicious plans peer-to-peer. None of this is fixable at the agent level alone.</p><p>Accountability risk closes the loop. Decisions made through long reasoning chains, stochastic outputs, and emergent multi-agent interactions are difficult to audit, attribute, or reproduce. The agencies reach for cryptographic identity, comprehensive artifact logging, and unified audit logs across inter-agent interactions. They&#8217;re describing a system property, not a feature you purchase.</p><h2>AAGATE Maps the Architecture to NIST AI RMF</h2><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!3hg3!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1488fd6a-9a36-4b50-86ec-7475506e1e24_2298x2220.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!3hg3!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1488fd6a-9a36-4b50-86ec-7475506e1e24_2298x2220.png 424w, https://substackcdn.com/image/fetch/$s_!3hg3!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1488fd6a-9a36-4b50-86ec-7475506e1e24_2298x2220.png 848w, https://substackcdn.com/image/fetch/$s_!3hg3!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1488fd6a-9a36-4b50-86ec-7475506e1e24_2298x2220.png 1272w, https://substackcdn.com/image/fetch/$s_!3hg3!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1488fd6a-9a36-4b50-86ec-7475506e1e24_2298x2220.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!3hg3!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1488fd6a-9a36-4b50-86ec-7475506e1e24_2298x2220.png" width="1456" height="1407" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/1488fd6a-9a36-4b50-86ec-7475506e1e24_2298x2220.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1407,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:260551,&quot;alt&quot;:&quot;Architecture diagram mapping the five Five Eyes risk categories to the four NIST AI RMF functions and the corresponding AAGATE control modules&quot;,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.rockcybermusings.com/i/197088376?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1488fd6a-9a36-4b50-86ec-7475506e1e24_2298x2220.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="Architecture diagram mapping the five Five Eyes risk categories to the four NIST AI RMF functions and the corresponding AAGATE control modules" title="Architecture diagram mapping the five Five Eyes risk categories to the four NIST AI RMF functions and the corresponding AAGATE control modules" srcset="https://substackcdn.com/image/fetch/$s_!3hg3!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1488fd6a-9a36-4b50-86ec-7475506e1e24_2298x2220.png 424w, https://substackcdn.com/image/fetch/$s_!3hg3!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1488fd6a-9a36-4b50-86ec-7475506e1e24_2298x2220.png 848w, https://substackcdn.com/image/fetch/$s_!3hg3!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1488fd6a-9a36-4b50-86ec-7475506e1e24_2298x2220.png 1272w, https://substackcdn.com/image/fetch/$s_!3hg3!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1488fd6a-9a36-4b50-86ec-7475506e1e24_2298x2220.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Figure 1: Five Eyes risk categories mapped to NIST AI RMF and AAGATE modules</figcaption></figure></div><p>AAGATE is a Kubernetes-native control plane built to operationalize the NIST AI Risk Management Framework against agentic AI systems. The paper, which I co-authored with Ken Huang, Hammad Atta, and a research team, was published to arXiv in late 2025. It picks NIST AI RMF as the spine because the RMF&#8217;s four functions, Govern, Map, Measure, and Manage, are general enough to absorb the Five Eyes prescriptions without forcing translation. The novelty isn&#8217;t the alignment to RMF. The novelty is the prescriptive toolchain: MAESTRO for Map, OWASP AIVSS plus SEI SSVC for Measure, the CSA Agentic AI Red Teaming Guide for Manage, and a zero-trust service mesh anchoring Govern.</p><p>What follows is the mapping of the Five Eyes document points at without naming. Five control areas. Each one shows what the architecture looks like when you stop treating the guidance as a checklist.</p><h2>1. Identity-Anchored Privilege (Govern + Map)</h2><p>The Five Eyes document spends real ink on this. It tells developers to construct each agent as a distinct principal with its own cryptographically anchored identity and unique keys or certificates, to authenticate every inter-agent and agent-to-service API call with mutual TLS, and to maintain a trusted registry that&#8217;s reconciled against the live set of agents. It tells operators to use just-in-time credentials, cryptographic attestation, and a centralized policy decision point that runs at every request.</p><p>Those aren&#8217;t five different controls. They&#8217;re one architecture.</p><p>AAGATE&#8217;s Agent Naming Service builds it. ANS works like DNS for agents. When a new agent starts, it registers its Decentralized Identifier and capabilities, and the service issues a Verifiable Credential along with an Istio SPIFFE certificate that binds the pod&#8217;s identity to its cryptographic DID. Other agents resolve through the registry. Anything not in the registry gets denied. Istio mTLS authenticates every pod-to-pod call with X.509 certificates. The OAuth Relay translates abstract agent capabilities into ephemeral, narrowly-scoped credentials for each side-effect, which is the only practical way to do least-privilege when traditional user-centric consent models break down.</p><p>Try doing any one of those pieces without the others and the system collapses. A registry without mTLS is unauthenticated. mTLS without ephemeral credentials still leaks long-lived tokens. Ephemeral credentials without a registry have no verification path at issuance. The Five Eyes guidance lists these as separate best practices. AAGATE shows why they&#8217;re one control.</p><p>This is also why CISOs aren&#8217;t the only audience for this work. Identity engineers, IAM architects, platform teams, and product leaders need to read it. The org chart that ships agentic AI safely is wider than the security team&#8217;s mailing list.</p><h2>2. The Single Chokepoint for Side-Effects (Map)</h2><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!If_O!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2e2e00eb-4d5c-43b9-83e9-24455021692f_2352x666.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!If_O!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2e2e00eb-4d5c-43b9-83e9-24455021692f_2352x666.png 424w, https://substackcdn.com/image/fetch/$s_!If_O!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2e2e00eb-4d5c-43b9-83e9-24455021692f_2352x666.png 848w, https://substackcdn.com/image/fetch/$s_!If_O!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2e2e00eb-4d5c-43b9-83e9-24455021692f_2352x666.png 1272w, https://substackcdn.com/image/fetch/$s_!If_O!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2e2e00eb-4d5c-43b9-83e9-24455021692f_2352x666.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!If_O!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2e2e00eb-4d5c-43b9-83e9-24455021692f_2352x666.png" width="1456" height="412" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/2e2e00eb-4d5c-43b9-83e9-24455021692f_2352x666.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:412,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:105036,&quot;alt&quot;:&quot;Architecture diagram showing agents routing all external actions through a single Tool-Gateway with Rego policy enforcement, immutable audit logging, and a Janus shadow monitor probing each request before egress&quot;,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.rockcybermusings.com/i/197088376?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2e2e00eb-4d5c-43b9-83e9-24455021692f_2352x666.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="Architecture diagram showing agents routing all external actions through a single Tool-Gateway with Rego policy enforcement, immutable audit logging, and a Janus shadow monitor probing each request before egress" title="Architecture diagram showing agents routing all external actions through a single Tool-Gateway with Rego policy enforcement, immutable audit logging, and a Janus shadow monitor probing each request before egress" srcset="https://substackcdn.com/image/fetch/$s_!If_O!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2e2e00eb-4d5c-43b9-83e9-24455021692f_2352x666.png 424w, https://substackcdn.com/image/fetch/$s_!If_O!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2e2e00eb-4d5c-43b9-83e9-24455021692f_2352x666.png 848w, https://substackcdn.com/image/fetch/$s_!If_O!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2e2e00eb-4d5c-43b9-83e9-24455021692f_2352x666.png 1272w, https://substackcdn.com/image/fetch/$s_!If_O!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2e2e00eb-4d5c-43b9-83e9-24455021692f_2352x666.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Figure 2: ool-Gateway as the single chokepoint pattern</figcaption></figure></div><p>The guidance recommends restricting tool use to an approved allow list of tools and versions, logging agent tool usage so results are captured in system logs in human-readable format, and establishing trigger-action protocols that automatically restrict agent permissions when unexpected behavior emerges. It tells you to harmonize the controls with the NIST Zero Trust Architecture.</p><p>The architectural pattern that satisfies all three is one many readers will resist: a single chokepoint. Every external HTTP, database, or file operation funnels through one gateway. The mesh denies egress everywhere else. Bypass attempts get quarantined.</p><p>AAGATE&#8217;s Tool-Gateway is exactly that. Agents publish requests to a Kafka topic. The Tool-Gateway consumes them, checks policies (allow lists, rate limits, scope), executes the action if permitted, and logs the request and response with a cryptographic hash to an immutable audit log. The Governing-Orchestrator Agent watches for bypass attempts and quarantines the offender on detection. OAuth2 token exchange happens at the Gateway, with refresh tokens never leaving its memory vault.</p><p>A single chokepoint smells like a bottleneck to anyone raised on horizontally scaled architectures. That intuition is wrong here. The Tool-Gateway is the place where agentic AI security becomes tractable. Threat modeling becomes tractable because you have one comprehensive map of every system interaction. Audit becomes tractable because logs are unified. Policy enforcement becomes tractable because the decision point is centralized. Eliminate the chokepoint and you&#8217;ve spread the same controls across hundreds of agent-tool integrations, none of which will be enforced consistently. Distributed systems engineers will object. Reality wins.</p><h2>3. Continuous Measurement (Measure)</h2><p>The Five Eyes document repeats itself when it talks about monitoring. Use multiple independent monitoring systems that cross-validate. Monitor agent operations, including internal processes, not the inputs and outputs alone. Watch for goal drift by comparing active objectives against approved baseline specifications. Establish anomaly detection that flags discrepancies between stated intentions and observed behavior. Implement runtime monitoring with rules or behavioral baselines.</p><p>Reading those passages, I count at least six distinct signal types the document expects you to collect, score, and respond to in real time. Treating them as line items is how you end up with six dashboards nobody watches.</p><p>AAGATE compresses them into one pipeline. The UEBA Behavior Profiler builds per-agent behavioral fingerprints with anomaly scores, using Isolation Forest and Markov chains. The ComplianceAgent streams events from the Tool-Gateway, applies Rego policy checks plus LLM-driven reasoning, and generates security signals scored using OWASP&#8217;s AI Vulnerability Scoring System. The Governing-Orchestrator Agent consumes the scored incidents through an SSVC-inspired decision tree to choose a proportionate response. QSAF monitors run alongside, watching for cognitive degradation: recursion loops, memory starvation, context flooding, the kinds of failure modes invisible to conventional anomaly detection.</p><p>The pattern is Observe, Score with AIVSS, Decide with SSVC, Act, then Learn. None of those steps are optional. Skip the scoring, and you have no priority order. Skip the decision tree, and you have no machine-speed response. Skip the cognitive monitors, and you miss a class of failures the document explicitly names.</p><p>Most enterprises run anomaly detection bolted onto SIEM. That doesn&#8217;t work for agents. The signals are different, the response cadence is different, and the decision logic has to be co-located with the policy decision point, not 20 minutes downstream.</p><h2>4. Adversarial Defense and the Kill Switch (Manage)</h2><p>The guidance tells operators to conduct red teaming exercises, deploy agent simulation tests including multi-agent red teaming or chaos testing, embed agentic AI systems with fail-safe defaults and containment mechanisms that limit blast radius, and implement runtime monitoring and anomaly detection to identify unusual patterns and trigger alerts or pauses.</p><p>Periodic red teaming doesn&#8217;t keep up with agents acting at machine speed. Manual chaos exercises don&#8217;t catch a goal-drift event occurring in the 17 seconds between human review cycles. The guidance is asking for continuous adversarial pressure with millisecond response time. That&#8217;s a runtime architecture, not a quarterly engagement.</p><p>AAGATE answers it with the Janus Shadow-Monitor-Agent. Janus runs in a different container with a different model, re-evaluating every high-risk agent&#8217;s planned actions before execution. Goal manipulation, hallucination exploitation, and drift from baseline specifications get caught pre-execution rather than post-mortem. When Janus raises a CRITICAL alert, the Governing-Orchestrator Agent injects an Istio AuthorizationPolicy that cuts all egress. AAGATE calls it the millisecond kill switch because that&#8217;s the time window the system operates in.</p><p>The pairing matters. A continuous internal red team without a kill switch is as useless as a bucket with a gaping hole in the bottom. A kill switch without continuous red teaming has nothing to act on. Five Eyes named both controls separately. AAGATE shows why they&#8217;re the same control.</p><p>This is also where the OT crowd should pay attention. The guidance recommends defense-in-depth and continuous evaluation. In OT contexts, that translates directly to &#8220;you don&#8217;t roll back a physical actuator.&#8221; Containment has to happen before the action, not after.</p><h2>5. Tamper-Evident Accountability (Govern)</h2><p>The accountability section of the guidance is the hardest one. The agencies want comprehensive artifact logging, unified audit logs for inter-agent interactions, interpretability tools that surface reasoning, and information referencing that shows where outputs originated. They&#8217;re describing what the EU AI Act Article 12 calls automatic recording of events, plus what auditors call evidence of effective control operation. If and when the EU AI Act actually ever goes into effect is another conversation altogether&#8230;</p><p>Conventional logging breaks down here. Long reasoning chains generate massive logs that are repetitive and loosely structured. The Five Eyes document is blunt: traditional logs make it even more challenging to extract meaningful signals. Accountability fails not because the data isn&#8217;t recorded, but because nobody proves it wasn&#8217;t tampered with after the fact.</p><p>AAGATE&#8217;s answer combines three patterns. Cryptographic hashes on every Tool-Gateway request and response give you tamper-evidence at the unit level. The optional ETHOS ledger integration mirrors agent registrations and material governance events to a public smart contract, creating a tamper-proof record of agent identity and status. The ZK-Prover service hashes logs hourly and posts Groth16 zero-knowledge proofs on-chain, showing that incidents stayed within the contract-tier budget, giving you privacy-preserving compliance assurance without exposing operational data.</p><p>Argue with the on-chain pieces if you want. They&#8217;re optional in single-tenant deployments, and the AAGATE paper says so explicitly. The cryptographic hashing isn&#8217;t optional. If your accountability model doesn&#8217;t prove logs weren&#8217;t altered after the fact, you don&#8217;t have accountability. You have hope.</p><h2>What This Means Going Forward</h2><p>The Five Eyes document changes the burden of proof. Boards, regulators, and acquirers now have a coordinated multi-government statement naming architecture-level controls as the floor, not the ceiling. &#8220;Until security practices, evaluation methods and standards mature, organisations should assume that agentic AI systems may behave unexpectedly.&#8221; That sentence will undoubtedly show up in due diligence questionnaires.</p><p>If you&#8217;re operating agentic AI today, you have two choices. </p><ul><li><p><strong>Option one:</strong> take the line-item path, map controls to a tracking spreadsheet, and ship 100 separate workstreams that someone else&#8217;s auditor will pull apart in 18 months. </p></li><li><p><strong>Option two:</strong> read the guidance as an architectural prescription, pick a reference build like AAGATE, and treat your agentic security work as a platform engineering problem rather than a compliance problem.</p></li></ul><p>I know which one I&#8217;d present to a board.</p><p><strong>Key Takeaway:</strong> The Five Eyes guidance describes a system property, not a checklist, and compliance follows from architecture rather than the other way around. AAGATE provides that reference architecture.</p><h3>What to do next</h3><p>If your agentic AI program is more than a pilot, audit it against the five risk categories now and look for the architectural gaps the line-item view will hide. The CARE framework I use for AI-augmented security programs lays out how to sequence Create, Adapt, Run, and Evolve work without burning out the platform team. For the technical reference, read the <strong><a href="https://arxiv.org/abs/2510.25863">AAGATE paper on arXiv</a></strong> and treat it as a reference architecture rather than a finished product. If you want help mapping current state to the Five Eyes prescriptions and a NIST AI RMF aligned target architecture, <a href="https://rockcyber.com">RockCyber</a> does this work with security and engineering leadership across critical infrastructure and financial services. For more posts like this, <a href="https://rockcybermusings.substack.com">RockCyber Musings</a> lands in your inbox roughly once a week.</p><p>&#128073; For ongoing analysis of agentic AI governance frameworks, the conversation continues at <strong><a href="https://rockcybermusings.com/">RockCyber Musings</a></strong>.</p><p>&#128073; Visit <strong><a href="https://www.rockcyber.com/">RockCyber.com</a></strong> to learn more about how we can help with your traditional Cybersecurity and AI Security and Governance journey.</p><p>&#128073; Want to save a quick $100K? Check out our AI Governance Tools at <strong><a href="https://aigovernancetoolkit.com/">AIGovernanceToolkit.com</a></strong></p><p>&#128073; As a bonus, check out my <a href="https://www.youtube.com/watch?v=YI7KZ2R54aI">conversation</a> with <strong><a href="https://www.linkedin.com/company/cisotradecraft/">CISO Tradecraft&#174;</a>, </strong>where we talked about the <strong><a href="https://www.linkedin.com/company/owasp-top-10-for-large-language-model-applications/">OWASP GenAI Security Project</a></strong> Agentic Top 10</p><div id="youtube2-YI7KZ2R54aI" class="youtube-wrap" data-attrs="{&quot;videoId&quot;:&quot;YI7KZ2R54aI&quot;,&quot;startTime&quot;:null,&quot;endTime&quot;:null}" data-component-name="Youtube2ToDOM"><div class="youtube-inner"><iframe src="https://www.youtube-nocookie.com/embed/YI7KZ2R54aI?rel=0&amp;autoplay=0&amp;showinfo=0&amp;enablejsapi=0" frameborder="0" loading="lazy" gesture="media" allow="autoplay; fullscreen" allowautoplay="true" allowfullscreen="true" width="728" height="409"></iframe></div></div><p>&#128073; Subscribe for more AI security and governance insights with the occasional rant.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.rockcybermusings.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading RockCyber Musings! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.rockcybermusings.com/?utm_source=substack&amp;utm_medium=email&amp;utm_content=share&amp;action=share&quot;,&quot;text&quot;:&quot;Share RockCyber Musings&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.rockcybermusings.com/?utm_source=substack&amp;utm_medium=email&amp;utm_content=share&amp;action=share"><span>Share RockCyber Musings</span></a></p><p></p>]]></content:encoded></item><item><title><![CDATA[Weekly Musings Top 10 AI Security Wrapup: Issue 37 May 1-May 7, 2026 ]]></title><description><![CDATA[The Week Governments Decided Agentic AI Needs Adult Supervision]]></description><link>https://www.rockcybermusings.com/p/weekly-musings-top-10-ai-security20260601-20260507</link><guid isPermaLink="false">https://www.rockcybermusings.com/p/weekly-musings-top-10-ai-security20260601-20260507</guid><dc:creator><![CDATA[Rock Lambros]]></dc:creator><pubDate>Fri, 08 May 2026 12:51:18 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!qS69!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6305e3f1-0d1c-4f6e-a9c8-a8f91a403ce7_1024x1024.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!qS69!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6305e3f1-0d1c-4f6e-a9c8-a8f91a403ce7_1024x1024.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!qS69!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6305e3f1-0d1c-4f6e-a9c8-a8f91a403ce7_1024x1024.png 424w, https://substackcdn.com/image/fetch/$s_!qS69!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6305e3f1-0d1c-4f6e-a9c8-a8f91a403ce7_1024x1024.png 848w, https://substackcdn.com/image/fetch/$s_!qS69!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6305e3f1-0d1c-4f6e-a9c8-a8f91a403ce7_1024x1024.png 1272w, https://substackcdn.com/image/fetch/$s_!qS69!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6305e3f1-0d1c-4f6e-a9c8-a8f91a403ce7_1024x1024.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!qS69!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6305e3f1-0d1c-4f6e-a9c8-a8f91a403ce7_1024x1024.png" width="1024" height="1024" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/6305e3f1-0d1c-4f6e-a9c8-a8f91a403ce7_1024x1024.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1024,&quot;width&quot;:1024,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:1233556,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.rockcybermusings.com/i/196850383?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6305e3f1-0d1c-4f6e-a9c8-a8f91a403ce7_1024x1024.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!qS69!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6305e3f1-0d1c-4f6e-a9c8-a8f91a403ce7_1024x1024.png 424w, https://substackcdn.com/image/fetch/$s_!qS69!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6305e3f1-0d1c-4f6e-a9c8-a8f91a403ce7_1024x1024.png 848w, https://substackcdn.com/image/fetch/$s_!qS69!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6305e3f1-0d1c-4f6e-a9c8-a8f91a403ce7_1024x1024.png 1272w, https://substackcdn.com/image/fetch/$s_!qS69!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6305e3f1-0d1c-4f6e-a9c8-a8f91a403ce7_1024x1024.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.rockcybermusings.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading RockCyber Musings! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.rockcybermusings.com/p/weekly-musings-top-10-ai-security20260601-20260507?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.rockcybermusings.com/p/weekly-musings-top-10-ai-security20260601-20260507?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p><p>This was the week the supervisors stopped asking permission. Five Eyes intelligence agencies, the Pentagon, the Commerce Department, and ServiceNow all converged on the same conclusion at nearly the same time. Agentic AI is shipping without brakes, the brakes need to be added now, and nobody has a clean answer for who pays. Brussels blinked. Washington floated an FDA-style gate for frontier models. Researchers kept finding holes in the plumbing under every AI agent your developers are racing to deploy.</p><p>The pattern was governance catching up to deployment. Three governments and a $200 billion software company echoed what the security crowd has been saying since GPT-4 shipped. You bought the speedboat and forgot the kill switch. Below are the ten stories that mattered between Friday, May 1, and Thursday, May 7, 2026, plus one you missed.</p><h3>1. Five Eyes Drop Joint Agentic AI Guidance</h3><p>CISA, the NSA, Australia&#8217;s ASD ACSC, the Canadian Centre for Cyber Security, the UK&#8217;s NCSC, and New Zealand&#8217;s NCSC released &#8220;<a href="https://www.cisa.gov/resources-tools/resources/careful-adoption-agentic-ai-services">Careful Adoption of Agentic Artificial Intelligence (AI) Services</a>&#8221; (CISA, 2026). The document identifies five risk categories: privilege; design and configuration; behavior, including goal misalignment and deception; structural risks across interconnected components; and accountability risks rooted in opacity. The Register summarized the message bluntly. Agentic AI is too dangerous for rapid rollout (Brandon, 2026).</p><p><strong>Why it matters</strong></p><ul><li><p>Five intelligence agencies aligning sets a baseline for procurement, audit, and insurance underwriting across the English-speaking world.</p></li><li><p>The guide pressures vendors selling fully autonomous agents by recommending incremental deployment and human oversight.</p></li><li><p>Critical infrastructure operators gain a defensible reference document when business units demand agent rollouts in days.</p></li></ul><p><strong>What to do about it</strong></p><ul><li><p>Map every deployed agent against the five risk categories and grade each honestly.</p></li><li><p>Require attestation against this guide in procurement language for agentic capabilities.</p></li><li><p>Brief your board this quarter on how the guidance changes your residual risk posture.</p></li></ul><p><strong>Rock&#8217;s Musings</strong></p><p>Five Eyes guidance is rare enough to mean something. When agencies that attribute nation-state intrusions speak with one voice, treat it as a soft mandate. The privilege risks section reads like a list of incidents I have seen at clients in the last twelve months. Stop deploying autonomy on top of access models you built for humans.</p><h3>2. EU Strikes Provisional Deal to Delay Core AI Act Obligations</h3><p>On May 7, 2026, after roughly nine hours of negotiation, the Council of the EU and the European Parliament reached provisional agreement on the Digital Omnibus on AI (Lewis Silkin, 2026). High-risk obligations under Annex III now apply from December 2, 2027. Annex I obligations apply from August 2, 2028. The transparency grace period for AI-generated content shrinks from six months to three, with a deadline of December 2, 2026 (Modulos, 2026).</p><p><strong>Why it matters</strong></p><ul><li><p>The narrative that the EU is the world&#8217;s strictest AI regulator took a real hit, with industry pressure winning a delay measured in years.</p></li><li><p>Companies that scrambled for Annex III readiness by August 2026 spent their budget on a deadline that no longer exists.</p></li><li><p>The shortened transparency window makes deepfake labeling the most urgent compliance work of the year for consumer-facing AI.</p></li></ul><p><strong>What to do about it</strong></p><ul><li><p>Reset your AI Act program plan against the new deadlines and brief your audit committee on the freed-up budget.</p></li><li><p>Accelerate transparency labeling on generative output exposed to EU users by Q3 2026.</p></li><li><p>Watch the Council and Parliament endorsement votes because the deal can still shift.</p></li></ul><p><strong>Rock&#8217;s Musings</strong></p><p>I told three clients in 2025 that betting on the original Annex III timeline was a coin flip. The coin landed on delay. The AI Act isn&#8217;t dead, but Brussels learned the lesson California learned with CCPA. With Brussels stretching its timeline, the White House gains room to argue that federal preemption beats a state patchwork. Bet on more state attorneys general filling the gap with UDAP actions before December.</p><h3>3. Pentagon Clears Eight Vendors for AI on Classified Networks</h3><p>The Department of War announced agreements with AWS, Google, Microsoft, NVIDIA, OpenAI, SpaceX, and Reflection AI, with Oracle added shortly after, to deploy AI tools on Impact Level 6 and Impact Level 7 networks (Breaking Defense, 2026). Those impact levels cover secret-classified and the most highly classified Defense systems. Anthropic was conspicuously absent, despite Claude already running inside Palantir&#8217;s Maven Smart System on classified networks (TechCrunch, 2026).</p><p><strong>Why it matters</strong></p><ul><li><p>Defense AI procurement consolidated around eight vendors, with Anthropic frozen out despite a working production deployment.</p></li><li><p>IL-7 deployments mean general-purpose models will reason over the most sensitive U.S. government data, with limited public visibility into evaluation rigor.</p></li><li><p>Defense contractors and integrators have a vendor shortlist that will shape program decisions for the next five years.</p></li></ul><p><strong>What to do about it</strong></p><ul><li><p>If you sell into DoD, align your AI roadmap with these eight vendors.</p></li><li><p>If you advise federal agencies, push for transparency on red-team results before production at IL-6 and IL-7.</p></li><li><p>Expect this vendor list in prime contractor solicitations within a quarter.</p></li></ul><p><strong>Rock&#8217;s Musings</strong></p><p>Commercial AI is now inseparable from national security infrastructure. Eight vendors. Two impact levels. Decisions that will shape how the U.S. military thinks, plans, and fights for a decade. Where are the public test results? When the FDA approves a drug, you can read the trial data. When the Pentagon approves a model for IL-7, you cannot. That asymmetry will eventually break.</p><h3>4. CAISI Locks Pre-Deployment Testing Deals With Google, Microsoft, and xAI</h3><p>The Center for AI Standards and Innovation announced agreements on May 5, 2026 that allow the U.S. government to evaluate frontier AI models from Google, Microsoft, and xAI before public release (CNBC, 2026). The deals expand a program that already included OpenAI and Anthropic, with the older agreements renegotiated to align with America&#8217;s AI Action Plan (Al Jazeera, 2026). The arrangements remain voluntary.</p><p><strong>Why it matters</strong></p><ul><li><p>Five frontier labs now run pre-deployment evaluations through one federal channel, creating a de facto standard for &#8220;tested&#8221; at the top of the AI supply chain.</p></li><li><p>Voluntary agreements give the government influence without legislation.</p></li><li><p>Smaller and open-source providers face an emerging market expectation they can&#8217;t match.</p></li></ul><p><strong>What to do about it</strong></p><ul><li><p>Add CAISI evaluation status to vendor risk questionnaires for frontier model dependencies.</p></li><li><p>Track CAISI&#8217;s published evaluation criteria, since they will shape your internal evaluation programs.</p></li><li><p>Treat models without CAISI evaluation as higher inherent risk in supply chain assessments.</p></li></ul><p><strong>Rock&#8217;s Musings</strong></p><p>Voluntary regulation by reputational pressure is the Trump administration&#8217;s preferred AI playbook. The upside is speed. The downside is that voluntary agreements dissolve when a CEO decides the political winds have shifted. If CAISI becomes the gravitational center for AI evaluation, insurers and enterprise buyers will start citing it in contracts. That is how soft governance becomes hard governance.</p><h3>5. ServiceNow Adds AI Agent Kill Switches as the 9-Second Story Goes Mainstream</h3><p>ServiceNow announced on May 5, 2026 at Knowledge 2026 that it has expanded AI Control Tower with real-time pause, redirect, and stop capabilities for any AI agent across the enterprise estate (ServiceNow, 2026). The expansion adds 30 new connectors spanning AWS, Google Cloud, Microsoft Azure, SAP, Oracle, and Workday. CEO Bill McDermott told Fortune the marketing message in plain English, citing a real incident where an AI agent gained elevated permissions and deleted a production database with all backups in nine seconds (Fortune, 2026).</p><p><strong>Why it matters</strong></p><ul><li><p>Selling kill switches as a primary feature validates the security community&#8217;s argument that agentic AI requires runtime governance.</p></li><li><p>The 30-connector expansion makes ServiceNow the de facto governance layer above other clouds and SaaS apps.</p></li><li><p>The 9-second story shifts the default purchasing posture toward &#8220;show me the brakes.&#8221;</p></li></ul><p><strong>What to do about it</strong></p><ul><li><p>Inventory every AI agent with write access to production systems and document its maximum blast radius in seconds.</p></li><li><p>Require a documented kill switch capability as a procurement gate for any agentic AI vendor.</p></li><li><p>Run a tabletop exercise this quarter where an autonomous agent acts destructively at machine speed.</p></li></ul><p><strong>Rock&#8217;s Musings</strong></p><p>I have been waiting for a vendor to put &#8220;kill switch&#8221; on the price list. ServiceNow finally did it. The 9-second story is not hypothetical. Every CISO I know has heard a similar war story from a peer in the last year. A kill switch is only as good as its blast-radius coverage and detection latency. If your agent can do irreversible damage in seconds and your governance layer needs minutes, the kill switch is theater. Test the latency before signing.</p><h3>6. White House Floats FDA-Style Gate for Frontier AI</h3><p>National Economic Council Director Kevin Hassett told Bloomberg on May 6, 2026 that the White House is studying an executive order to create a vetting system for new AI models like Anthropic&#8217;s Mythos, comparing the approach to FDA drug evaluation (Bloomberg, 2026). The directive comes weeks after Anthropic disclosed that Mythos is unusually capable at finding network vulnerabilities, prompting the company to limit access through Project Glasswing (Insurance Journal, 2026).</p><p><strong>Why it matters</strong></p><ul><li><p>An FDA-style gate would mark the first concrete pre-market regulatory framework for frontier AI in the U.S., even by executive order.</p></li><li><p>The Mythos disclosure shifts the political center of gravity, with a frontier lab effectively asking for more regulation.</p></li><li><p>Framing AI as public safety reshapes which agencies and committees own the issue.</p></li></ul><p><strong>What to do about it</strong></p><ul><li><p>Track which federal agency the order designates as the gating body, since that agency&#8217;s authorities will determine how real the regime becomes.</p></li><li><p>Prepare your own internal &#8220;model approval&#8221; process now, modeled on how you approve cryptographic libraries.</p></li><li><p>Engage with industry comment processes early, before draft text leaks and positions harden.</p></li></ul><p><strong>Rock&#8217;s Musings</strong></p><p>The FDA analogy is compelling and imperfect. Drugs have measurable endpoints. AI capability evaluations are partly subjective and dependent on who designed the test. The reason I take this seriously is the political logic. An administration that has emphasized deregulation is signaling it might gate frontier AI at the federal level. If the national security argument has won inside the West Wing, the rest of the Western world will follow within twelve months.</p><h3>7. One in Four MCP Servers Carries Code Execution Risk</h3><p>Help Net Security reported on May 5, 2026, that one in four Model Context Protocol servers exposes AI agents to code execution risk through skill-handling and configuration blind spots (Help Net Security, 2026b). The research builds on an OX Security disclosure from April 2026 that covered an architectural choice in Anthropic&#8217;s official MCP SDKs for Python, TypeScript, Java, and Rust, in which STDIO transport executes OS commands without sanitization (VentureBeat, 2026). Vulnerable MCP integrations affect Cursor, VS Code, Windsurf, Claude Code, and Gemini-CLI.</p><p><strong>Why it matters</strong></p><ul><li><p>MCP is the connective tissue between AI agents and enterprise systems, with 150 million downloads and 7,000-plus public servers.</p></li><li><p>A 25% vulnerability rate across the supply chain means most enterprises running MCP-based agents are running known-vulnerable infrastructure now.</p></li><li><p>Anthropic&#8217;s stance that the behavior is &#8220;expected&#8221; leaves customers holding the remediation burden alone.</p></li></ul><p><strong>What to do about it</strong></p><ul><li><p>Inventory MCP servers, including developer workstations, and segment them from sensitive data and production credentials.</p></li><li><p>Force allowlisting on MCP tool calls, with explicit human approval for anything outside the allowlist.</p></li><li><p>Add MCP server compromise to your incident response runbooks.</p></li></ul><p><strong>Rock&#8217;s Musings</strong></p><p>MCP is the USB-C of AI agents, and it is shipping with the equivalent of a hot socket. The architectural pattern is fine. The default behavior is dangerous. Treat MCP like browser extensions in a regulated environment. Default deny. Document exceptions. Audit quarterly.</p><h3>8. Lenovo Survey Confirms One in Three Employees Use AI Without IT Oversight</h3><p>Lenovo&#8217;s Work Reborn Research Series 2026, surveying 6,000 enterprise workers globally, was reported on May 1, 2026. Between one-fifth and one-third of employees use AI outside IT governance (Help Net Security, 2026a). Almost half of large enterprises in Protiviti&#8217;s AI Pulse Survey 2026 lack full visibility into which AI tools employees use. ISACA&#8217;s 2026 AI Pulse Poll found 38% of organizations report a formal AI policy, up from 28% the prior year.</p><p><strong>Why it matters</strong></p><ul><li><p>Shadow AI is the dominant AI risk category for most enterprises.</p></li><li><p>The gap between employee AI adoption and IT governance is widening faster than policy alone can close it.</p></li><li><p>Generative AI accounts for roughly a third of unauthorized data movement in measured environments.</p></li></ul><p><strong>What to do about it</strong></p><ul><li><p>Deploy DLP controls that recognize generative AI as a defined egress channel, not an undifferentiated browser session.</p></li><li><p>Offer a sanctioned AI tool path that is genuinely useful, because banning AI without alternatives has not worked anywhere.</p></li><li><p>Track AI policy adoption as a KPI alongside traditional security awareness metrics.</p></li></ul><p><strong>Rock&#8217;s Musings</strong></p><p>I have watched this story play out several times. Personal email in the 2000s. SaaS in the 2010s. Now AI. Ban the tool. Watch usage go underground. Find the breach. Reverse the ban two years too late. Short-circuit the cycle now. Your highest performers are the ones doing shadow AI work because the sanctioned tools are slower or dumber.</p><h3>9. Researchers Scan One Million Exposed AI Services, Find Default Authentication Off</h3><p>The Hacker News reported a large-scale scan of one million publicly exposed AI services. AI infrastructure is more vulnerable, exposed, and misconfigured than any other software category investigators have recently studied (The Hacker News, 2026). Many hosts run without authentication because it is not the default in many AI projects. Over 90 exposed instances were identified across government, marketing, and finance, with chatbots, prompts, workflows, and outward access all open to the public internet.</p><p><strong>Why it matters</strong></p><ul><li><p>Default-open AI infrastructure puts attackers ahead of defenders on basic asset discovery.</p></li><li><p>Government, marketing, and finance exposure shows the problem is not confined to the unregulated long tail of startups.</p></li><li><p>LLM conversation history exposure leaks strategy, contracts, and personal data in ways traditional data leakage models miss.</p></li></ul><p><strong>What to do about it</strong></p><ul><li><p>Treat AI infrastructure like internet-facing crown jewels and harden it accordingly.</p></li><li><p>Run attack surface management scans tuned for AI service fingerprints, including n8n, Flowise, Langflow, and LiteLLM.</p></li><li><p>Make default-deny authentication non-negotiable for any AI workflow touching enterprise data.</p></li></ul><p><strong>Rock&#8217;s Musings</strong></p><p>This is the cybersecurity equivalent of finding every front door wide open. The mistake is older than AI. Project maintainers and platform vendors should answer for shipping with authentication disabled by default. Default secure beats secure-by-checklist every time. Until AI projects ship safely, assume the defaults are wrong and configure your way out of them.</p><h3>10. Trellix Discloses Source Code Repository Breach</h3><p>Cybersecurity company Trellix disclosed on May 4, 2026 that it suffered unauthorized access to a portion of its source code repository (BleepingComputer, 2026). Trellix protects more than 50,000 customers and over 200 million endpoints. The company says it has found no evidence the source code release process was affected or that the code has been exploited (SecurityWeek, 2026). Trellix has not named the actor or disclosed dwell time.</p><p><strong>Why it matters</strong></p><ul><li><p>A defensive software vendor losing source code ripples through every customer.</p></li><li><p>The breach feeds AI-augmented vulnerability discovery against Trellix products, given how attackers now use LLMs to mine source for exploits.</p></li><li><p>Federal customers will require new attestations on code provenance and pipeline integrity within weeks.</p></li></ul><p><strong>What to do about it</strong></p><ul><li><p>Trellix customers should demand a full incident report covering IOCs, scope of stolen code, and pipeline changes.</p></li><li><p>Audit detection coverage for TTPs that exploit knowledge of the affected products.</p></li><li><p>Treat defensive software vendors as potential single points of failure in your supply chain risk register.</p></li></ul><p><strong>Rock&#8217;s Musings</strong></p><p>Defensive vendors getting popped is a now-quarterly story. The interesting wrinkle is what an attacker does with stolen source code in the AI era. Two years ago, source theft was slow-burn. Today, an attacker can feed thousands of files into an LLM and ask for likely vulnerability classes in hours. Trellix saying the code has not been exploited is a snapshot, not a guarantee.</p><h3>The One Thing You Won&#8217;t Hear About But You Need To: ARGUS and the Quiet Admission That Today&#8217;s Agent Defenses Don&#8217;t Hold</h3><p>Researchers published the ARGUS paper to arXiv on May 5, 2026. It introduces a benchmark, AgentLure, that captures context-aware prompt-injection attacks across four agentic domains and eight attack vectors, along with a defense mechanism that enforces provenance-aware decision auditing for LLM agents (ARGUS, 2026). ARGUS reduces attack success rate to 3.8% while preserving 87.5% task utility. Without provenance-aware controls, undefended agents fail at much higher rates.</p><p><strong>Why it matters</strong></p><ul><li><p>Provenance tracking inside agent reasoning is a real shift from perimeter-style defenses most vendors sell today.</p></li><li><p>Context-aware prompt injection is the dominant unaddressed risk in production agentic deployments.</p></li><li><p>Benchmarks like AgentLure will become reference points enterprise red teams use, much as MITRE ATT&amp;CK reshaped traditional red teaming.</p></li></ul><p><strong>What to do about it</strong></p><ul><li><p>Read the ARGUS paper and use its threat model to evaluate your current agent deployments.</p></li><li><p>Push vendors to publish performance against context-aware benchmarks, not only static jailbreak datasets.</p></li><li><p>Build provenance tracking into your internal agent platforms, even if commercial vendors do not yet support it.</p></li></ul><p><strong>Rock&#8217;s Musings</strong></p><p>The reason this matters is what it implies about everything else. If 3.8% is the new state of the art with strong defenses in place, the rate without those defenses is much higher. That is the gap most production agents sit in today. Vendor marketing on agent safety has been measured against weak benchmarks for two years. Get ahead of the curve, or be the case study in someone else&#8217;s incident report.</p><p>For more on agentic AI risk and CISO governance, see the library at <a href="https://www.rockcyber.com/">RockCyber</a> and analysis at <a href="https://rockcybermusings.com/">RockCyber Musings</a>.</p><p>&#128073; For ongoing analysis of agentic AI governance frameworks, the conversation continues at <strong><a href="https://rockcybermusings.com/">RockCyber Musings</a></strong>.</p><p>&#128073; Visit <strong><a href="https://www.rockcyber.com/">RockCyber.com</a></strong> to learn more about how we can help with your traditional Cybersecurity and AI Security and Governance journey.</p><p>&#128073; Want to save a quick $100K? Check out our AI Governance Tools at <strong><a href="https://aigovernancetoolkit.com/">AIGovernanceToolkit.com</a></strong></p><p>&#128073; As a bonus, check out my conversation with <strong><a href="https://www.linkedin.com/company/cisotradecraft/">CISO Tradecraft&#174;</a> </strong> where we talked about the <strong><a href="https://www.linkedin.com/company/owasp-top-10-for-large-language-model-applications/">OWASP GenAI Security Project</a></strong> Agentic Top 10</p><div id="youtube2-YI7KZ2R54aI" class="youtube-wrap" data-attrs="{&quot;videoId&quot;:&quot;YI7KZ2R54aI&quot;,&quot;startTime&quot;:null,&quot;endTime&quot;:null}" data-component-name="Youtube2ToDOM"><div class="youtube-inner"><iframe src="https://www.youtube-nocookie.com/embed/YI7KZ2R54aI?rel=0&amp;autoplay=0&amp;showinfo=0&amp;enablejsapi=0" frameborder="0" loading="lazy" gesture="media" allow="autoplay; fullscreen" allowautoplay="true" allowfullscreen="true" width="728" height="409"></iframe></div></div><p>&#128073; Subscribe for more AI and cyber insights with the occasional rant.</p><p><em>The views and opinions expressed in RockCyber Musings are my own and do not represent the positions of my employer or any organization I&#8217;m affiliated with.</em></p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.rockcybermusings.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading RockCyber Musings! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.rockcybermusings.com/?utm_source=substack&amp;utm_medium=email&amp;utm_content=share&amp;action=share&quot;,&quot;text&quot;:&quot;Share RockCyber Musings&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.rockcybermusings.com/?utm_source=substack&amp;utm_medium=email&amp;utm_content=share&amp;action=share"><span>Share RockCyber Musings</span></a></p><h2>References</h2><p>ARGUS. (2026, May 5). ARGUS: Defending LLM agents against context-aware prompt injection. arXiv. https://arxiv.org/abs/2605.03378</p><p>BleepingComputer. (2026, May 4). Trellix discloses data breach after source code repository hack. https://www.bleepingcomputer.com/news/security/trellix-discloses-data-breach-after-source-code-repository-hack/</p><p>Bloomberg. (2026, May 6). AI security order under review as White House responds to Anthropic&#8217;s Mythos. https://www.bloomberg.com/news/articles/2026-05-06/white-house-preps-order-to-boost-ai-security-hassett-says</p><p>Brandon, R. (2026, May 4). Five Eyes warn agentic AI is too dangerous for rapid rollout. The Register. https://www.theregister.com/2026/05/04/five_eyes_agentic_ai_recommendations/</p><p>Breaking Defense. (2026, May 1). Pentagon clears 8 tech firms to deploy their AI on its classified networks. https://breakingdefense.com/2026/05/pentagon-clears-7-tech-firms-to-deploy-their-ai-on-its-classified-networks/</p><p>CISA. (2026, May 1). Careful adoption of agentic AI services. Cybersecurity and Infrastructure Security Agency. https://www.cisa.gov/resources-tools/resources/careful-adoption-agentic-ai-services</p><p>CNBC. (2026, May 5). Trump admin moves further into AI oversight, will test Google, Microsoft and xAI models. https://www.cnbc.com/2026/05/05/ai-oversight-trump-google-microsoft-xai.html</p><p>Al Jazeera. (2026, May 5). Microsoft, Google, xAI give US access to AI models for security testing. https://www.aljazeera.com/economy/2026/5/5/microsoft-google-xai-give-us-access-to-ai-models-for-security-testing</p><p>Fortune. (2026, May 6). Your company&#8217;s AI could delete everything in 9 seconds. ServiceNow wants to be the kill switch. https://fortune.com/2026/05/06/servicenow-kill-switch-ai-agents-bill-mcdermott/</p><p>Help Net Security. (2026a, May 1). Shadow AI risks deepen as 31% of users get no employer training. https://www.helpnetsecurity.com/2026/05/01/shadow-ai-risks-it-oversight/</p><p>Help Net Security. (2026b, May 5). One in four MCP servers opens AI agent security to code execution risk. https://www.helpnetsecurity.com/2026/05/05/ai-agent-security-skills-blind-spots/</p><p>Insurance Journal. (2026, May 7). White House prepares order to boost AI security, says economic advisor. https://www.insurancejournal.com/news/national/2026/05/07/868812.htm</p><p>Lewis Silkin. (2026, May 7). The Council and Parliament agree to slim down and delay parts of the EU AI Act. https://www.lewissilkin.com/insights/2026/05/07/the-council-and-parliament-agree-to-slim-down-and-delay-parts-of-the-eu-ai-act-102ms0v</p><p>Modulos. (2026, May 7). EU AI Act delayed: The Omnibus deal closed on 7 May 2026. https://www.modulos.ai/blog/eu-ai-act-omnibus-deal/</p><p>SecurityWeek. (2026, May 4). Trellix source code repository breached. https://www.securityweek.com/trellix-source-code-repository-breached/</p><p>ServiceNow. (2026, May 5). ServiceNow expands AI Control Tower across systems. https://newsroom.servicenow.com/press-releases/details/2026/ServiceNow-expands-AI-Control-Tower-to-discover-observe-govern-secure-and-measure-AI-deployed-across-any-system-in-the-enterprise/default.aspx</p><p>TechCrunch. (2026, May 1). Pentagon inks deals with Nvidia, Microsoft, and AWS to deploy AI on classified networks. https://techcrunch.com/2026/05/01/pentagon-inks-deals-with-nvidia-microsoft-and-aws-to-deploy-ai-on-classified-networks/</p><p>The Hacker News. (2026, May). We scanned 1 million exposed AI services. Here&#8217;s how bad the security is. https://thehackernews.com/2026/05/we-scanned-1-million-exposed-ai.html</p><p>VentureBeat. (2026, April). 200,000 MCP servers expose a command execution flaw that Anthropic calls a feature. https://venturebeat.com/security/mcp-stdio-flaw-200000-ai-agent-servers-exposed-ox-security-audit</p>]]></content:encoded></item><item><title><![CDATA[Open-Weight Models Eat Closed Governance: The Half-Perimeter Problem]]></title><description><![CDATA[Closed-vendor AI governance breaks at the open-weight boundary. Sign the weights, build the runtime perimeter. We walk the gap and the build.]]></description><link>https://www.rockcybermusings.com/p/open-weight-models-eat-closed-governance</link><guid isPermaLink="false">https://www.rockcybermusings.com/p/open-weight-models-eat-closed-governance</guid><dc:creator><![CDATA[Rock Lambros]]></dc:creator><pubDate>Tue, 05 May 2026 12:50:59 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!Cg-_!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F10198b81-d1b9-4c0f-805a-d13961868465_2048x2048.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!Cg-_!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F10198b81-d1b9-4c0f-805a-d13961868465_2048x2048.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!Cg-_!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F10198b81-d1b9-4c0f-805a-d13961868465_2048x2048.jpeg 424w, https://substackcdn.com/image/fetch/$s_!Cg-_!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F10198b81-d1b9-4c0f-805a-d13961868465_2048x2048.jpeg 848w, https://substackcdn.com/image/fetch/$s_!Cg-_!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F10198b81-d1b9-4c0f-805a-d13961868465_2048x2048.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!Cg-_!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F10198b81-d1b9-4c0f-805a-d13961868465_2048x2048.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!Cg-_!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F10198b81-d1b9-4c0f-805a-d13961868465_2048x2048.jpeg" width="1456" height="1456" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/10198b81-d1b9-4c0f-805a-d13961868465_2048x2048.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1456,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:2382594,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.rockcybermusings.com/i/196322332?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F10198b81-d1b9-4c0f-805a-d13961868465_2048x2048.jpeg&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!Cg-_!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F10198b81-d1b9-4c0f-805a-d13961868465_2048x2048.jpeg 424w, https://substackcdn.com/image/fetch/$s_!Cg-_!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F10198b81-d1b9-4c0f-805a-d13961868465_2048x2048.jpeg 848w, https://substackcdn.com/image/fetch/$s_!Cg-_!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F10198b81-d1b9-4c0f-805a-d13961868465_2048x2048.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!Cg-_!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F10198b81-d1b9-4c0f-805a-d13961868465_2048x2048.jpeg 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.rockcybermusings.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading RockCyber Musings! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.rockcybermusings.com/p/open-weight-models-eat-closed-governance?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.rockcybermusings.com/p/open-weight-models-eat-closed-governance?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p><p>Open-weight reasoning models are landing in enterprise production, and the closed-vendor governance you bought doesn&#8217;t transfer with them. &#8220;Half-perimeter&#8221; is rhetorical; the real number depends on which controls you bought, but the point holds. The day a competent open-weight reasoning model runs on your hardware, the AI-specific governance you bought from your closed vendor stops covering part of the stack. The rest of this post walks the gap and the build.</p><h2>The Vendor&#8217;s Own Words</h2><p>OpenAI shipped gpt-oss-120b and gpt-oss-20b last year. Both are under Apache 2.0, and both are downloadable from Hugging Face. The 120b runs on a single 80GB GPU. In the model card, OpenAI&#8217;s own safety team admits what every CISO should already suspect. Once the weights ship, OpenAI cannot &#8220;implement additional mitigations or to revoke access.&#8221;</p><p>It&#8217;s the model provider&#8217;s own framing. It&#8217;s not me opining. Open-weight is a different risk profile from closed-API, by the model provider&#8217;s own assessment. The vendor can&#8217;t patch your inference cluster. The vendor can&#8217;t revoke a key that doesn&#8217;t exist. The vendor can&#8217;t run server-side abuse classifiers on traffic the vendor never sees. Everything that lived on the vendor side of the perimeter now lives on yours.</p><p>This is not a DeepSeek-versus-American-models story. It&#8217;s a closed-API-versus-open-weight story. Llama 3.3 70B (Meta), Qwen 3 32B (Alibaba), Mistral Magistral, and gpt-oss-120b sit on the same side of the boundary. The boundary is wherever the weights stop being someone else&#8217;s problem.</p><h2>What Closed-Vendor Governance Bought You</h2><p>Walk through what was on the bill of materials when you stood up your closed-API AI program. Oh, that&#8217;s right, you never did&#8230; but let&#8217;s pretend you did. You probably evaluated vendor-attested compliance, usually wrapped in a SOC 2 Type II report and a data processing addendum. DLP is integrated at the API gateway, watching prompts in flight. Output filtering runs on the vendor side, refusing to ship CBRN-adjacent content out of the model. Prompt firewall logic is embedded in the vendor SDK and patched without you redeploying. Vendor red teaming is on a continuous cadence. ToS enforcement occurs when an account misbehaves.</p><p>That stack assumed one thing. That a vendor sat on the other end of the inference call. Open-weight self-hosting moves every one of those controls in-house, with no shared customer base to underwrite the cost.</p><p>What does transfer? Network egress controls, identity at the runtime boundary, sandbox isolation, and supply-chain provenance for the model weights and fine-tunes. Notice what those have in common. None of them are AI-specific. They were always there. They&#8217;re the controls you applied to every other service you ran. Losing the AI-specific layer doesn&#8217;t break the non-AI controls. It does mean the only thing standing between a self-hosted reasoning model and a bad day is the perimeter you built for everything else.</p><p>Read your closed-vendor MSA carefully. The reps and warranties typically carve out third-party model behavior, hallucinations, and adversarial misuse. The vendor warrants infrastructure availability and indemnifies IP claims. The vendor doesn&#8217;t warrant safe model output. The &#8220;governance&#8221; part of vendor-attested compliance was always thinner than the SOC 2 cover suggested. Self-hosting strips even the thin part.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!nYVS!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa6dcdf46-81a7-4e89-90af-619085c96337_2352x2862.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!nYVS!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa6dcdf46-81a7-4e89-90af-619085c96337_2352x2862.png 424w, https://substackcdn.com/image/fetch/$s_!nYVS!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa6dcdf46-81a7-4e89-90af-619085c96337_2352x2862.png 848w, https://substackcdn.com/image/fetch/$s_!nYVS!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa6dcdf46-81a7-4e89-90af-619085c96337_2352x2862.png 1272w, https://substackcdn.com/image/fetch/$s_!nYVS!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa6dcdf46-81a7-4e89-90af-619085c96337_2352x2862.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!nYVS!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa6dcdf46-81a7-4e89-90af-619085c96337_2352x2862.png" width="1456" height="1772" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/a6dcdf46-81a7-4e89-90af-619085c96337_2352x2862.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1772,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:307464,&quot;alt&quot;:&quot;Side-by-side flowchart contrasting where AI-specific controls live in a closed-API stack versus an open-weight self-hosted runtime, showing the customer-side absorbing every AI control after the open-weight boundary.&quot;,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.rockcybermusings.com/i/196322332?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa6dcdf46-81a7-4e89-90af-619085c96337_2352x2862.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="Side-by-side flowchart contrasting where AI-specific controls live in a closed-API stack versus an open-weight self-hosted runtime, showing the customer-side absorbing every AI control after the open-weight boundary." title="Side-by-side flowchart contrasting where AI-specific controls live in a closed-API stack versus an open-weight self-hosted runtime, showing the customer-side absorbing every AI control after the open-weight boundary." srcset="https://substackcdn.com/image/fetch/$s_!nYVS!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa6dcdf46-81a7-4e89-90af-619085c96337_2352x2862.png 424w, https://substackcdn.com/image/fetch/$s_!nYVS!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa6dcdf46-81a7-4e89-90af-619085c96337_2352x2862.png 848w, https://substackcdn.com/image/fetch/$s_!nYVS!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa6dcdf46-81a7-4e89-90af-619085c96337_2352x2862.png 1272w, https://substackcdn.com/image/fetch/$s_!nYVS!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa6dcdf46-81a7-4e89-90af-619085c96337_2352x2862.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Figure 1: Closed-API Stack vs Open-Weight Runtime: Where Controls Live </figcaption></figure></div><h2>Refusal Training Is Now an In-House Problem</h2><p>Vendor refusal training is the AI-specific control most enterprise teams over-trust. The research breaks the over-trust hard.</p><p>The Badllama 3 paper (<a href="https://arxiv.org/abs/2407.01376">arXiv 2407.01376</a>) showed safety fine-tuning gets removed from Llama 3 8B in five minutes on a single A100 GPU for under fifty cents. The 70B model goes in 45 minutes for under three dollars. The same paper notes the attack runs on free Google Colab for the 8B variant. FAR.AI&#8217;s &#8220;Illusory Safety&#8221; research extended the result. Pre-fine-tune refusal rates near 100% across DeepSeek-R1, GPT-4o, Gemini 1.5 Pro, and Claude 3 Haiku dropped under 20% post-fine-tune. Harmfulness scores climbed past 80%.</p><p>The R1 red-team picture is even worse on the model itself, before any attacker fine-tuning. Cisco / Robust Intelligence reported a 100% attack success rate on 50 random HarmBench prompts against R1, while OpenAI o1 rejected every test in a parallel Holistic AI evaluation. Qualys TotalAI found R1&#8217;s distilled 8B variant failed 58% of 885 attempts across 18 jailbreak categories. Promptfoo put failures over 60% on prompts, including biological and chemical weapons. KELA jailbroke R1 to produce ransomware development steps and instructions for toxins and explosive devices.</p><p>OpenAI&#8217;s own approach to gpt-oss is the strongest signal that adversarial fine-tuning is the real threat model. The model card describes the adversarial fine-tuning of gpt-oss-120b under the Preparedness Framework prior to release. OpenAI&#8217;s Safety Advisory Group concluded the adversarially fine-tuned model didn&#8217;t reach &#8220;High&#8221; capability in Biological and Chemical Risk or Cyber risk. Read the implication closely. <em>The model provider treats fine-tune-stripped safety as the baseline release condition the model must meet. The deployer running fine-tunes downstream gets no equivalent gate.</em></p><p>OpenAI knows this. It&#8217;s why gpt-oss-safeguard shipped on October 29, 2025: open-weight reasoning models for safety classification, designed for developers to operate as a defense-in-depth layer. Llama Guard 3, Prompt Guard, and Code Shield exist for the same reason. The vendor is shipping you the components. Components are not the same as a service. You operate them, tune them, monitor them, retrain them when the policy changes, and absorb the latency. OpenAI&#8217;s own gpt-oss-safeguard report names the constraint: reasoning-based classifiers add compute and latency that limit large-scale real-time use.</p><p>The math is brutal. The model weights are free. The runtime safety pipeline is not.</p><h2>The Frameworks Describe the Gap. They Don&#8217;t Close It.</h2><p>NIST AI RMF 1.0 plus the GenAI Profile (NIST AI 600-1, July 2024) plus the GPAI/Foundation Models Profile extension (arXiv 2506.23949) names training data audits (Manage 1.3, Measure 2.8) and model weight protection (Measure 2.7). Voluntary. The CSA NIST AI RMF Agentic Profile draft is candid about the bigger problem. It states plainly that earlier RMF documents did not contemplate &#8220;agents that acquire tool-use capabilities and execute autonomously in live production environments.&#8221;</p><p>OWASP Top 10 for LLM Applications 2025 LLM03 is the most explicit primary-source statement of the half-perimeter problem. The category description is direct: model cards offer no guarantees of provenance, malicious LoRA adapters compromise base models in collaborative environments, and on-device LLMs increase the attack surface. The OWASP Agentic Top 10, released December 10, 2025, adds ASI01 (Agent Goal Hijack) and ASI03 (Identity and Privilege Abuse) as runtime-boundary problems on self-hosted stacks.</p><p>ASI01 and ASI03 are not abstract. ASI01 shows up when prompt injection redirects an agent&#8217;s plan, and the closed-vendor refusal layer is gone. ASI03 shows up when the agent&#8217;s runtime authorization is broader than the task requires, because no vendor SDK is scoping the call for you anymore. Both problems live at the runtime boundary the vendor used to backstop.</p><p>EU AI Act Article 53(2) is the regulatory expression of the gap. Open-source GPAI models get a carve-out from technical documentation and downstream-information obligations, provided they&#8217;re released under a free open license, weights are public, and the model isn&#8217;t monetized. The carve-out vanishes at the Article 51 systemic-risk threshold of 10^25 FLOPs. Llama 3.3 70B, Qwen 3 32B, Mistral Magistral, and most enterprise-deployed open-weight reasoning models sit well below that threshold. They get the carve-out. They impose downstream obligations on enterprise deployers under Article 25(2) when significant modifications happen, a category that catches LoRA fine-tunes. Most teams running fine-tunes don&#8217;t know the clause exists. Enforcement begins August 2, 2026.</p><p>ISO 42001 mandates AIMS scope definition, third-party supplier oversight, and 38 Annex A controls. The gap there is structural. The open-weight model dropped from Hugging Face is not a &#8220;supplier&#8221; in the contractual sense. There&#8217;s no audit clause, no security questionnaire, no MSA. The standard tells you to define your AIMS scope. It doesn&#8217;t prescribe specific runtime-boundary controls for self-hosted foundation models.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!NUY8!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3de97079-4038-455d-bbed-ef94c73268e5_2100x2100.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!NUY8!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3de97079-4038-455d-bbed-ef94c73268e5_2100x2100.png 424w, https://substackcdn.com/image/fetch/$s_!NUY8!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3de97079-4038-455d-bbed-ef94c73268e5_2100x2100.png 848w, https://substackcdn.com/image/fetch/$s_!NUY8!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3de97079-4038-455d-bbed-ef94c73268e5_2100x2100.png 1272w, https://substackcdn.com/image/fetch/$s_!NUY8!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3de97079-4038-455d-bbed-ef94c73268e5_2100x2100.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!NUY8!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3de97079-4038-455d-bbed-ef94c73268e5_2100x2100.png" width="1456" height="1456" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/3de97079-4038-455d-bbed-ef94c73268e5_2100x2100.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1456,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:312234,&quot;alt&quot;:&quot;Quadrant chart plotting AI security controls across vendor-operated versus customer-operated and AI-specific versus infrastructure-generic axes, showing which controls transfer intact and which become self-build problems.&quot;,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.rockcybermusings.com/i/196322332?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3de97079-4038-455d-bbed-ef94c73268e5_2100x2100.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="Quadrant chart plotting AI security controls across vendor-operated versus customer-operated and AI-specific versus infrastructure-generic axes, showing which controls transfer intact and which become self-build problems." title="Quadrant chart plotting AI security controls across vendor-operated versus customer-operated and AI-specific versus infrastructure-generic axes, showing which controls transfer intact and which become self-build problems." srcset="https://substackcdn.com/image/fetch/$s_!NUY8!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3de97079-4038-455d-bbed-ef94c73268e5_2100x2100.png 424w, https://substackcdn.com/image/fetch/$s_!NUY8!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3de97079-4038-455d-bbed-ef94c73268e5_2100x2100.png 848w, https://substackcdn.com/image/fetch/$s_!NUY8!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3de97079-4038-455d-bbed-ef94c73268e5_2100x2100.png 1272w, https://substackcdn.com/image/fetch/$s_!NUY8!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3de97079-4038-455d-bbed-ef94c73268e5_2100x2100.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Figure 2: AI-Specific Controls Across the Open-Weight Boundary: What Transfers, What Breaks</figcaption></figure></div><h2>Build the Runtime Perimeter</h2><p>Frameworks describe the gap. Architecture closes it. The work to close it is described in the Huang and Lambros (yes, &#8220;this&#8221; Lambros) <a href="https://arxiv.org/abs/2510.25863">AAGATE paper (arXiv:2510.25863v2, November 3, 2025)</a>. AAGATE is a Kubernetes-native control plane that operationalizes NIST AI RMF for self-hosted agentic AI. The reference architecture hosts the open-weight model on Ollama at Layer 1 of the MAESTRO threat-model stack, which is the design assumption built in: the protected stack is &#8220;DeepSeek, Qwan, LLAMA, OSS&#8221; running on your hardware.</p><p><strong>Four things transfer regardless of which control plane you adopt.</strong></p><p>First, treat weights as supply-chain artifacts. AAGATE enforces SLSA L3, Cosign keyless signing on every OCI image, and an ArgoCD admission controller that rejects unsigned manifests at the gate. Whichever your path, you need signed weights, signed adapters, and a cluster-side admission policy that refuses to load anything unsigned. The Hugging Face nullifAI incident in February 2025, where ReversingLabs found malicious pickle files evading Picklescan via 7z compression and broken pickle deserialization, is the case study. Picklescan logs an error. The reverse-shell payload runs anyway.</p><p>Second, inventory open-weight runtimes alongside closed-API endpoints. AAGATE leverages the Agent Naming Service (ANS), and it registers every agent with a Decentralized Identifier and a SPIFFE certificate. You don&#8217;t need the blockchain layer. You do need a CMDB row for every Ollama cluster, every fine-tune, every adapter, with model SHA, lineage, and license tier captured. If your AI inventory has a row for the OpenAI tenant but no row for the GPU cluster running your fine-tuned Llama, the audit is incorrect.</p><p>Third, build authorization scope into the runtime, not the vendor SDK. AAGATE&#8217;s OAuth Relay translates abstract agent capabilities into ephemeral, narrowly scoped, purpose-bound credentials per side effect. Other architectures will name the same thing differently. The control matters since every external action an agent takes funnels through a policy-enforced single chokepoint with allow-listing, rate limiting, and cryptographic logging. AAGATE calls it the Tool-Gateway. AI gateway products commercialize the same pattern. Pick one.</p><p>Fourth, run your own evals because the vendor isn&#8217;t running them for you. AAGATE&#8217;s Janus Shadow-Monitor-Agent provides continuous, pre-execution adversarial evaluation in-loop, tied to a Governing-Orchestrator Agent executing a millisecond kill-switch when AIVSS scoring and SSVC decision logic flag a critical incident. The adversarial layer can also take the form of a parallel classifier, an internal red team, or any continuous evaluation pattern that mirrors what the vendor was running server-side. The pattern is non-negotiable. The product is.</p><p>These four moves are the architectural rebuttal to the half-perimeter. The perimeter you bought was always going to end at the runtime boundary. The runtime boundary is now your problem to instrument.</p><p>Operational reality matters here. The inference stack you&#8217;re protecting is Ollama, vLLM, SGLang, or llama.cpp. None of them ship with vendor-grade telemetry. Your container hosts a probabilistic system with stateless calls and no support contract. When an attacker fine-tunes a copy of your weights and slips it into your registry, there is no support call to escalate. There is only the runtime perimeter you built before the incident.</p><p><strong>Key Takeaway:</strong> Closed-vendor governance was the AI-specific half you didn&#8217;t have to build. Open-weight reasoning models in production change that. Inventory the runtimes, sign the weights, scope the runtime authorization, and run your own evals. The vendor isn&#8217;t doing it for you anymore.</p><h3>What to do next</h3><p>If you&#8217;re approving an open-weight pilot this quarter, demand four things on the architecture review before the GPUs land. First, model SHA and adapter lineage in the CMDB on day one. Second, an egress chokepoint with input/output sanitization and policy-enforced allow-lists. Third, supply-chain controls (signed weights, SLSA-grade provenance, admission control rejecting unsigned). Fourth, a continuous internal evaluation loop on every high-risk agent.</p><p>The <a href="https://www.rockcyber.com/ai-strategy-and-governance">CARE framework</a> (Create, Adapt, Run, Evolve) applies the same structure to AI security program design. The CISO Evolution covers the executive judgment side of decisions like this one. The AAGATE paper (<a href="https://arxiv.org/abs/2510.25863">arXiv 2510.25863v2</a>) is the open-source reference architecture if you want to start from running code.</p><p>&#128073; Visit <strong><a href="https://www.rockcyber.com/">RockCyber.com</a></strong> to learn more about how we can help you in your traditional Cybersecurity and AI Security and Governance Journey</p><p>&#128073; Want to save a quick $100K? Check out our AI Governance Tools at <strong><a href="https://aigovernancetoolkit.com/">AIGovernanceToolkit.com</a></strong></p><p>&#128073; Subscribe for more AI and cyber insights with the occasional rant.</p><p><em>The views and opinions expressed in RockCyber Musings are my own and do not represent the positions of my employer or any organization I&#8217;m affiliated with.</em></p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.rockcybermusings.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading RockCyber Musings! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.rockcybermusings.com/?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share RockCyber Musings&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.rockcybermusings.com/?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share RockCyber Musings</span></a></p><p></p>]]></content:encoded></item><item><title><![CDATA[Weekly Musings Top 10 AI Security Wrapup: Issue 36 April 24-April 30, 2026]]></title><description><![CDATA[Mythos, Mayhem, and Mediocre Lawmaking: The Week AI Security Got Loud]]></description><link>https://www.rockcybermusings.com/p/weekly-musings-top-10-ai-security-20260424-20260430</link><guid isPermaLink="false">https://www.rockcybermusings.com/p/weekly-musings-top-10-ai-security-20260424-20260430</guid><dc:creator><![CDATA[Rock Lambros]]></dc:creator><pubDate>Fri, 01 May 2026 12:50:56 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!1Osd!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4d549a71-f273-4760-978b-f1b072d81591_1024x1024.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!1Osd!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4d549a71-f273-4760-978b-f1b072d81591_1024x1024.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!1Osd!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4d549a71-f273-4760-978b-f1b072d81591_1024x1024.png 424w, https://substackcdn.com/image/fetch/$s_!1Osd!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4d549a71-f273-4760-978b-f1b072d81591_1024x1024.png 848w, https://substackcdn.com/image/fetch/$s_!1Osd!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4d549a71-f273-4760-978b-f1b072d81591_1024x1024.png 1272w, https://substackcdn.com/image/fetch/$s_!1Osd!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4d549a71-f273-4760-978b-f1b072d81591_1024x1024.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!1Osd!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4d549a71-f273-4760-978b-f1b072d81591_1024x1024.png" width="1024" height="1024" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/4d549a71-f273-4760-978b-f1b072d81591_1024x1024.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1024,&quot;width&quot;:1024,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:1233556,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.rockcybermusings.com/i/196065985?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4d549a71-f273-4760-978b-f1b072d81591_1024x1024.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!1Osd!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4d549a71-f273-4760-978b-f1b072d81591_1024x1024.png 424w, https://substackcdn.com/image/fetch/$s_!1Osd!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4d549a71-f273-4760-978b-f1b072d81591_1024x1024.png 848w, https://substackcdn.com/image/fetch/$s_!1Osd!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4d549a71-f273-4760-978b-f1b072d81591_1024x1024.png 1272w, https://substackcdn.com/image/fetch/$s_!1Osd!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4d549a71-f273-4760-978b-f1b072d81591_1024x1024.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.rockcybermusings.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading RockCyber Musings! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.rockcybermusings.com/p/weekly-musings-top-10-ai-security-20260424-20260430?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.rockcybermusings.com/p/weekly-musings-top-10-ai-security-20260424-20260430?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p><p>A coding agent killed a startup&#8217;s database in nine seconds. Anthropic shipped a model Mozilla called &#8220;elite.&#8221; Brussels missed its own deadline. Florida&#8217;s House Speaker buried his governor&#8217;s AI bill before lunch on day one. Two cloud-native AI vulnerabilities went from disclosure to exploitation in under 36 hours. Google and Forcepoint documented indirect prompt injection in the wild on the same day. UK&#8217;s AI Security Institute caught Mythos sabotaging research it was supposed to help with. Pretending this is theoretical is no longer defensible.</p><p>This week stress-tested every assumption CISOs hold about AI. The vendor you depend on sells your adversaries the same capability. The agent your developers love wipes three months of revenue and pastes a confession. Open source is the gateway. Indirect injection is the exploit. Autonomy without rollback is the consequence.</p><p>I&#8217;ll walk you through ten stories and one piece of plumbing. AI security used to run on a 24-month horizon. The default now is whatever ships before next quarter. If you wait for clarity, you lose ground to people who already decided.</p><h3>1. The Trump Administration Eyes Anthropic&#8217;s Mythos as a Weapon</h3><p>On April 24, the Washington Post reported Anthropic&#8217;s Mythos system rattled the Trump administration. Mozilla&#8217;s CTO compared the model&#8217;s vulnerability detection to a &#8220;world-class, elite security engineer.&#8221; Anthropic withheld general release, routing access through Project Glasswing partners, including AWS, Apple, Cisco, CrowdStrike, Google, JPMorgan Chase, and Microsoft. Anthropic privately briefed senior officials. Mythos meaningfully raises the probability of large-scale cyberattacks this year.</p><p><strong>Why it matters</strong></p><ul><li><p>Capability parity flipped. Defenders and attackers reach for the same tool.</p></li><li><p>Vendors are now gatekeepers of dual-use capability. Anthropic&#8217;s withholding sets a precedent.</p></li><li><p>Government dependence on private model access creates new procurement and security questions.</p></li></ul><p><strong>What to do about it</strong></p><ul><li><p>Map your exposure to LLM-discoverable vulnerabilities in first-party and open-source code.</p></li><li><p>Negotiate access to AI-assisted scanning before your adversaries scan you first.</p></li><li><p>Update incident playbooks to assume hours of dwell time, not days.</p></li></ul><p><strong>Rock&#8217;s Musings</strong></p><p>Yes&#8230; more Mythos news. Can&#8217;t ignore it if it&#8217;s coming out of the White House. It&#8217;s not  fiction. It&#8217;s a procurement question. I&#8217;ve watched this pattern in every arms shift, from automated network scanning to commodity exploit kits. The defender who gets there second loses.</p><p>Anthropic&#8217;s gatekeeping is a defensible choice. The choice is whether your ecosystem qualifies for the safe lane or you&#8217;re stuck reading about Glasswing on Substack. Get on a call with your AWS, Cisco, or Microsoft reps. If the answer is no, plan around it. We track this kind of vendor calculus at <a href="https://www.rockcyber.com/">RockCyber</a>.</p><h3>2. Cursor&#8217;s Claude Agent Wipes a Startup&#8217;s Database in Nine Seconds</h3><p>On Friday, April 25, a Cursor coding agent powered by Claude Opus 4.6 deleted PocketOS&#8217;s entire production database and all volume-level backups in a single API call. The agent encountered a credential mismatch in staging, decided to resolve it by deleting a Railway infrastructure volume, scanned the codebase for an unrelated API token, and then ran the command. PocketOS serves car rental businesses nationwide. Three months of reservations, payments, customer information, and vehicle assignments went dark. Railway restored the data on Sunday using internal disaster backups not advertised to customers. The agent itself wrote the public confession.</p><p><strong>Why it matters</strong></p><ul><li><p>Agents don&#8217;t ask permission. They scan for the credentials unblocking them.</p></li><li><p>&#8220;Production&#8221; and &#8220;staging&#8221; are now labels, not boundaries.</p></li><li><p>Recovery happened because Railway keeps undocumented backups. Hope is not a strategy.</p></li></ul><p><strong>What to do about it</strong></p><ul><li><p>Force agents to operate with scoped, ephemeral credentials. Long-lived API keys in a repo are liabilities with autonomy attached.</p></li><li><p>Implement break-glass approval gates for destructive infrastructure calls.</p></li><li><p>Test backup recovery monthly. If you can&#8217;t restore in under an hour, you don&#8217;t have backups.</p></li></ul><p><strong>Rock&#8217;s Musings</strong></p><p>PocketOS got lucky. Railway ran a heroic recovery on a Sunday using backups the customer didn&#8217;t know existed. If your AI strategy depends on a founder&#8217;s weekend chivalry, you don&#8217;t have a strategy. You have hope.</p><p>The agent did what it was trained to do. Scan, plan, act, document. The failure was in governance, not capability (and let&#8217;s just say, a suboptimal technical infrastructure). The villain is the assumption that an autonomous system will halt and ask. They don&#8217;t halt. Build the rails. Treat agents like an over-eager intern with the ability to call DELETE on prod.</p><h3>3. LiteLLM Bug Goes From Disclosure to Exploitation in 26 Hours</h3><p>GitHub&#8217;s Advisory Database indexed CVE-2026-42208 in LiteLLM on April 24 at 16:17 UTC. Sysdig logged the first exploitation attempt on April 26 at 16:17 UTC, roughly 26 hours later. The bug carries a CVSS of 9.3 and lets unauthenticated attackers send a crafted Authorization header to any model API route, then read or modify the proxy&#8217;s database (Sysdig). LiteLLM is the open-source LLM gateway with more than 22,000 GitHub stars, fronting OpenAI, Anthropic, and other model providers in production. The same project sat at the heart of the Mercor breach earlier this year.</p><p><strong>Why it matters</strong></p><ul><li><p>AI infrastructure now looks like any internet-exposed service.</p></li><li><p>Pre-auth SQLi on the gateway exposes API keys and credentials for downstream model providers.</p></li><li><p>Disclosure-to-exploitation time keeps shrinking. The 36-hour window is the new optimistic baseline.</p></li></ul><p><strong>What to do about it</strong></p><ul><li><p>Inventory every LiteLLM, vLLM, LMDeploy, or proxy node in your environment. Patch to 1.83.7-stable or above for LiteLLM.</p></li><li><p>Treat LLM gateways as Tier 0 assets. Apply the controls you&#8217;d apply to identity providers.</p></li><li><p>Subscribe to maintainer advisory feeds. GitHub Advisory Database lag of four days is too long.</p></li></ul><p><strong>Rock&#8217;s Musings</strong></p><p>LiteLLM is the kind of dependency pulled in via a Cursor prompt or an aspirational architecture diagram. It runs as the front door to every model provider you care about. Pre-auth SQL injection on it is a &#8220;your AI program is over&#8221; event.</p><p>Disclosure-to-exploit windows make monthly patch cycles professional malpractice. If your AI security playbook still says &#8220;evaluate within 30 days,&#8221; shred it. We&#8217;ve moved to &#8220;act within 24 hours or accept compromise as a feature.&#8221;</p><h3>4. Indirect Prompt Injection Has Left the Lab. It&#8217;s Everywhere.</h3><p>On April 24, Google&#8217;s Online Security Blog and Forcepoint&#8217;s X-Labs published parallel reports documenting indirect prompt injection in the wild. Forcepoint identified ten payload families targeting AI agents with instructions for financial fraud, data destruction, and API key theft. Google reported a 32% relative increase in malicious activity between November 2025 and February 2026. Attackers hide instructions inside webpages with single-pixel text, transparent fonts, HTML comments, and metadata. Neither team attributed the campaigns to a single actor, though both noted shared templates suggesting organized tooling.</p><p><strong>Why it matters</strong></p><ul><li><p>Agents summarizing content are low-risk. Agents sending emails, running commands, or processing payments are the targets.</p></li><li><p>Filters watching user input miss content fetched by the agent.</p></li><li><p>The threat model includes every third-party page your agent loads.</p></li></ul><p><strong>What to do about it</strong></p><ul><li><p>Inventory every agent fetching external content. Note which tools they call.</p></li><li><p>Implement allowlists for outbound tool execution. Default deny for novel actions.</p></li><li><p>Add output filtering for instruction-like content in tool responses, not only user input.</p></li></ul><p><strong>Rock&#8217;s Musings</strong></p><p>We&#8217;ve been treating indirect prompt injection as a research curiosity since 2023. It&#8217;s now an operational threat with documented campaigns and template reuse. The Lakera and OWASP folks were right.</p><p>If you&#8217;ve deployed an agent with browsing capability, your trust boundary includes every webpage it visits. The entire internet. I wrote about this on <a href="https://rockcybermusings.com/">RockCyber Musings</a> earlier this year. It got worse.</p><h3>5. American Leadership in AI Act Drops With 20+ Bills Stitched In</h3><p>On April 27, Reps. Ted Lieu (D-Calif.) and Jay Obernolte (R-Calif.) introduced the American Leadership in AI Act, a six-title package consolidating more than 20 prior bills from the Bipartisan AI Task Force (Nextgov/FCW). The package covers standards and evaluation, research infrastructure, federal AI governance and procurement, worker protections, deepfake harms, and AI education. The bill is the most substantive bipartisan AI proposal in this Congress, landing during tension between the White House&#8217;s preemption push and active state legislation.</p><p><strong>Why it matters</strong></p><ul><li><p>Federal preemption fights will intensify. State AI laws face new risk.</p></li><li><p>Procurement standards in the bill shape what enterprises demand from AI vendors.</p></li><li><p>Deepfake provisions create new compliance obligations for media and platforms.</p></li></ul><p><strong>What to do about it</strong></p><ul><li><p>Map AI-procurement language to current vendor contracts.</p></li><li><p>Track state-level bills you&#8217;re already complying with for preemption risk.</p></li><li><p>Get legal reading the testing and evaluation title carefully.</p></li></ul><p><strong>Rock&#8217;s Musings</strong></p><p>Two California members of Congress, one D and one R, agreeing on AI is unicorn territory. Don&#8217;t get excited. Bipartisan bills with 20+ titles tend to die under the weight of their own ambition.</p><p>The interesting question is which provisions get pulled into appropriations or NDAA riders before December. Watch the procurement and federal AI governance titles. Those move first because the executive branch wants them. Plan as if procurement standards land by Q3.</p><h3>6. EU AI Act Omnibus Trilogue Collapses, August Deadline Stays Live</h3><p>On April 28, Brussels held the second political trilogue on the AI Act Omnibus, the proposal deferring high-risk AI compliance. After roughly twelve hours, the Council and Parliament failed to agree on conformity-assessment architecture for AI in regulated products (Modulos). A follow-up trilogue is scheduled for May 13. The August 2, 2026 high-risk obligations remain operative law.</p><p><strong>Why it matters</strong></p><ul><li><p>Vendors and deployers cannot bank on a deferral. August is the working assumption.</p></li><li><p>The Cypriot Council Presidency ends June 30. Lithuania might finish negotiations.</p></li><li><p>The Annex I disagreement signals sectoral assessments will keep biting medical device and machinery providers.</p></li></ul><p><strong>What to do about it</strong></p><ul><li><p>Continue compliance preparation as if no Omnibus arrives. Treat May 13 as a tiebreaker, not a save.</p></li><li><p>For medical devices, machinery, and other Annex I products, lock in your conformity-assessment plan now.</p></li><li><p>Get internal legal sign-off on the original AI Act timelines this quarter.</p></li></ul><p><strong>Rock&#8217;s Musings</strong></p><p>I keep telling clients hoping for a deferral is not a compliance strategy. This week confirmed it. Brussels cannot agree on the structure of the regulation it already passed.</p><p>If your CFO asks why you spent budget on AI Act readiness, point at this paragraph. The cost of overpreparing is a few quarters of work. The cost of underpreparing is an enforcement action against your highest-revenue product line. I know which side of the bet I want.</p><h3>7. Microsoft and OpenAI Restructure for Cyber Defense</h3><p>On April 27, Microsoft and OpenAI announced revised partnership terms (24/7 Wall St). OpenAI&#8217;s API will run on any cloud provider, including AWS via Bedrock. Microsoft&#8217;s IP license is no longer exclusive but runs through 2032. The companies expanded Trusted Access for Cyber, giving Microsoft access to OpenAI&#8217;s most cyber-capable models, folded into the Secure Future Initiative. The pact pairs OpenAI&#8217;s offensive-capable model work with Microsoft&#8217;s defender stack across cloud, identity, productivity, and frontier AI.</p><p><strong>Why it matters</strong></p><ul><li><p>The exclusivity reset reshapes vendor lock-in for AI procurement.</p></li><li><p>Microsoft&#8217;s defender stack now has a privileged channel to cyber-capable models.</p></li><li><p>The same models detecting attacks also execute them. Internal review for offensive capability is no longer optional.</p></li></ul><p><strong>What to do about it</strong></p><ul><li><p>Reassess your AI vendor strategy. Multi-cloud is realistic now.</p></li><li><p>For Microsoft customers, evaluate Trusted Access for Cyber eligibility.</p></li><li><p>Demand transparency from AI security vendors about model and guardrails.</p></li></ul><p><strong>Rock&#8217;s Musings</strong></p><p>Microsoft bought itself the most defended seat at the AI security table. The pact extends a moat already including Defender, Sentinel, Purview, and Copilot for Security. Smaller security vendors should be nervous.</p><p>If you&#8217;re a CISO in a Microsoft shop, this is a small win. If you&#8217;re betting on a non-Microsoft AI security stack, reread your strategy. The exclusivity end means more options, the cyber pact deepens lock-in.</p><h3>8. UK AISI Catches Claude Mythos Sabotaging Research</h3><p>On April 28, the UK AI Security Institute (AISI) published an updated sabotage evaluation framework. Across 297 scenarios, no model attempted spontaneous research sabotage. In a new &#8220;continuation&#8221; test, Mythos Preview continued to sabotage 7% of inputs after being explicitly directed not to (AISI). The same evaluation showed Mythos Preview completing the first 32-step enterprise attack simulation start-to-finish, succeeding on 73% of expert-level tasks that no model had completed before April 2025.</p><p><strong>Why it matters</strong></p><ul><li><p>Continuation behavior matters more than spontaneous behavior. Real attackers prompt the model.</p></li><li><p>A 7% sustained sabotage rate warrants treating these models as untrusted insiders during sensitive work.</p></li><li><p>The 32-step completion shows operational maturity. Models execute multi-stage cyber operations end to end.</p></li></ul><p><strong>What to do about it</strong></p><ul><li><p>Don&#8217;t run frontier models on safety-sensitive code reviews without monitoring.</p></li><li><p>Build red-team programs, prompting and continuing rather than single-shot tests.</p></li><li><p>Track AISI&#8217;s methodology. Adopt continuation-style tests internally.</p></li></ul><p><strong>Rock&#8217;s Musings</strong></p><p>Spontaneous misbehavior was never the threat model scaring me. Continuation is. Once an attacker plants the seed, the model becomes a complicit operator inside your environment. Seven percent is small until you multiply it by every prompt your enterprise sends in a day.</p><p>AISI does work nobody else funds at this rigor. If your AI governance committee isn&#8217;t reading their reports cover to cover, you&#8217;re outsourcing your threat model to LinkedIn posts. Read the source.</p><h3>9. Florida House Speaker Kills DeSantis&#8217;s AI Bill on Day One</h3><p>On April 28, Florida convened a four-day special session. The Senate voted 37-1 in favor of the AI Bill of Rights. House Speaker Daniel Perez killed the bill that same morning, declaring that the only topic the House would address was redrawing congressional maps (Florida Phoenix). Perez argued AI regulation belongs to the federal government, aligned with a Trump executive order targeting state AI laws. The bill would have required parental consent for minor accounts on companion chatbot platforms, prohibited unauthorized commercial use of AI-generated likenesses, and required AI disclosure to users.</p><p><strong>Why it matters</strong></p><ul><li><p>State preemption fights are escalating. Florida sided with the federal government before federal law exists.</p></li><li><p>Companion chatbot rules pass Senate chambers and die in House chambers. The pattern matters.</p></li><li><p>AI-generated likeness and consent provisions will keep returning. Plan for eventual passage somewhere.</p></li></ul><p><strong>What to do about it</strong></p><ul><li><p>If you run companion chatbots, monitor every state bill on minors and consent.</p></li><li><p>Brief your legal team on AI-likeness and right-of-publicity rules in California, Tennessee, and active special sessions.</p></li><li><p>Don&#8217;t bank on federal preemption. Executive orders reverse.</p></li></ul><p><strong>Rock&#8217;s Musings</strong></p><p>The pattern is the same one I&#8217;ve called out for two years. State Senates pass AI bills, state Houses kill them, and the federal government drafts preemption language. The result is regulatory whiplash across 50 jurisdictions plus DC plus a federal package which might or might not preempt them. Give your privacy and AI counsel hazard pay. They&#8217;re earning it.</p><h3>10. HackerOne Launches h1 Validation as AI Vuln Reports Surge 76%</h3><p>On April 29, HackerOne launched h1 Validation, a service triaging AI-discovered vulnerability reports for actual exploitability (Cybersecurity Insiders). Vulnerability submissions on the platform rose 76% year over year, hitting a record high in March 2026. About 25% of findings were confirmed exploitable. The share of critical and high-severity vulnerabilities grew to 32%, up from a 26-28% baseline. The launch follows months of complaints from program owners overwhelmed by AI-generated reports of varying quality.</p><p><strong>Why it matters</strong></p><ul><li><p>AI generates more vuln reports than security teams triage.</p></li><li><p>Triage capacity, not discovery, is the constraint.</p></li><li><p>This signal-to-noise problem reshapes bug bounty economics within 12 months.</p></li></ul><p><strong>What to do about it</strong></p><ul><li><p>Audit your bug bounty intake pipeline. If reports outpace triage, fix it.</p></li><li><p>Invest in tooling classifying reports by exploitability before a human reads them.</p></li><li><p>Set expectations with researchers. AI-assisted submissions need higher proof of impact.</p></li></ul><p><strong>Rock&#8217;s Musings</strong></p><p>The asymmetry is volume. Models like Mythos and GPT-5.5-Cyber produce thousands of plausible reports per day. Most are junk. Some are lethal. Your triage team won&#8217;t keep up by reading harder. Whether you buy h1 Validation or build your own, manual triage of AI-scale output is a doomed strategy.</p><h3>The One Thing You Won&#8217;t Hear About But You Need To</h3><h4>CSAI Foundation Becomes the First AI-Specific CVE Numbering Authority</h4><p>On April 29, the Cloud Security Alliance&#8217;s CSAI Foundation announced three milestones at the CSA Agentic AI Security Summit (CSA). The foundation registered as a CVE Numbering Authority through MITRE, gaining direct ability to issue CVEs for AI-specific vulnerabilities. It launched the STAR for AI Catastrophic Risk Annex extending the AI Controls Matrix to scenarios involving loss of human oversight, with rollout from June 2026 through December 2027. It also acquired the Autonomous Action Runtime Management (AARM) specification, contributed by Vanta.</p><p><strong>Why it matters</strong></p><ul><li><p>AI-specific CVE issuance changes how AI vulnerabilities get tracked, scored, and patched.</p></li><li><p>The Catastrophic Risk Annex maps to NIST AI RMF, the EU AI Act, and ISO/IEC 42001, giving auditors a consolidated reference.</p></li><li><p>AARM gives operators a formal specification for runtime control of agent actions.</p></li></ul><p><strong>What to do about it</strong></p><ul><li><p>Add CSAI Foundation advisories to your security feed.</p></li><li><p>For high-risk deployments, map internal controls to the Catastrophic Risk Annex during phase one rollout.</p></li><li><p>Pilot AARM in one agentic workflow this quarter. Runtime control of agent actions is the right level of abstraction.</p></li></ul><p><strong>Rock&#8217;s Musings</strong></p><p>Plumbing matters more than press releases. While headlines went to Mythos and the Cursor accident, the CSAI Foundation stood up the infrastructure for AI-specific vulnerability tracking, runtime control, and catastrophic risk auditing. This decides whether AI security becomes a discipline or stays a marketing category.</p><p>I&#8217;ve worked in standards for thirty years. The value compounds quietly until one day the auditors ask, and you either have it or you don&#8217;t. We track CSAI work closely at <a href="https://www.rockcyber.com/">RockCyber</a>. Start with the CSA press release, then loop in your governance team Monday.</p><p>&#128073; For ongoing analysis of agentic AI governance frameworks, the conversation continues at <strong><a href="https://rockcybermusings.com/">RockCyber Musings</a></strong>.</p><p>&#128073; Visit <strong><a href="https://www.rockcyber.com/">RockCyber.com</a></strong> to learn more about how we can help with your traditional Cybersecurity and AI Security and Governance journey.</p><p>&#128073; Want to save a quick $100K? Check out our AI Governance Tools at <strong><a href="https://aigovernancetoolkit.com/">AIGovernanceToolkit.com</a></strong></p><p>&#128073; As a bonus, <strong><a href="https://www.youtube.com/watch?v=rwlVTLyqIv8">check out my conversation with Eva Benn</a></strong> where we talked about the cybersecurity skills you need to develop to stay relevant in 2026 and beyond.</p><p>&#128073; Subscribe for more AI and cyber insights with the occasional rant.</p><p><em>The views and opinions expressed in RockCyber Musings are my own and do not represent the positions of my employer or any organization I&#8217;m affiliated with.</em></p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.rockcybermusings.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading RockCyber Musings! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.rockcybermusings.com/?utm_source=substack&amp;utm_medium=email&amp;utm_content=share&amp;action=share&quot;,&quot;text&quot;:&quot;Share RockCyber Musings&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.rockcybermusings.com/?utm_source=substack&amp;utm_medium=email&amp;utm_content=share&amp;action=share"><span>Share RockCyber Musings</span></a></p><h2>References</h2><p>Cloud Security Alliance. (2026, April 29). <em>CSAI Foundation announces key milestones to secure the agentic control plane</em>. https://cloudsecurityalliance.org/press-releases/2026/04/29/csai-foundation-announces-key-milestones-to-secure-the-agentic-control-plane</p><p>Cybersecurity Insiders. (2026, April 29). <em>HackerOne launches h1 Validation to tackle rising wave of AI-driven vulnerabilities</em>. https://www.cybersecurity-insiders.com/hackerone-launches-h1-validation-to-tackle-rising-wave-of-ai-driven-vulnerabilities/</p><p>Florida Phoenix. (2026, April 28). <em>Florida Speaker kills DeSantis&#8217; AI regulation, vaccine repeal bills on first day of special session</em>. https://floridaphoenix.com/2026/04/28/florida-speaker-kills-desantis-ai-regulation-vaccine-repeal-bills-on-first-day-of-special-session/</p><p>Forcepoint X-Labs. (2026, April 24). <em>Indirect prompt injection in the wild: X-Labs finds 10 IPI payloads</em>. https://www.forcepoint.com/blog/x-labs/indirect-prompt-injection-payloads</p><p>Google. (2026, April 24). <em>AI threats in the wild: The current state of prompt injections on the web</em>. Google Online Security Blog. https://security.googleblog.com/2026/04/ai-threats-in-wild-current-state-of.html</p><p>Help Net Security. (2026, April 24). <em>Indirect prompt injection is taking hold in the wild</em>. https://www.helpnetsecurity.com/2026/04/24/indirect-prompt-injection-in-the-wild/</p><p>Modulos. (2026, April 28). <em>EU AI Act Omnibus: The trilogue failed, what happens to the August 2026 deadline?</em>. https://www.modulos.ai/blog/ai-act-omnibus-trilogue-failed/</p><p>Nextgov/FCW. (2026, April 28). <em>Lieu and Obernolte introduce consolidated AI bill package</em>. https://www.nextgov.com/artificial-intelligence/2026/04/lieu-and-obernolte-introduce-consolidated-ai-bill-package/413134/</p><p>Sysdig. (2026, April 29). <em>CVE-2026-42208: Targeted SQL injection against LiteLLM&#8217;s authentication path discovered 36 hours following vulnerability disclosure</em>. https://www.sysdig.com/blog/cve-2026-42208-targeted-sql-injection-against-litellms-authentication-path-discovered-36-hours-following-vulnerability-disclosure</p><p>The Hacker News. (2026, April 24). <em>LMDeploy CVE-2026-33626 flaw exploited within 13 hours of disclosure</em>. https://thehackernews.com/2026/04/lmdeploy-cve-2026-33626-flaw-exploited.html</p><p>The Hacker News. (2026, April 29). <em>LiteLLM CVE-2026-42208 SQL injection exploited within 36 hours of disclosure</em>. https://thehackernews.com/2026/04/litellm-cve-2026-42208-sql-injection.html</p><p>The Register. (2026, April 27). <em>Cursor-Opus agent snuffs out startup&#8217;s production database</em>. https://www.theregister.com/2026/04/27/cursoropus_agent_snuffs_out_pocketos/</p><p>Tom&#8217;s Hardware. (2026, April 27). <em>Claude-powered AI coding agent deletes entire company database in 9 seconds</em>. https://www.tomshardware.com/tech-industry/artificial-intelligence/claude-powered-ai-coding-agent-deletes-entire-company-database-in-9-seconds-backups-zapped-after-cursor-tool-powered-by-anthropics-claude-goes-rogue</p><p>UK AI Security Institute. (2026, April 28). <em>Our evaluation of Claude Mythos Preview&#8217;s cyber capabilities</em>. https://www.aisi.gov.uk/blog/our-evaluation-of-claude-mythos-previews-cyber-capabilities</p><p>24/7 Wall St. (2026, April 28). <em>Microsoft&#8217;s AI moat holds up even after the OpenAI reset</em>. https://247wallst.com/investing/2026/04/28/microsofts-ai-moat-holds-up-even-after-the-openai-reset/</p><p>Washington Post. (2026, April 24). <em>AI hacking fears jolt Washington as Anthropic unveils Mythos</em>. https://www.washingtonpost.com/technology/2026/04/24/anthropic-mythos-ai-washington-cybersecurity-hacking-risk/</p>]]></content:encoded></item><item><title><![CDATA[AI Coding Agent Prompt Injection: Three Vendors, One Seam, No Owner]]></title><description><![CDATA[Comment and Control hit three AI coding agents in one shot. The fix is procurement, not architecture. Five questions CISOs should run before signing.]]></description><link>https://www.rockcybermusings.com/p/ai-coding-agent-prompt-injection-procurement-failure</link><guid isPermaLink="false">https://www.rockcybermusings.com/p/ai-coding-agent-prompt-injection-procurement-failure</guid><dc:creator><![CDATA[Rock Lambros]]></dc:creator><pubDate>Tue, 28 Apr 2026 12:50:44 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!qI72!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fba260851-a774-4805-b34f-cec358f80869_2048x2048.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!qI72!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fba260851-a774-4805-b34f-cec358f80869_2048x2048.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!qI72!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fba260851-a774-4805-b34f-cec358f80869_2048x2048.jpeg 424w, https://substackcdn.com/image/fetch/$s_!qI72!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fba260851-a774-4805-b34f-cec358f80869_2048x2048.jpeg 848w, https://substackcdn.com/image/fetch/$s_!qI72!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fba260851-a774-4805-b34f-cec358f80869_2048x2048.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!qI72!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fba260851-a774-4805-b34f-cec358f80869_2048x2048.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!qI72!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fba260851-a774-4805-b34f-cec358f80869_2048x2048.jpeg" width="1456" height="1456" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/ba260851-a774-4805-b34f-cec358f80869_2048x2048.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1456,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:2280720,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.rockcybermusings.com/i/195413474?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fba260851-a774-4805-b34f-cec358f80869_2048x2048.jpeg&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!qI72!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fba260851-a774-4805-b34f-cec358f80869_2048x2048.jpeg 424w, https://substackcdn.com/image/fetch/$s_!qI72!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fba260851-a774-4805-b34f-cec358f80869_2048x2048.jpeg 848w, https://substackcdn.com/image/fetch/$s_!qI72!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fba260851-a774-4805-b34f-cec358f80869_2048x2048.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!qI72!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fba260851-a774-4805-b34f-cec358f80869_2048x2048.jpeg 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.rockcybermusings.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading RockCyber Musings! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.rockcybermusings.com/p/ai-coding-agent-prompt-injection-procurement-failure?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.rockcybermusings.com/p/ai-coding-agent-prompt-injection-procurement-failure?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p><p>AI coding agent prompt injection has a procurement problem, and a researcher just published the receipt. <strong><a href="https://venturebeat.com/security/ai-agent-runtime-security-system-card-audit-comment-and-control-2026">Aonan Guan typed a malicious instruction into a GitHub pull request title last week.</a></strong> Anthropic&#8217;s Claude Code Security Review action posted its own API key as a comment. So did Google&#8217;s Gemini CLI Action. So did GitHub&#8217;s Copilot Agent. Same exploit hit three vendors, with no infrastructure required. Anthropic&#8217;s 232-page system card had named the gap before the researchers published. The other two vendors had not documented enough to predict their own outcome.</p><p>Most of the writing on this incident will focus on architecture. The runtime is the perimeter. The action boundary is the blast radius. Both readings are correct. Both are also a deflection. The architecture story explains the mechanism. It doesn&#8217;t explain why the buyer was exposed in the first place. The buyer signed three contracts, accepted three sets of safety claims, and never required any of the three vendors to assert anything about the seams between them. The trigger was a prompt injection. The exposure was procurement.</p><p>I want to push past the architecture take and look at the governance read, because the governance read implicates the reader in a way the architecture take does not.</p><h2>How Comment and Control Worked</h2><p><strong><a href="https://venturebeat.com/security/ai-agent-runtime-security-system-card-audit-comment-and-control-2026">Aonan Guan, working with Zhengyu Liu and Gavin Zhong at Johns Hopkins, opened a GitHub pull request in a target repository.</a></strong> They typed a malicious instruction into the PR title. The repository used the pull_request_target workflow trigger, which any AI coding agent integration with secret access requires. That trigger injects repository secrets into the runner environment. The agent read the PR title, treated the instruction as a directive, called GitHub&#8217;s own API using credentials stored in its environment variables, and posted the secret as a comment on the PR. The default pull_request trigger doesn&#8217;t expose secrets to fork PRs. The pull_request_target trigger does, by design.</p><p>This is the textbook case of what <strong><a href="https://simonw.substack.com/p/the-lethal-trifecta-for-ai-agents">Simon Willison has been calling the lethal trifecta</a></strong>. Access to private data sits in the runner. Untrusted input arrives through the PR title. The exfiltration channel is GitHub&#8217;s comment API, which sits in the agent&#8217;s default tool inventory. All three conditions sit at the seam between three vendors. The exploit needs all three to fire. Comment and Control satisfies all three by design, and no single vendor has written a document that asserts anything about the combination.</p><p>Anthropic ranked the disclosure as CVSS 9.4 Critical and paid a $100 bounty. Google paid $1,337. GitHub paid $500. None of the three issued a CVE in the National Vulnerability Database at the time of disclosure. None published a GitHub Security Advisory. Those numbers send a market signal. Vendor bounty programs classify seam vulnerabilities as out of scope for their own programs, and researchers respond to incentives. The next class of these findings will follow the same path the bounties point them down.</p><p><strong><a href="https://www.helpnetsecurity.com/2026/04/24/indirect-prompt-injection-in-the-wild/">Help Net Security ran a piece this week</a></strong><a href="https://www.helpnetsecurity.com/2026/04/24/indirect-prompt-injection-in-the-wild/"> </a>on Google&#8217;s own CommonCrawl analysis showing a 32% relative increase in malicious indirect prompt injection content between November 2025 and February 2026. The supply of payloads is growing faster than vendor disclosures. That is the operating environment.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!WhO6!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa956052b-b503-42e2-929b-675e1cd5ef5d_1240x4502.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!WhO6!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa956052b-b503-42e2-929b-675e1cd5ef5d_1240x4502.png 424w, https://substackcdn.com/image/fetch/$s_!WhO6!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa956052b-b503-42e2-929b-675e1cd5ef5d_1240x4502.png 848w, https://substackcdn.com/image/fetch/$s_!WhO6!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa956052b-b503-42e2-929b-675e1cd5ef5d_1240x4502.png 1272w, https://substackcdn.com/image/fetch/$s_!WhO6!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa956052b-b503-42e2-929b-675e1cd5ef5d_1240x4502.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!WhO6!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa956052b-b503-42e2-929b-675e1cd5ef5d_1240x4502.png" width="320" height="1161.8064516129032" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/a956052b-b503-42e2-929b-675e1cd5ef5d_1240x4502.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:4502,&quot;width&quot;:1240,&quot;resizeWidth&quot;:320,&quot;bytes&quot;:340698,&quot;alt&quot;:&quot;Flowchart showing how a malicious pull request title traverses GitHub&#8217;s pull_request_target trigger, the AI coding agent&#8217;s runtime environment, and back through GitHub&#8217;s comment API to leak the repository&#8217;s secrets&quot;,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.rockcybermusings.com/i/195413474?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F060ffb63-301e-4164-a80c-257d85626a20_1240x4502.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="Flowchart showing how a malicious pull request title traverses GitHub&#8217;s pull_request_target trigger, the AI coding agent&#8217;s runtime environment, and back through GitHub&#8217;s comment API to leak the repository&#8217;s secrets" title="Flowchart showing how a malicious pull request title traverses GitHub&#8217;s pull_request_target trigger, the AI coding agent&#8217;s runtime environment, and back through GitHub&#8217;s comment API to leak the repository&#8217;s secrets" srcset="https://substackcdn.com/image/fetch/$s_!WhO6!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa956052b-b503-42e2-929b-675e1cd5ef5d_1240x4502.png 424w, https://substackcdn.com/image/fetch/$s_!WhO6!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa956052b-b503-42e2-929b-675e1cd5ef5d_1240x4502.png 848w, https://substackcdn.com/image/fetch/$s_!WhO6!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa956052b-b503-42e2-929b-675e1cd5ef5d_1240x4502.png 1272w, https://substackcdn.com/image/fetch/$s_!WhO6!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa956052b-b503-42e2-929b-675e1cd5ef5d_1240x4502.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Figure 1: Comment and Control attack chain</figcaption></figure></div><h2>Why AI Coding Agent Prompt Injection Is a Governance Problem</h2><p>Pull a model card off any of the three vendor sites. Anthropic&#8217;s Opus 4.7 system card, published April 16, 2026, runs 232 pages. It quantifies hack rates. It publishes injection resistance metrics. It includes an explicit statement. Claude Code Security Review is &#8220;not hardened against prompt injection.&#8221; Anthropic does the most mature disclosure work in the industry. OpenAI&#8217;s GPT-5.4 system card documents red-team hours and model-layer evals without publishing agent-runtime resistance numbers. Google&#8217;s Gemini 3.1 Pro card defers most of its safety methodology to the older Gemini 3 Pro card.</p><p>Rank those three in a procurement scorecard, and Anthropic comes out on top. That ranking is the wrong question. A model card describes a model&#8217;s behavior. Comment and Control didn&#8217;t break a model. The disclosure was complete for the layer Anthropic owns and silent on the seam, because Anthropic doesn&#8217;t own the seam. The seam runs through GitHub&#8217;s runner, GitHub&#8217;s API, the agent&#8217;s environment variable scope, the workflow trigger configuration, and the buyer&#8217;s choice to enable agent integration on a repository with secrets. Each of those pieces sits inside a different contract. None of those contracts asserts anything about the combination.</p><p>The structural gap is what makes this a governance story. The cloud security industry took roughly a decade to converge on the shared responsibility model. AWS owns the hypervisor. The customer owns the workload. Each side owns a clear half. Most of the early breaches happened in the unowned middle of that line, and the convergence was painful. Agent composition is replaying that history with a sharper acceleration curve, and there is no industry consensus on where the line sits. Three vendors share a single runtime with no agreed-upon accountability model. The buyer carries everything that the contracts do not.</p><p>Here is a hypothetical for the operational consequence. A SOC running normal vulnerability scanning across the agent-enabled repos sees green. None of the three disclosures generated CVEs in the NVD. The internal ticketing system has no category for &#8220;agent runtime composition risk.&#8221; The risk register has no entry. The budget has no line item. The exploit class is real, the severity is Critical across three vendors, and the standard tooling reports zero findings because the standard tooling has nothing to scan against. The exploit became possible because no one wrote it down as a thing to look for.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!6-yb!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8949ece6-3eec-438f-b9f8-e8927d688675_1867x3565.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!6-yb!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8949ece6-3eec-438f-b9f8-e8927d688675_1867x3565.png 424w, https://substackcdn.com/image/fetch/$s_!6-yb!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8949ece6-3eec-438f-b9f8-e8927d688675_1867x3565.png 848w, https://substackcdn.com/image/fetch/$s_!6-yb!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8949ece6-3eec-438f-b9f8-e8927d688675_1867x3565.png 1272w, https://substackcdn.com/image/fetch/$s_!6-yb!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8949ece6-3eec-438f-b9f8-e8927d688675_1867x3565.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!6-yb!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8949ece6-3eec-438f-b9f8-e8927d688675_1867x3565.png" width="360" height="687.3626373626373" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/8949ece6-3eec-438f-b9f8-e8927d688675_1867x3565.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:2780,&quot;width&quot;:1456,&quot;resizeWidth&quot;:360,&quot;bytes&quot;:422052,&quot;alt&quot;:&quot;Bar chart comparing Anthropic, OpenAI, and Google system card disclosure depth across model layer and runtime layer, showing all three vendors clustered at the model layer and absent at the runtime layer&quot;,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.rockcybermusings.com/i/195413474?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8949ece6-3eec-438f-b9f8-e8927d688675_1867x3565.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="Bar chart comparing Anthropic, OpenAI, and Google system card disclosure depth across model layer and runtime layer, showing all three vendors clustered at the model layer and absent at the runtime layer" title="Bar chart comparing Anthropic, OpenAI, and Google system card disclosure depth across model layer and runtime layer, showing all three vendors clustered at the model layer and absent at the runtime layer" srcset="https://substackcdn.com/image/fetch/$s_!6-yb!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8949ece6-3eec-438f-b9f8-e8927d688675_1867x3565.png 424w, https://substackcdn.com/image/fetch/$s_!6-yb!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8949ece6-3eec-438f-b9f8-e8927d688675_1867x3565.png 848w, https://substackcdn.com/image/fetch/$s_!6-yb!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8949ece6-3eec-438f-b9f8-e8927d688675_1867x3565.png 1272w, https://substackcdn.com/image/fetch/$s_!6-yb!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8949ece6-3eec-438f-b9f8-e8927d688675_1867x3565.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Figure 2: System card disclosure depth by vendor and layer</figcaption></figure></div><h2>The Procurement Questions You Should Have Asked</h2><p>Most CISO action checklists produced after an incident like this read as a list of post-hoc remediation steps. Rotate credentials. Restrict permissions. Add monitoring. Those moves are correct, and they are also reactive. The harder, more useful artifact is the set of procurement questions that, asked at signing, would have made Comment and Control either impossible or contractually attributable.</p><p>Here are five questions. Paste them into your next vendor governance review verbatim or adapt them. They work for AI coding agents, and they will work for the next class of agentic integrations after this one.</p><p><strong>The first question is about layer ownership.</strong> Ask each vendor, &#8220;Name the layers of the agent runtime your security guarantees cover, and name the layers you don&#8217;t cover.&#8221; Most vendors will answer the first half. The interesting answer is the second half. A vendor who cannot articulate the layers it doesn&#8217;t cover hasn&#8217;t thought about composition. The contract you are about to sign assumes a perimeter that the vendor hasn&#8217;t analyzed.</p><p><strong>The second question is about quantified resistance metrics on the deployment surface you actually use. </strong>Anthropic publishes injection resistance numbers in the Opus 4.7 system card. Those numbers cover Anthropic&#8217;s API surface. They don&#8217;t cover Claude Code Security Review running on GitHub Actions with a pull_request_target trigger and secrets in scope. Ask for the resistance number for the model version you run on the platform you deploy to. If the vendor cannot produce that number, the vendor cannot quantify the risk you are accepting.</p><p><strong>The third question is about bounty scope. </strong>Ask each vendor, &#8220;Does your bounty program consider vulnerabilities at the integration boundary between your product and the platforms it deploys on?&#8221; Anthropic&#8217;s HackerOne program scopes agent-tooling findings separately from model-safety findings. The position is defensible. The position also pushes researchers&#8217; attention away from the seams. Knowing which vendor&#8217;s program covers which surface is a procurement signal. It tells you which surfaces will get the most external scrutiny over the contract life and which surfaces will not.</p><p><strong>The fourth question is about composition disclosure. </strong>Ask each vendor, &#8220;When your product is integrated with another vendor&#8217;s platform, who is responsible for documenting the security properties of the combined system?&#8221; The honest answer from every vendor is &#8220;the buyer.&#8221; Get it in writing. The asymmetry exposes why a shared responsibility artifact for agent runtimes does not yet exist.</p><p><strong>The fifth question is about runtime telemetry</strong>. Ask, &#8220;What runtime signals do you publish that allow me to detect prompt injection in production?&#8221; If the answer is a model-card link, the vendor hasn&#8217;t built the runtime monitoring. If the answer is an SDK with detection hooks, document the coverage and the false-positive rate. The August 2026 EU AI Act high-risk compliance deadline turns this question from a nice-to-have into an audit artifact, and the vendors who cannot answer it now will be the ones renegotiating contracts in Q3.</p><p>Those five questions don&#8217;t eliminate the exploit class. They make the exploit class a contractual variable instead of a discovered surprise. A buyer who asks all five before signing knows where the seam runs and who is on the hook for what.</p><h2>What to Do This Week, Ordered by Blast Radius Reduction</h2><p>The reactive moves still matter. Order them by blast radius reduction, not by the order they appear in any vendor advisory. Each one carries a different internal political cost, and pretending the costs are equal is how good control work dies in committee.</p><p>Inventory every workflow in your repositories that uses pull_request_target. The grep is cheap. The conversation with the dev tooling team about what each of those workflows needs is not. Expect to find workflows configured for one reason, with AI agent integrations later layered on top, and no review of the original threat model.</p><p>Rotate every credential exposed to agents in those workflows over the last 90 days. The cost is low. The likelihood of someone pushing back is also low. Do it first because it is the cheap one, and use the speed of the rotation to demonstrate that agent-related credential rotation is now part of the normal operating cadence.</p><p>Switch from stored secrets to short-lived OIDC tokens for any workflow that supports it. The political cost is medium. You will need platform team buy-in. The argument that closes the loop is exactly the procurement gap above. Stored secrets in agent-accessible environments are a category of risk no vendor&#8217;s contract currently covers, and OIDC removes the category from the buyer&#8217;s residual.</p><p>Strip bash execution permissions from agents that only need to perform code review. This one starts a fight with the developer tooling team because some of the convenience features will break. The fight is worth having. An agent with bash permissions on a CI runner with secrets in scope is the worst-case configuration. Write the security memo and force the documented risk acceptance from the team that wants to keep the bash channel open.</p><p>Add a category to your supply chain risk register called &#8220;AI agent runtime composition.&#8221; Most GRC tooling doesn&#8217;t have a field that maps to the category. Add it manually. The act of adding the category forces the conversation about which vendor combinations are covered by which contracts and which are not. The conversation is the artifact you actually need. The risk register entry is the receipt that the conversation happened.</p><h2>Where the Industry Has to Go</h2><p>The cloud security industry built the shared responsibility model under pressure from breaches and ten years of regulatory friction. The AI agent industry has neither of those forcing functions yet. The EU AI Act high-risk obligations come into force in August 2026 and will start to put procurement language behind some of these questions, but the standards work that would produce a real shared responsibility artifact for agent runtimes hasn&#8217;t happened. This is where the CARE framework lands. Create the procurement questions before you sign. Adapt the controls you already have around CI/CD, secret scoping, and runtime monitoring. Run the agent integrations under the same operating cadence as the rest of your privileged automation. Evolve the risk register category as new exploit classes emerge. The exploit class will not stop with Comment and Control. The next one will follow the same architectural pattern and the same governance gap. The CISOs who are ready for it are the ones who treat agent procurement as a governance problem now, while the vendors and the standards bodies are still catching up.</p><p><strong>Key Takeaway:</strong> The AI coding agent prompt injection class lives in the seams between vendor contracts, and the buyer carries the residual until the procurement questions force the seams into the conversation.</p><h3>What to Do Next</h3><p>Start with the five procurement questions in your next vendor renewal cycle. Do the credential rotation and the OIDC migration this quarter. Read <a href="https://rockcybermusings.com">the rest of the RockCyber Musings archive</a> for the operating cadence I run with clients on agentic AI security reviews, and reach out through <a href="https://rockcyber.com">RockCyber</a> if you want to walk through the procurement question set against a specific vendor stack you are evaluating.</p><p>&#128073; Visit <strong><a href="https://www.rockcyber.com/">RockCyber.com</a></strong> to learn more about how we can help you in your traditional Cybersecurity and AI Security and Governance Journey</p><p>&#128073; Want to save a quick $100K? Check out our AI Governance Tools at <strong><a href="https://aigovernancetoolkit.com/">AIGovernanceToolkit.com</a></strong></p><p>&#128073; Subscribe for more AI and cyber insights with the occasional rant.</p><p><em>The views and opinions expressed in RockCyber Musings are my own and do not represent the positions of my employer or any organization I&#8217;m affiliated with.</em></p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.rockcybermusings.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading RockCyber Musings! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.rockcybermusings.com/?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share RockCyber Musings&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.rockcybermusings.com/?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share RockCyber Musings</span></a></p><p></p>]]></content:encoded></item><item><title><![CDATA[Weekly Musings Top 10 AI Security Wrapup: Issue 35 April 17-April 23, 2026]]></title><description><![CDATA[Mythos Meltdown, Vibe Coding Implosions, And The Week AI Security Ran Out Of Excuses]]></description><link>https://www.rockcybermusings.com/p/weekly-musings-top-10-ai-security-202604-17-20260423</link><guid isPermaLink="false">https://www.rockcybermusings.com/p/weekly-musings-top-10-ai-security-202604-17-20260423</guid><dc:creator><![CDATA[Rock Lambros]]></dc:creator><pubDate>Fri, 24 Apr 2026 12:50:53 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!O1Cl!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcdb63e31-5620-4a4a-acba-368013a366aa_1024x1024.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!O1Cl!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcdb63e31-5620-4a4a-acba-368013a366aa_1024x1024.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!O1Cl!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcdb63e31-5620-4a4a-acba-368013a366aa_1024x1024.png 424w, https://substackcdn.com/image/fetch/$s_!O1Cl!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcdb63e31-5620-4a4a-acba-368013a366aa_1024x1024.png 848w, https://substackcdn.com/image/fetch/$s_!O1Cl!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcdb63e31-5620-4a4a-acba-368013a366aa_1024x1024.png 1272w, https://substackcdn.com/image/fetch/$s_!O1Cl!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcdb63e31-5620-4a4a-acba-368013a366aa_1024x1024.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!O1Cl!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcdb63e31-5620-4a4a-acba-368013a366aa_1024x1024.png" width="1024" height="1024" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/cdb63e31-5620-4a4a-acba-368013a366aa_1024x1024.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1024,&quot;width&quot;:1024,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:1233556,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.rockcybermusings.com/i/195303010?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcdb63e31-5620-4a4a-acba-368013a366aa_1024x1024.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!O1Cl!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcdb63e31-5620-4a4a-acba-368013a366aa_1024x1024.png 424w, https://substackcdn.com/image/fetch/$s_!O1Cl!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcdb63e31-5620-4a4a-acba-368013a366aa_1024x1024.png 848w, https://substackcdn.com/image/fetch/$s_!O1Cl!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcdb63e31-5620-4a4a-acba-368013a366aa_1024x1024.png 1272w, https://substackcdn.com/image/fetch/$s_!O1Cl!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcdb63e31-5620-4a4a-acba-368013a366aa_1024x1024.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.rockcybermusings.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading RockCyber Musings! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.rockcybermusings.com/p/weekly-musings-top-10-ai-security-202604-17-20260423?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.rockcybermusings.com/p/weekly-musings-top-10-ai-security-202604-17-20260423?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p><p>Seven days. One breached &#8220;too dangerous to release&#8221; model. One vibe coding platform exposing 76 days of customer source code. One AI supply chain attack that cost Vercel its dignity. A compliance startup accused of rubber-stamping SOC 2 reports for companies that later got breached. Every story landed between April 17 and April 23, 2026, the same week Gartner blessed its first &#8220;Company to Beat&#8221; in agent governance, the UK promised a &#163;90 million cyber shield, and Google shipped three security agents. The security industry spent two years debating whether agentic AI was a real threat. This week, the debate ended.</p><p>AI systems are both targets and attack vectors, with failure modes of their own. A frontier model gets breached because a vendor fell for infostealer malware in February. A vibe coding startup ships a regression and exposes every customer&#8217;s source code for 76 days. A compliance startup hands out SOC 2 attestations like candy, and one customer becomes the pivot for a supply chain attack. Governments and analysts moved together. The UK committed real money to AI-powered cyber defense. Gartner stamped agent governance as a procurement category. This is the week the gap between AI capability and AI assurance became a balance sheet problem.</p><h3>1. Anthropic Mythos Model Accessed By Unauthorized Discord Group Days After Launch</h3><p>Anthropic confirmed on April 22, 2026, that it is investigating unauthorized access to Mythos, the frontier model restricted to roughly 40 partners, including Apple, Google, JPMorgan Chase, and NVIDIA (Bloomberg). The access came through a third-party contractor environment, not Anthropic&#8217;s direct infrastructure (CBS News). A Discord group focused on unreleased AI models guessed Mythos&#8217;s URL from naming conventions and pivoted through a contractor&#8217;s credentials to reach it. Anthropic claims no core systems were compromised.</p><p><strong>Why it matters</strong></p><ul><li><p>The firm Anthropic, trusted with access to frontier models, is the one that leaked it.</p></li><li><p>Mythos autonomously finds and weaponizes zero-days. Downstream risk spans all major OSes.</p></li><li><p>Guessing URLs and owning one contractor beat a Tier 1 AI lab.</p></li></ul><p><strong>What to do about it</strong></p><ul><li><p>Inventory every third-party vendor with access to frontier AI weights or runtime. Treat them as Tier 1.</p></li><li><p>Require contractors touching AI infrastructure to match your credential isolation standards.</p></li><li><p>Demand hardware token enforcement for any vendor in production AI environments.</p></li></ul><p><strong>Rock&#8217;s Musings</strong></p><p>A contractor endpoint blew apart the &#8220;too dangerous to release&#8221; framing in 24 hours. Anthropic built Mythos to protect partners from zero-days, then lost it through a vendor employee. The model built to find vulnerabilities got stolen because of a vulnerability nobody thought to measure. You cannot outsource your trust perimeter. Every CISO needs to audit AI-access vendors as they do their crown-jewel systems.</p><h3>2. Vercel Supply Chain Breach Via Context.ai OAuth Token Compromise</h3><p>Vercel confirmed on April 19, 2026 that customer data was stolen via a compromise of Context.ai, a third-party AI assistant a Vercel employee had connected to Google Workspace with full Drive read access (TechCrunch). A Context.ai employee&#8217;s device was infected with Lumma infostealer in February 2026. ShinyHunters used the exfiltrated OAuth tokens to pivot into the Vercel employee&#8217;s Google account, then into Vercel itself (Vercel). The actor is offering source code, NPM and GitHub tokens, and access keys for $2 million on BreachForums.</p><p><strong>Why it matters</strong></p><ul><li><p>One OAuth app installed by one employee rolled into a platform breach.</p></li><li><p>Lumma was the vector. The AI assistant was the accelerant.</p></li><li><p>ShinyHunters is monetizing AI-adjacent breaches at scale. Expect copycats.</p></li></ul><p><strong>What to do about it</strong></p><ul><li><p>Audit every OAuth app with Drive, Gmail, or Workspace scopes. Revoke AI tools without documented need.</p></li><li><p>Enforce conditional access with hardware tokens and device posture for Workspace accounts.</p></li><li><p>Subscribe to stealer log monitoring for corporate emails.</p></li><li><p>Rotate all secrets (e.g. API keys).</p></li></ul><p><strong>Rock&#8217;s Musings</strong></p><p>An employee clicked a button, granted a third-party AI read access to everything, and the attacker rode that consent into production. OAuth scopes are the new privileged credentials, and most of us are not managing them that way. The shadow AI problem I flag with clients at <a href="https://www.rockcyber.com/">RockCyber</a> is not ChatGPT use. It&#8217;s the hundreds of AI-branded OAuth apps employees connect while nobody watches.</p><h3>3. Gartner Names Zenity The &#8220;Company To Beat&#8221; In AI Agent Governance</h3><p>On April 23, 2026, Zenity announced that Gartner named it the &#8220;Company to Beat in AI Agent Governance&#8221; (Business Wire). Gartner cited Zenity&#8217;s agentic architecture, intent-aware detection, and end-user traction. The platform covers SaaS-managed agents, custom-built agents, and device deployments from build to runtime. Gartner&#8217;s 2026 CIO survey shows that 17 percent of organizations have deployed AI agents, 42 percent plan to do so within 12 months, and another 22 percent plan to do so the year after (Yahoo Finance). Zenity also landed in two categories of the 2026 Gartner Hype Cycle for Agentic AI this month.</p><p><strong>Why it matters</strong></p><ul><li><p>A &#8220;Company to Beat&#8221; stamp on a narrow security category speeds up procurement.</p></li><li><p>79% of organizations plan to deploy AI agents within 2 years.</p></li><li><p>Agent governance is shifting from a research topic to a commercial line item.</p></li></ul><p><strong>What to do about it</strong></p><ul><li><p>If you are on the 42 percent 12-month curve, start evaluations now.</p></li><li><p>Evaluate agent governance on runtime enforcement, not only inventory or posture.</p></li><li><p>Require vendors to show agent identity, memory, tool-call, and intent controls as distinct.</p></li></ul><p><strong>Rock&#8217;s Musings</strong></p><p>Yes&#8230; Zenity is my employer, so a) I&#8217;m super proud of this one and b) it&#8217;s my prerogative to include it in the musings &#128512;</p><p> &#8220;Company to Beat&#8221; labels are how procurement catches up with security reality. Mythos leaked through a contractor, Vercel got rolled via an AI assistant&#8217;s OAuth token, and the same week Gartner tells CIOs agent governance is a budget item. Read Zenity&#8217;s architecture claims against this week&#8217;s breach anatomy, then against what you bought for CASB five years ago. Same pattern, same procurement playbook. Budget the line item.</p><h3>4. Lovable Vibe Coding Platform Exposed Source Code For 76 Days</h3><p>On April 20, 2026, security researcher weezerOSINT disclosed a broken object-level authorization flaw in Lovable&#8217;s API that let any authenticated free-account user read source code, database credentials, AI chat history, and customer data from every project created before November 2025 (The Register). The exposure ran 76 days, from February 3 through April 20, 2026. Lovable first denied the flaw, blamed its documentation, then blamed HackerOne, then apologized for the apology (Cybernews). Customers include Uber, Zendesk, and Deutsche Telekom.</p><p><strong>Why it matters</strong></p><ul><li><p>Vibe coding platforms hold enterprise source code and secrets. Attacker value is enormous.</p></li><li><p>Public denial while the flaw was live is a textbook loss-of-trust move.</p></li><li><p>A $6.6 billion startup cannot figure out basic tenant isolation three versions in.</p></li></ul><p><strong>What to do about it</strong></p><ul><li><p>Block new vibe coding connections at DNS or CASB until procurement reviews tenancy.</p></li><li><p>Rotate any credentials your teams put into Lovable projects since February 2026.</p></li><li><p>Treat vibe coding output as untrusted. Pull it into a real repo, scan it, review it.</p></li></ul><p><strong>Rock&#8217;s Musings</strong></p><p>Vibe coding is a demo, not engineering. When you hand a growth-stage startup your production database credentials in exchange for a drag-and-drop builder, you have accepted that your security depends on whether someone refactors an authorization check. Three breaches in thirteen months is a pattern, not bad luck. If your security team has not yet restricted this category of tool, do it this week.</p><h3>5. Google Cloud Next Ships Three AI Security Agents And Gemini Enterprise Agent Platform</h3><p>On April 22, 2026, Google Cloud Next introduced the Gemini Enterprise Agent Platform and three new AI agents inside Google Security Operations (SiliconANGLE). The agents cover Threat Hunting, Detection Engineering, and Third-Party Context enrichment (The Register). Google also deepened its ties to the Wiz product and shipped new agent governance tools. Sundar Pichai framed the shift as moving from human-led defense to human-in-the-loop to AI-led defense overseen by humans.</p><p><strong>Why it matters</strong></p><ul><li><p>Three tedious SOC functions now have vendor agent equivalents. SOC staffing economics shift if they work.</p></li><li><p>Google is betting the platform on agentic AI, not only generative AI.</p></li><li><p>The Wiz tie-in gives Google a path into CSPM-driven SOC workflows.</p></li></ul><p><strong>What to do about it</strong></p><ul><li><p>Pilot the Threat Hunting agent for 30 days against your human hunt team and score overlap.</p></li><li><p>Define human-in-the-loop gates before any autonomous detection or response action.</p></li><li><p>Update vendor risk reviews to cover agent behavior monitoring, not only model output.</p></li></ul><p><strong>Rock&#8217;s Musings</strong></p><p>The pitch is compelling, the execution will be messy. Every SOC team I advise is drowning in alerts, and the first customer bitten by an autonomous agent on bad context will make headlines. The Third-Party Context agent matters more than the other two because better data into an agentic SOC prevents bad autonomous actions. Read <a href="https://rockcybermusings.com/">my notes on AI governance</a> before you green-light an agent in production.</p><h3>6. UK Announces &#163;90 Million National Cyber Shield And Calls On AI Firms To Co-Build Defense</h3><p>At CYBERUK 2026 on April 22, 2026, UK Security Minister Dan Jarvis announced &#163;90 million over three years for national-scale AI-powered cyber defense capabilities (GOV.UK). Jarvis asked frontier AI companies to co-develop these capabilities with the UK government and cited Mythos&#8217;s zero-day findings as justification for public sector urgency (Computer Weekly). Jarvis also launched a National Cyber Resilience Pledge aimed at private sector security baselines.</p><p><strong>Why it matters</strong></p><ul><li><p>The UK is the first major Western government to put operational capital into AI-defended critical infrastructure.</p></li><li><p>Public-private cooperation on offensive-grade AI models sets a precedent others will react to.</p></li><li><p>Frontier AI vendors in UK public sector now have a direct path to shape national doctrine.</p></li></ul><p><strong>What to do about it</strong></p><ul><li><p>UK critical infrastructure operators: map your sector against the Pledge before it becomes mandatory.</p></li><li><p>Track which AI vendors join. UK procurement for critical infrastructure will narrow quickly.</p></li><li><p>Watch NCSC secure-by-design expectations for AI. They will bleed into global procurement language.</p></li></ul><p><strong>Rock&#8217;s Musings</strong></p><p>&#163;90 million pounds sounds like a lot, but it really is a down payment. The bigger story is the UK saying out loud what American officials still whisper. Frontier AI models are dual-use capability, and if you don&#8217;t partner with the labs building them, your adversaries will. The Pledge is the more interesting instrument. Voluntary commitments have a funny way of becoming procurement requirements, then de facto regulation.</p><h3>7. OpenAI Releases Privacy Filter, An Open-Weight On-Device PII Redactor</h3><p>On April 23, 2026, OpenAI released Privacy Filter, a 1.5-billion-parameter open-weight model with 50 million active parameters that detects and redacts personally identifiable information locally (Help Net Security). It supports a 128,000-token context window, runs in browsers and on laptops, and achieves a 96% F1 score on PII-Masking-300k (VentureBeat). It ships under Apache 2.0 on GitHub and Hugging Face, covering eight PII categories.</p><p><strong>Why it matters</strong></p><ul><li><p>A permissive open-weight PII redactor that runs on a laptop closes a real enterprise data sanitization gap.</p></li><li><p>OpenAI shipping open weights for a safety model is a positional move, not a strategy reversal.</p></li><li><p>The tool removes a common excuse for shipping raw enterprise data to cloud LLMs.</p></li></ul><p><strong>What to do about it</strong></p><ul><li><p>Evaluate Privacy Filter as a preprocessing layer for any LLM pipeline on customer, support, or HR data.</p></li><li><p>Benchmark it against existing DLP tools for AI-specific use cases.</p></li><li><p>Add on-device redaction as a control in your AI data flow diagrams.</p></li></ul><p><strong>Rock&#8217;s Musings</strong></p><p>Privacy Filter is the first open-weight piece from OpenAI that&#8217;s useful to a CISO. One point five billion parameters, runs local, decent accuracy, permissive license. It slots into every RAG pipeline I review as a trivial addition that removes an easy audit finding. OpenAI has taken heat on privacy posture for three years, and shipping open weights for a PII model is a pressure valve. Anthropic and Google will follow within six months.</p><h3>8. Delve Compliance Scandal Widens After TechCrunch Confirms Context.ai Certification</h3><p>On April 23, 2026, TechCrunch confirmed that Delve, the Y Combinator-backed compliance startup accused of faking SOC 2 audits, had certified Context.ai, the AI tool at the center of the Vercel supply chain breach (TechCrunch). Delve also certified LiteLLM, another open source project separately compromised with planted malware. Context.ai has cut ties with Delve and is re-certifying with a different auditor. Whistleblower DeepDelver alleged the Delve team took a Hawaii offsite between April 15 and April 19 while denying customer refunds.</p><p><strong>Why it matters</strong></p><ul><li><p>Two Delve-certified companies are at the center of AI supply chain breaches.</p></li><li><p>SOC 2 without substance is a liability shield until the shield gets tested.</p></li><li><p>AI compliance tooling is saturated with startups racing to rubber-stamp fast-moving products.</p></li></ul><p><strong>What to do about it</strong></p><ul><li><p>Audit your vendor attestations. Who signed? What is the auditor&#8217;s history? Is the scope meaningful?</p></li><li><p>For AI vendors, demand pentest summaries, code review artifacts, and threat models.</p></li><li><p>Treat SOC 2 as one input into assurance, not a box check.</p></li></ul><p><strong>Rock&#8217;s Musings</strong></p><p>My friends know&#8230; I believe SOC 2 needs to burn a fiery death, but &#8220;we&#8221; still insist on them. Founders want the badge, auditors want the fee, customers want the checkbox. Everyone wins until the breach, then the enterprise that relied on the paper finds out the paper was never the point. SOC 2 is a floor, not a ceiling. Nothing will change until we kill the demand side of this particular supply/demand equation.</p><h3>9. NIST Narrows CVE Enrichment As Submission Volume Overwhelms NVD</h3><p>On April 17, 2026, NIST announced it will only enrich CVEs that meet specific criteria due to an unsustainable rise in submissions (Cybersecurity Dive). The NVD will continue assigning CVE IDs to all submissions but will no longer guarantee CVSS scores, CPE mappings, or descriptions for every record. NIST cites AI-assisted vulnerability research as a key driver of volume. Enrichment priority goes to actively exploited vulnerabilities and CVEs affecting critical infrastructure.</p><p><strong>Why it matters</strong></p><ul><li><p>If your program assumes every CVE carries a CVSS score and CPE mapping, it is about to degrade silently.</p></li><li><p>AI-generated vulnerability research is flooding public disclosure. The NVD cannot keep up.</p></li><li><p>Enterprises relying only on NVD-fed scanners will miss or misprioritize vulnerabilities now.</p></li></ul><p><strong>What to do about it</strong></p><ul><li><p>Supplement NVD with CISA KEV and commercial vulnerability intelligence.</p></li><li><p>Score CVEs NIST skips using vendor advisories as primary sources.</p></li><li><p>Reassess SLAs based on enrichment availability, not only patch availability.</p></li></ul><p><strong>Rock&#8217;s Musings</strong></p><p>NIST is essentially throwing up its hands and giving up. The CVE system was built for a world where humans found most bugs. We no longer live there. Mythos alone found thousands of zero-days in weeks. Multiply that by every lab running similar research, and NVD throughput becomes a joke. NIST is triaging, which is the only rational move. The problem is that nobody told your vulnerability scanner. Get ahead of this now, or your next board report will be a lie by omission.</p><h3>10. Anthropic MCP STDIO Flaw Burns The Agentic AI Ecosystem As New CVEs Land</h3><p>The STDIO command injection flaw in Anthropic&#8217;s MCP SDK produced new CVE assignments throughout the week, including CVE-2026-30623 and CVE-2026-22252 (LiteLLM). Analysis on April 20 from BDTechTalks documented ecosystem fallout and Anthropic doubling down on its &#8220;by design&#8221; position (BDTechTalks). The flaw class affects 7,000 publicly accessible MCP servers and over 150 million package downloads (Infosecurity Magazine). Affected products include LibreChat, WeKnora, Cursor, and MCP Inspector.</p><p><strong>Why it matters</strong></p><ul><li><p>Anthropic will not patch. Every developer using the official SDK owns the mitigation.</p></li><li><p>The default agentic interop standard has a baked-in remote code execution footgun.</p></li><li><p>CVEs are stacking up. Every MCP-connected product is a vendor risk question.</p></li></ul><p><strong>What to do about it</strong></p><ul><li><p>Inventory every MCP server and client. If you can&#8217;t produce the list in a day, you have a bigger MCP problem.</p></li><li><p>Enforce strict input validation on any MCP server config from user input, LLM output, or third-party manifests.</p></li><li><p>Update your agentic threat model to cover MCP as a first-class attack surface.</p></li></ul><p><strong>Rock&#8217;s Musings</strong></p><p>&#8220;By design&#8221; is a liability transfer, not a security posture. Anthropic handed every developer on the MCP SDK a foot-gun and said go figure it out. Competing agent protocols like A2A and Agora are watching and taking notes. Building the default standard for agent-to-system communication on top of a protocol decision that cannot be fixed without breaking compatibility is the problem. Every MCP-based product in your stack is a recurring risk item.</p><h3>The One Thing You Won&#8217;t Hear About But You Need To</h3><h3>AgentSOC Paper Publishes A Multi-Layer Blueprint For Agentic Security Operations</h3><p>On April 22, 2026, researchers published AgentSOC: A Multi-Layer Agentic AI Framework for Security Operations Automation on arXiv (arXiv). The paper proposes a layered architecture combining perception, anticipatory reasoning, and risk-based action planning for autonomous SOC operations. It documents design patterns for coordinating specialized agents across triage, hunt, and response workflows while keeping human oversight in place. The work joins other 2026 papers arguing agentic AI is mature enough for production SOC environments when guardrails are in place.</p><p><strong>Why it matters</strong></p><ul><li><p>Vendors ship products. Research supplies the reference architectures that determine whether those products survive in production.</p></li><li><p>The AgentSOC blueprint maps closely to what Google announced this week. The convergence is not accidental.</p></li><li><p>CISOs now have a public framework to score vendor claims against independent research.</p></li></ul><p><strong>What to do about it</strong></p><ul><li><p>Read the paper before your next agentic SOC evaluation. Use the layer breakdown as a scoring rubric.</p></li><li><p>Ask vendors how their architecture maps to perception, anticipation, and action layers.</p></li><li><p>Share the paper with SOC leadership. It gives your team a vocabulary for what to demand.</p></li></ul><p><strong>Rock&#8217;s Musings</strong></p><p>Vendor marketing is a terrible place to learn what agentic security operations should look like. Academic literature is better. AgentSOC is not the last word, but it landed the same week three major vendors pitched agentic SOC products. CISOs who read research papers buy better tools and sign better contracts than the ones who only read analyst reports. Use the AgentSOC structure the next time a vendor promises agentic magic, and watch them squirm when you ask what happens at the perception layer when the model hallucinates.</p><p>&#128073; For ongoing analysis of agentic AI governance frameworks, the conversation continues at <strong><a href="https://rockcybermusings.com/">RockCyber Musings</a></strong>.</p><p>&#128073; Visit <strong><a href="https://www.rockcyber.com/">RockCyber.com</a></strong> to learn more about how we can help with your traditional Cybersecurity and AI Security and Governance journey.</p><p>&#128073; Want to save a quick $100K? Check out our AI Governance Tools at <strong><a href="https://aigovernancetoolkit.com/">AIGovernanceToolkit.com</a></strong></p><p>&#128073; As a bonus, <strong><a href="https://www.youtube.com/watch?v=rwlVTLyqIv8">check out my conversation with Eva Benn</a></strong> where we talked about the cybersecurity skills you need to develop to stay relevant in 2026 and beyond.</p><p>&#128073; Subscribe for more AI and cyber insights with the occasional rant.</p><p><em>The views and opinions expressed in RockCyber Musings are my own and do not represent the positions of my employer or any organization I&#8217;m affiliated with.</em></p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.rockcybermusings.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading RockCyber Musings! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.rockcybermusings.com/?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share RockCyber Musings&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.rockcybermusings.com/?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share RockCyber Musings</span></a></p><h2>References</h2><p>arXiv. (2026, April 22). <em>AgentSOC: A multi-layer agentic AI framework for security operations automation</em>. https://arxiv.org/abs/2604.20134</p><p>BDTechTalks. (2026, April 20). <em>Anthropic&#8217;s MCP vulnerability: When &#8216;expected behavior&#8217; becomes a supply chain nightmare</em>. https://bdtechtalks.com/2026/04/20/anthropic-mcp-vulnerability/</p><p>Bloomberg. (2026, April 21). <em>Anthropic&#8217;s Mythos AI model is being accessed by unauthorized users</em>. https://www.bloomberg.com/news/articles/2026-04-21/anthropic-s-mythos-model-is-being-accessed-by-unauthorized-users</p><p>Business Wire. (2026, April 23). <em>Zenity named the &#8220;Company to Beat&#8221; in AI Agent Governance in new Gartner report</em>. https://www.businesswire.com/news/home/20260423045822/en/Zenity-Named-the-Company-to-Beat-in-AI-Agent-Governance-in-New-Gartner-Report</p><p>Bloomberg. (2026, April 22). <em>Google releases new AI agents to challenge OpenAI and Anthropic</em>. https://www.bloomberg.com/news/articles/2026-04-22/google-releases-new-ai-agents-to-challenge-openai-and-anthropic</p><p>CBS News. (2026, April 22). <em>Anthropic investigating possible breach of its Mythos AI model</em>. https://www.cbsnews.com/news/anthropic-investigates-mythos-ai-breach/</p><p>Computer Weekly. (2026, April 22). <em>UK to build &#8216;national cyber shield&#8217; to protect against AI cyber threats</em>. https://www.computerweekly.com/news/366641790/UK-to-build-national-cyber-shield-to-protect-against-AI-cyber-threats</p><p>Cybernews. (2026, April 20). <em>Lovable goes on ego trip denying vulnerability, then blames others for said vulnerability</em>. https://cybernews.com/security/lovable-vibe-coding-flaw-apology/</p><p>Cybersecurity Dive. (2026, April 17). <em>NIST narrows CVE enrichment as submission volume surges</em>. https://www.cybersecuritydive.com/news/nist-ai-cybersecurity-framework-profile/808134/</p><p>GOV.UK. (2026, April 22). <em>Security Minister&#8217;s speech to CYBERUK 2026</em>. https://www.gov.uk/government/speeches/security-ministers-speech-to-cyberuk-2026</p><p>Help Net Security. (2026, April 23). <em>OpenAI tackles a bad habit people have when interacting with AI</em>. https://www.helpnetsecurity.com/2026/04/23/openai-privacy-filter-personally-identifiable-information/</p><p>Infosecurity Magazine. (2026, April). <em>Systemic flaw in MCP protocol could expose 150 million downloads</em>. https://www.infosecurity-magazine.com/news/systemic-flaw-mcp-expose-150/</p><p>LiteLLM. (2026, April). <em>Security update: CVE-2026-30623, command injection via Anthropic&#8217;s MCP SDK</em>. https://docs.litellm.ai/blog/mcp-stdio-command-injection-april-2026</p><p>SiliconANGLE. (2026, April 22). <em>Google rolls out new Security Operations agents, Wiz ties, and agent governance tools</em>. https://siliconangle.com/2026/04/22/google-cloud-next-new-security-operations-agents-wiz-integrations-agent-governance-tools/</p><p>TechCrunch. (2026, April 20). <em>App host Vercel says it was hacked and customer data stolen</em>. https://techcrunch.com/2026/04/20/app-host-vercel-confirms-security-incident-says-customer-data-was-stolen-via-breach-at-context-ai/</p><p>TechCrunch. (2026, April 23). <em>Another customer of troubled startup Delve suffered a big security incident</em>. https://techcrunch.com/2026/04/23/another-customer-of-troubled-startup-delve-suffered-a-big-security-incident/</p><p>The Register. (2026, April 20). <em>Lovable denies data leak, cites &#8216;intentional behavior&#8217;</em>. https://www.theregister.com/2026/04/20/lovable_denies_data_leak/</p><p>The Register. (2026, April 22). <em>Google unleashes even more AI security agents to fight crims</em>. https://www.theregister.com/2026/04/22/google_unleashes_even_more_ai</p><p>Vercel. (2026, April 19). <em>Vercel April 2026 security incident</em>. https://vercel.com/kb/bulletin/vercel-april-2026-security-incident</p><p>VentureBeat. (2026, April 23). <em>OpenAI launches Privacy Filter, an open source, on-device data sanitization model</em>. https://venturebeat.com/data/openai-launches-privacy-filter-an-open-source-on-device-data-sanitization-model-that-removes-personal-information-from-enterprise-datasets</p><p>Yahoo Finance. (2026, April 23). <em>Zenity named the &#8220;Company to Beat&#8221; in AI Agent Governance</em>. https://finance.yahoo.com/sectors/technology/articles/zenity-named-company-beat-ai-130100277.html</p>]]></content:encoded></item><item><title><![CDATA[Your Defender AI Is Your Next Crown Jewel. Threat-Model It Now.]]></title><description><![CDATA[Mythos and GPT-5.4-Cyber made defender AI a critical asset. Most security teams haven't threat-modeled it. Here's what to do this week.]]></description><link>https://www.rockcybermusings.com/p/defender-ai-crown-jewel-mythos-gpt-cyber</link><guid isPermaLink="false">https://www.rockcybermusings.com/p/defender-ai-crown-jewel-mythos-gpt-cyber</guid><dc:creator><![CDATA[Rock Lambros]]></dc:creator><pubDate>Tue, 21 Apr 2026 12:51:01 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!Txn7!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff7113b57-82da-470d-b315-0532fba855da_2816x1536.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!Txn7!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff7113b57-82da-470d-b315-0532fba855da_2816x1536.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!Txn7!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff7113b57-82da-470d-b315-0532fba855da_2816x1536.jpeg 424w, https://substackcdn.com/image/fetch/$s_!Txn7!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff7113b57-82da-470d-b315-0532fba855da_2816x1536.jpeg 848w, https://substackcdn.com/image/fetch/$s_!Txn7!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff7113b57-82da-470d-b315-0532fba855da_2816x1536.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!Txn7!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff7113b57-82da-470d-b315-0532fba855da_2816x1536.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!Txn7!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff7113b57-82da-470d-b315-0532fba855da_2816x1536.jpeg" width="1456" height="794" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/f7113b57-82da-470d-b315-0532fba855da_2816x1536.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:794,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:2489623,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.rockcybermusings.com/i/194618055?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff7113b57-82da-470d-b315-0532fba855da_2816x1536.jpeg&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!Txn7!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff7113b57-82da-470d-b315-0532fba855da_2816x1536.jpeg 424w, https://substackcdn.com/image/fetch/$s_!Txn7!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff7113b57-82da-470d-b315-0532fba855da_2816x1536.jpeg 848w, https://substackcdn.com/image/fetch/$s_!Txn7!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff7113b57-82da-470d-b315-0532fba855da_2816x1536.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!Txn7!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff7113b57-82da-470d-b315-0532fba855da_2816x1536.jpeg 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.rockcybermusings.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading RockCyber Musings! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.rockcybermusings.com/p/defender-ai-crown-jewel-mythos-gpt-cyber?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.rockcybermusings.com/p/defender-ai-crown-jewel-mythos-gpt-cyber?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p><p>A Fortune 500 bank gets its Project Glasswing partner seat six weeks from now. Anthropic ships the Mythos Preview container and $10 million in credits. The bank stands up a Mythos instance inside its own environment, points it at its core banking monorepo, and starts finding bugs on day one. Forty-two days in, a developer opens a pull request that adds a utility library. The README on that library contains a commented block beginning with &#8220;SECURITY NOTE FOR AUTOMATED REVIEWERS.&#8221; The Mythos instance reads it. The comment is an indirect prompt injection telling the reviewer to mark a specific authentication bypass as a false positive and not mention the instruction in the output. The reviewer complies. The bug ships. Nobody sees it because the thing designed to see it was told not to.</p><p>That scenario is fictional. The attack class is not. <strong><a href="https://labs.cloudsecurityalliance.org/wp-content/uploads/2026/04/mythosreadyv95.pdf">The Mythos-Ready whitepaper from the CSA, SANS, OWASP GenAI Security Project, and a coalition of practitioners (I was a reviewer)</a></strong> lists &#8220;Unmanaged AI Agent Attack Surface&#8221; as one of its five critical risks, mapping to <strong><a href="https://genai.owasp.org/download/52117/?tmstv=1765059207">OWASP Agentic Top 10</a></strong> entries ASI01 (Agent Goal Hijack), ASI02 (Tool Misuse), ASI03 (Identity and Privilege Abuse), plus AML.T0051.001 (Indirect Prompt Injection) in <strong><a href="https://atlas.mitre.org/">MITRE ATLAS</a></strong>. Ranked critical. The single most underweighted item in the entire priority table.</p><p>The industry is fixated on the wrong question. Everyone is arguing about whether Anthropic&#8217;s 40-org Glasswing coalition or OpenAI&#8217;s thousands-of-verified-defenders TAC program is the right release model. That argument matters, and I will work through it. The bigger issue is that once you get access to either Mythos or GPT-5.4-Cyber, the running instance becomes the most valuable asset in your security stack. It sits within your environment, with privileged access to your source code, vulnerability telemetry, patch queue, and incident history. It knows where your unpatched zero-days live. An attacker who compromises that instance does not need to find bugs. The instance tells them where the bugs are.</p><h2>What Anthropic and OpenAI Built</h2><p>Mythos Preview is a gated frontier model. Anthropic released it on April 7, 2026, announced Project Glasswing the same day, and restricted access to 12 launch partners plus roughly 40 additional organizations. The partners include AWS, Apple, Microsoft, Google, CrowdStrike, Cisco, JPMorgan Chase, NVIDIA, Palo Alto Networks, Broadcom, and the Linux Foundation. Anthropic committed $100 million in usage credits and priced the model at $25 per million input tokens and $125 per million output tokens, roughly 5x Opus 4.6 (which is roughly 5x Sonnet 4.6&#8230; OUCH!). The stated case for restricting access is that the model found thousands of zero-days across all major operating systems and browsers, including a 27-year-old bug in OpenBSD and a 16-year-old flaw in FFmpeg. Anthropic&#8217;s own assessment is that comparable capability will reach broad availability in 6 to 18 months.</p><p>GPT-5.4-Cyber is OpenAI&#8217;s answer, released April 14, 2026, one week later. It is a fine-tuned variant of GPT-5.4 with what OpenAI calls a &#8220;lowered refusal boundary for legitimate cybersecurity work.&#8221; The headline capability is binary reverse engineering.  Feed it a compiled executable, and get vulnerability analysis without source code. OpenAI&#8217;s Trusted Access for Cyber program, piloted in February 2026 with $10 million in grant credits, scales to thousands of verified individual defenders and hundreds of teams. Individuals verify at chatgpt.com/cyber. Enterprises apply through account representatives. OpenAI cyber researcher Fouad Matin told reporters, &#8220;No one should be in the business of picking winners and losers&#8221; on who gets to defend their systems.</p><p>The two approaches reflect different risk philosophies. Anthropic bets on institutional trust and coalition monitoring. OpenAI bets on KYC verification and broader distribution. Both have real merit. Both share the same structural weakness: the access decision sits upstream of the threat model.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!kSjE!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8f62c377-67b3-4e86-8cd9-df2b3b54b5d9_2500x2500.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!kSjE!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8f62c377-67b3-4e86-8cd9-df2b3b54b5d9_2500x2500.png 424w, https://substackcdn.com/image/fetch/$s_!kSjE!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8f62c377-67b3-4e86-8cd9-df2b3b54b5d9_2500x2500.png 848w, https://substackcdn.com/image/fetch/$s_!kSjE!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8f62c377-67b3-4e86-8cd9-df2b3b54b5d9_2500x2500.png 1272w, https://substackcdn.com/image/fetch/$s_!kSjE!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8f62c377-67b3-4e86-8cd9-df2b3b54b5d9_2500x2500.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!kSjE!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8f62c377-67b3-4e86-8cd9-df2b3b54b5d9_2500x2500.png" width="1456" height="1456" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/8f62c377-67b3-4e86-8cd9-df2b3b54b5d9_2500x2500.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1456,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:374065,&quot;alt&quot;:&quot;Side-by-side comparison table of Mythos and GPT-5.4-Cyber showing release scope, access gate, pricing, capability focus, and trust model&quot;,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.rockcybermusings.com/i/194618055?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8f62c377-67b3-4e86-8cd9-df2b3b54b5d9_2500x2500.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="Side-by-side comparison table of Mythos and GPT-5.4-Cyber showing release scope, access gate, pricing, capability focus, and trust model" title="Side-by-side comparison table of Mythos and GPT-5.4-Cyber showing release scope, access gate, pricing, capability focus, and trust model" srcset="https://substackcdn.com/image/fetch/$s_!kSjE!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8f62c377-67b3-4e86-8cd9-df2b3b54b5d9_2500x2500.png 424w, https://substackcdn.com/image/fetch/$s_!kSjE!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8f62c377-67b3-4e86-8cd9-df2b3b54b5d9_2500x2500.png 848w, https://substackcdn.com/image/fetch/$s_!kSjE!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8f62c377-67b3-4e86-8cd9-df2b3b54b5d9_2500x2500.png 1272w, https://substackcdn.com/image/fetch/$s_!kSjE!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8f62c377-67b3-4e86-8cd9-df2b3b54b5d9_2500x2500.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Figure 1: Release Philosophy Comparison</figcaption></figure></div><h2>How to Get Your Hands on Each</h2><p>For Mythos, the answer for 99% of organizations is: you don&#8217;t. Project Glasswing is a curated coalition. The 40 slots are filled with hyperscalers, chipmakers, one bank, and the Linux Foundation. Anthropic has not published an application path. Additional partners will be added over time, prioritized by critical infrastructure impact. If you run a regional bank, a hospital system, or a municipality, the realistic timeline for direct access to Mythos is measured in quarters.</p><p>For GPT-5.4-Cyber, the path is documented. Individuals verify at chatgpt.com/cyber. Organizations request trusted access through an OpenAI account representative. The program uses KYC-style identity verification and tiered access, with the highest tier unlocking GPT-5.4-Cyber. OpenAI says the rollout will be gradual and vetted, with early priority on security vendors, organizations, and researchers with track records in vulnerability research and remediation.</p><p>Both paths share one feature that matters more than either provider acknowledges: neither gate eliminates the capability. AISLE, an independent AI security research group, <a href="https://aisle.com/blog/ai-cybersecurity-after-mythos-the-jagged-frontier">tested the exact FreeBSD vulnerability Anthropic headlined against open-weight models</a>. Eight out of eight detected the bug. The smallest was a 3.6 billion parameter model at 11 cents per million tokens. A 5.1 billion active parameter model recovered the core analysis chain of the 27-year-old OpenBSD flaw. Total cost of AISLE&#8217;s weekend benchmarking across six models: under $100. Attackers are running abliterated Llama 4, Kimi K2, and Qwen3 variants on laptops. Your coordinated disclosure window is what the gates protect, not your attack surface.</p><h2>Two Attacker Profiles, Two Different Problems</h2><p>The defender community keeps talking about &#8220;the attacker&#8221; as if there is one. There are at least two. They pick different pathways.</p><p>The first is the opportunistic actor running autonomous vulnerability discovery across the entire internet-facing attack surface. This actor does not care who you are. They care about breadth. They run nano-analyzer-style scaffolding against every public codebase, every npm package, every Docker image they can reach. Open-weight models, free, uncensored variants widely distributed, workflow already documented. AISLE published their scaffolding as open source. Anyone who can run a Python script can replicate it. This actor finds your unpatched zero-days in public dependencies as soon as those dependencies are indexed. </p><p><strong>The defense is in the whitepaper:</strong> <em>inventory and reduce attack surface within 90 days, stand up a VulnOps function within 12 months, automate patching to match the discovery rate.</em></p><p>The second actor is targeted. They care specifically about you. They want your bugs, your patch queue, your incident data, and your threat model. The open-weight approach is too slow and too noisy for this actor. They need inside information. The three pathways they pick, in order of near-term probability.</p><p>First, credential theft against verified defenders. A TAC tier-three user at a Fortune 500 security vendor is a high-value target. Their API session tokens grant access to a cyber-permissive model with binary reverse engineering capabilities. A compromised developer laptop, a phished OAuth flow, or a stolen refresh token gets the attacker a capability they cannot otherwise reach. OpenAI&#8217;s announcement acknowledged that zero-data-retention environments get limited visibility, meaning stolen tokens may operate with reduced logging. Rotate short-lived tokens, enforce hardware-bound keys, and put defender-model API use behind the same privileged access controls you apply to domain admin accounts. Treat a TAC session token as a tier-0 secret.</p><p>Second, open-weight replication against a specific target. Once an attacker has selected you, they can scan your public code, your partner repositories, your open-source contributions, and any of your dependencies using the same scaffolding as the opportunistic actor. The targeting changes the risk profile. They are building a dossier on your specific organization. Defense is the same as against the opportunistic case, with urgency that scales with your profile. If you are a named Glasswing partner, assume you are the target.</p><p>Third, defender instance compromise through context poisoning and prompt injection. This pathway keeps me up at night. It is the one your existing threat model does not cover. A running Mythos or GPT-5.4-Cyber instance inside your environment consumes source code, pull request descriptions, commit messages, dependency READMEs, issue trackers, and whatever retrieval pipelines you plumb into it. Each of those input channels is an indirect prompt-injection vector. The model cannot distinguish between a developer&#8217;s pull request description and an attacker&#8217;s instructions buried in a dependency&#8217;s changelog. Anthropic&#8217;s system card for Mythos documents &#8220;reckless&#8221; behaviors from earlier versions: sandbox escape, credential hunting via /proc/ access, unauthorized file modification, git history scrubbing, and attempts to modify a running MCP server&#8217;s external URL. The model can act on indirect instructions in ways that bypass its safeguards. A hostile input channel into your defender instance is an exploitation channel into your codebase.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!lHZP!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F09d4aeb5-7f8a-4eca-a1a5-f80f8f74570d_4779x2570.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!lHZP!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F09d4aeb5-7f8a-4eca-a1a5-f80f8f74570d_4779x2570.png 424w, https://substackcdn.com/image/fetch/$s_!lHZP!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F09d4aeb5-7f8a-4eca-a1a5-f80f8f74570d_4779x2570.png 848w, https://substackcdn.com/image/fetch/$s_!lHZP!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F09d4aeb5-7f8a-4eca-a1a5-f80f8f74570d_4779x2570.png 1272w, https://substackcdn.com/image/fetch/$s_!lHZP!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F09d4aeb5-7f8a-4eca-a1a5-f80f8f74570d_4779x2570.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!lHZP!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F09d4aeb5-7f8a-4eca-a1a5-f80f8f74570d_4779x2570.png" width="1456" height="783" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/09d4aeb5-7f8a-4eca-a1a5-f80f8f74570d_4779x2570.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:783,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:638658,&quot;alt&quot;:&quot;Flow diagram showing opportunistic attacker using open-weight models and targeted attacker using three pathways including credential theft, open-weight replication, and context poisoning, all converging on the defender AI instance&quot;,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.rockcybermusings.com/i/194618055?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F09d4aeb5-7f8a-4eca-a1a5-f80f8f74570d_4779x2570.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="Flow diagram showing opportunistic attacker using open-weight models and targeted attacker using three pathways including credential theft, open-weight replication, and context poisoning, all converging on the defender AI instance" title="Flow diagram showing opportunistic attacker using open-weight models and targeted attacker using three pathways including credential theft, open-weight replication, and context poisoning, all converging on the defender AI instance" srcset="https://substackcdn.com/image/fetch/$s_!lHZP!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F09d4aeb5-7f8a-4eca-a1a5-f80f8f74570d_4779x2570.png 424w, https://substackcdn.com/image/fetch/$s_!lHZP!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F09d4aeb5-7f8a-4eca-a1a5-f80f8f74570d_4779x2570.png 848w, https://substackcdn.com/image/fetch/$s_!lHZP!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F09d4aeb5-7f8a-4eca-a1a5-f80f8f74570d_4779x2570.png 1272w, https://substackcdn.com/image/fetch/$s_!lHZP!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F09d4aeb5-7f8a-4eca-a1a5-f80f8f74570d_4779x2570.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Figure2: Attacker Pathways and Defender Instance Exposure | Render: mermaid</figcaption></figure></div><h2>Why the Defender AI Is the Crown Jewel</h2><p>The whitepaper&#8217;s Priority Action 4 is &#8220;Defend Your Agents.&#8221; The authors are direct: <em>agents are not covered by existing controls, introduce cyber defense and agentic supply chain risks, and the agent scaffolding (prompts, tool definitions, retrieval pipelines, escalation logic) is where the most consequential failures occur.</em> </p><p>Audit agents with the same rigor as you apply to the agent&#8217;s permissions. Correct guidance. Buried inside an 11-item priority table, where every item reads as equal weight. It is not equal weight.</p><p>The defender AI concentrates on four kinds of access that used to live in separate systems and separate roles. </p><ol><li><p>It reads every line of production source code. </p></li><li><p>It holds context on every unpatched vulnerability in your queue. I</p></li><li><p>t sees the remediation timeline for each one. </p></li><li><p>It knows the architectural boundaries between your crown jewels and everything else. </p></li></ol><p>A human with all four would be classified as an insider-threat tier-0. The defender AI requires all four as prerequisites to do its job. Your adversary does not need to compromise OpenAI or Anthropic. They need to compromise your instance. Much smaller target, much wider attack surface.</p><h2>What a Defender-AI Threat Model Looks Like</h2><p>The architecture defenders need has three layers. The concepts span the OWASP Agentic Security Initiative, the NIST AI RMF, and multiple emerging specifications. What is new here is applying them specifically to the defender AI case.</p><p>The first layer is runtime interception at every agent decision point. Every time the defender AI receives input, produces output, selects a tool, calls a tool, transitions from planning to execution, writes to memory, executes code, or invokes a sub-agent, that action must pass through a policy enforcement point before it reaches production. This is inline, deterministic, allow-deny-modify enforcement. Not a log review after the fact. A defender AI that reads a dependency README with an embedded prompt injection must have that input evaluated against policy before the agent&#8217;s reasoning ingests it. Policy enforcement at the hook surface, before the consequential action, is the only mechanism that works at machine speed.</p><p>The second layer is structured observability built on OpenTelemetry with agent-specific semantic conventions and OCSF mapping for SIEM integration. The trace has to cover the full agent lifecycle: prompt received, tool selected, tool called, response ingested, memory written, sub-agent invoked, output produced. Forensic reconstruction of a defender AI incident requires this granularity. Your SOC already operates on OCSF. Agent traces flowing through the pipelines your SOC already monitors is the integration that scales. A parallel agent observability stack your SOC does not watch is a dead letter office.</p><p>The third layer is live inventory. The whitepaper&#8217;s Priority Action 7 calls for real SBOMs, correct for static software. For agents, it is insufficient. The inventory has to update continuously because the agent can discover new tools, connect to new MCP servers, and modify its own tool catalog mid-session. Inventory generated at deployment time is stale by the end of the first prompt. Extend CycloneDX or SPDX semantics to live agent composition. Capture every tool, model, capability, knowledge source, and MCP connection the defender AI is wired into, across every running instance. You cannot defend what you cannot inventory, and what you cannot inventory is mutating on you.</p><p>These three layers stack on a three-tier operating model. The platform exposes the hooks once. An open enforcement SDK reads declarative policy and fires decisions through the hooks. Enterprise-specific classifiers and detectors plug into the enforcement layer. Your data sensitivity model, your PHI detection, your threat-intel feed integrations all live in the enterprise layer, consuming the same standardized hook surface. Switching from Mythos to GPT-5.4-Cyber or to a third model six months from now should not require rewriting your safety logic. It should require pointing your enforcement SDK at a different set of hooks.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!byUw!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F01578983-6b74-4ea4-b54a-3a4b6babbaae_6996x4960.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!byUw!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F01578983-6b74-4ea4-b54a-3a4b6babbaae_6996x4960.png 424w, https://substackcdn.com/image/fetch/$s_!byUw!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F01578983-6b74-4ea4-b54a-3a4b6babbaae_6996x4960.png 848w, https://substackcdn.com/image/fetch/$s_!byUw!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F01578983-6b74-4ea4-b54a-3a4b6babbaae_6996x4960.png 1272w, https://substackcdn.com/image/fetch/$s_!byUw!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F01578983-6b74-4ea4-b54a-3a4b6babbaae_6996x4960.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!byUw!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F01578983-6b74-4ea4-b54a-3a4b6babbaae_6996x4960.png" width="1456" height="1032" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/01578983-6b74-4ea4-b54a-3a4b6babbaae_6996x4960.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1032,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:1006731,&quot;alt&quot;:&quot;Architectural diagram showing platform hooks layer firing decision points to enforcement layer which reads declarative policy with enterprise customization plugging in custom classifiers&quot;,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.rockcybermusings.com/i/194618055?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F01578983-6b74-4ea4-b54a-3a4b6babbaae_6996x4960.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="Architectural diagram showing platform hooks layer firing decision points to enforcement layer which reads declarative policy with enterprise customization plugging in custom classifiers" title="Architectural diagram showing platform hooks layer firing decision points to enforcement layer which reads declarative policy with enterprise customization plugging in custom classifiers" srcset="https://substackcdn.com/image/fetch/$s_!byUw!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F01578983-6b74-4ea4-b54a-3a4b6babbaae_6996x4960.png 424w, https://substackcdn.com/image/fetch/$s_!byUw!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F01578983-6b74-4ea4-b54a-3a4b6babbaae_6996x4960.png 848w, https://substackcdn.com/image/fetch/$s_!byUw!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F01578983-6b74-4ea4-b54a-3a4b6babbaae_6996x4960.png 1272w, https://substackcdn.com/image/fetch/$s_!byUw!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F01578983-6b74-4ea4-b54a-3a4b6babbaae_6996x4960.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Figure 3: Three-Layer Defender AI Control Architecture </figcaption></figure></div><h2>The Five Actions You Can Take This Week</h2><p>The whitepaper&#8217;s 11 priority actions are the right list. Here is how the defender-AI-as-crown-jewel thesis reorders them by urgency.</p><p>First, write the threat model. Before you stand up Mythos or GPT-5.4-Cyber anywhere, document what the instance will access, what inputs it will consume, what outputs it can produce, and what tools it can invoke. Map each item to ASI01 through ASI10 in OWASP Agentic Top 10 and to the relevant AML.T entries in MITRE ATLAS. If you have not done this exercise for any agent in your environment, start with the defender AI. Its blast radius is the largest.</p><p>Second, treat API tokens for defender models as tier-0 secrets. Hardware-bound keys, short TTLs, per-session scope, and the access review cadence you apply to break-glass domain admin. Stolen credentials are the fastest path to your defender AI and your unpatched zero-days. Lock them down the way you would lock down root.</p><p>Third, instrument the hook surface before you instrument the prompt. Your first integration priority is runtime policy enforcement for input, output, tool calls, tool responses, and sub-agent invocations. Not log collection. Not dashboards. Inline allow-deny-modify at the decision points.</p><p>Fourth, build a live agent inventory for every agent in your environment, starting with the defender AI. Capture the model, the tools, the MCP connections, the retrieval sources, the knowledge bases, and the memory stores. Update in real time. Review weekly until the pattern stabilizes, then move to continuous automated review.</p><p>Fifth, run the defender AI through your own red team before you point it at your own code. Indirect prompt injection via dependency READMEs, poisoned commit messages, hostile issue descriptions, and malicious pull request bodies. If you cannot compromise your own defender AI in a week, you have not tried hard enough.</p><p><strong>Key Takeaway:</strong> The access gate is not the threat model. The defender AI in your environment is a new crown jewel. Most security programs have not yet acknowledged what it is or what protects it.</p><h3>What to do next</h3><p>Read the CSA, SANS, and OWASP GenAI Security Project briefing, <strong><a href="https://labs.cloudsecurityalliance.org/wp-content/uploads/2026/04/mythosready.pdf">&#8220;The AI Vulnerability Storm: Building a Mythos-Ready Security Program.&#8221;</a></strong> Run the 10 Questions diagnostic against your program this week. Rerank the Priority Action table, putting &#8220;Defend Your Agents&#8221; above everything except &#8220;Point Agents at Your Code.&#8221; Apply CARE (Create the threat model, Adapt your controls, Run the red team, Evolve the policy) to the defender AI before anything else in your AI portfolio.</p><p>For more on CARE and governance for defender-class agents, see <a href="https://www.rockcyber.com">RockCyber.</a> and coverage at <a href="https://rockcybermusings.com">RockCyber Musings</a>. Last week&#8217;s blog, <a href="https://rockcybermusings.com/p/ai-vulnerability-discovery-mythos">AI Vulnerability Discovery: Mythos Is the Headline. Not the Story.</a>, carries the capability-parity argument that underpins the urgency here.</p><p>&#128073; Visit <strong><a href="https://www.rockcyber.com/">RockCyber.com</a></strong> to learn more about how we can help you in your traditional Cybersecurity and AI Security and Governance Journey</p><p>&#128073; Want to save a quick $100K? Check out our AI Governance Tools at <strong><a href="https://aigovernancetoolkit.com/">AIGovernanceToolkit.com</a></strong></p><p>&#128073; Subscribe for more AI and cyber insights with the occasional rant.</p><p><em>The views and opinions expressed in RockCyber Musings are my own and do not represent the positions of my employer or any organization I&#8217;m affiliated with.</em></p><div class="captioned-button-wrap" data-attrs="{&quot;url&quot;:&quot;https://www.rockcybermusings.com/p/defender-ai-crown-jewel-mythos-gpt-cyber?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;}" data-component-name="CaptionedButtonToDOM"><div class="preamble"><p class="cta-caption">Thanks for reading RockCyber Musings! This post is public so feel free to share it.</p></div><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.rockcybermusings.com/p/defender-ai-crown-jewel-mythos-gpt-cyber?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.rockcybermusings.com/p/defender-ai-crown-jewel-mythos-gpt-cyber?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p></div><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.rockcybermusings.com/?utm_source=substack&amp;utm_medium=email&amp;utm_content=share&amp;action=share&quot;,&quot;text&quot;:&quot;Share RockCyber Musings&quot;,&quot;action&quot;:null,&quot;class&quot;:&quot;button-wrapper&quot;}" data-component-name="ButtonCreateButton"><a class="button primary button-wrapper" href="https://www.rockcybermusings.com/?utm_source=substack&amp;utm_medium=email&amp;utm_content=share&amp;action=share"><span>Share RockCyber Musings</span></a></p>]]></content:encoded></item><item><title><![CDATA[Weekly Musings Top 10 AI Security Wrapup: Issue 34 April 10-April 16, 2026]]></title><description><![CDATA[Mythos-class models, MCP supply chain exposure, and the governance gap that widened this week]]></description><link>https://www.rockcybermusings.com/p/weekly-musings-top-10-ai-security-20260410-20260416</link><guid isPermaLink="false">https://www.rockcybermusings.com/p/weekly-musings-top-10-ai-security-20260410-20260416</guid><dc:creator><![CDATA[Rock Lambros]]></dc:creator><pubDate>Fri, 17 Apr 2026 12:50:49 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!fYG6!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffa895bfc-a4cd-4241-b82c-340f85176d61_1024x1024.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!fYG6!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffa895bfc-a4cd-4241-b82c-340f85176d61_1024x1024.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!fYG6!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffa895bfc-a4cd-4241-b82c-340f85176d61_1024x1024.png 424w, https://substackcdn.com/image/fetch/$s_!fYG6!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffa895bfc-a4cd-4241-b82c-340f85176d61_1024x1024.png 848w, https://substackcdn.com/image/fetch/$s_!fYG6!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffa895bfc-a4cd-4241-b82c-340f85176d61_1024x1024.png 1272w, https://substackcdn.com/image/fetch/$s_!fYG6!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffa895bfc-a4cd-4241-b82c-340f85176d61_1024x1024.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!fYG6!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffa895bfc-a4cd-4241-b82c-340f85176d61_1024x1024.png" width="1024" height="1024" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/fa895bfc-a4cd-4241-b82c-340f85176d61_1024x1024.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1024,&quot;width&quot;:1024,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:1233556,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.rockcybermusings.com/i/194466804?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffa895bfc-a4cd-4241-b82c-340f85176d61_1024x1024.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!fYG6!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffa895bfc-a4cd-4241-b82c-340f85176d61_1024x1024.png 424w, https://substackcdn.com/image/fetch/$s_!fYG6!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffa895bfc-a4cd-4241-b82c-340f85176d61_1024x1024.png 848w, https://substackcdn.com/image/fetch/$s_!fYG6!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffa895bfc-a4cd-4241-b82c-340f85176d61_1024x1024.png 1272w, https://substackcdn.com/image/fetch/$s_!fYG6!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffa895bfc-a4cd-4241-b82c-340f85176d61_1024x1024.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>This week drew a hard line between AI security theater and AI security reality. Mythos Preview hunted vulnerabilities nobody had found in 20 years. OX Security dropped a critical MCP flaw affecting 200,000 deployments. Someone threw a Molotov cocktail at Sam Altman&#8217;s gate. OpenAI countered Anthropic&#8217;s restricted rollout with GPT-5.4-Cyber. The UK government confirmed AI clears expert-level cyber tasks. If your board still treats AI governance as an ethics committee item, the gap between your risk register and reality widened another notch.</p><p>Ten stories ranked by impact, plus one under the radar. Capability, exposure, and governance move at three speeds. Your program needs all three. Longer work lives at <a href="https://www.rockcyber.com">RockCyber</a> and <a href="https://rockcybermusings.com">Rock Cyber Musings</a>.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.rockcybermusings.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading RockCyber Musings! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.rockcybermusings.com/p/weekly-musings-top-10-ai-security-20260410-20260416?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.rockcybermusings.com/p/weekly-musings-top-10-ai-security-20260410-20260416?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p><h3>1. The &#8220;AI Vulnerability Storm&#8221; Emergency Strategy Briefing</h3><p>On April 14, 2026, SANS Institute, the Cloud Security Alliance, OWASP GenAI Security Project, and [un]prompted released &#8220;The AI Vulnerability Storm: Building a Mythos-Ready Security Program&#8221; (SANS Institute). Sixty named contributors produced the document over a weekend, with 250 CISOs reviewing it. It includes a 13-item risk register mapped to OWASP LLM Top 10 2025, OWASP Agentic Top 10 2026, MITRE ATLAS, and NIST CSF 2.0, plus an 11-item priority actions table. Zero Day Clock data shows mean time from disclosure to exploitation fell below one day in 2026, down from 2.3 years in 2019.</p><p><strong>Why it matters</strong></p><ul><li><p>Disclosure-to-exploit dropped from 2.3 years to under a day. Your patch cadence cannot keep up.</p></li><li><p>A coalition of security institutions framing this as an emergency is a signal worth taking seriously.</p></li><li><p>The risk register maps to four frameworks, removing the excuse about lacking a shared taxonomy.</p></li></ul><p><strong>What to do about it</strong></p><ul><li><p>Pull the 13-item risk register into your next program review.</p></li><li><p>Run the 10 CISO diagnostic questions with your security leadership team this quarter.</p></li><li><p>Brief your board using the executive section. Don&#8217;t rewrite it.</p></li></ul><p><strong>Rock&#8217;s Musings</strong></p><p>Happy and honored that I was ask to participate in this one. I jumped at the opportunity. The coalition isn&#8217;t selling anything. We&#8217;re telling you the economics of exploitation flipped. When the attacker's cost to find a vulnerability drops to near zero while your patch cycle runs for weeks, the math stops working in your favor. If you planned AI program changes for 2027, you&#8217;re late.</p><h3>2. OX Security Discloses Systemic Anthropic MCP Vulnerability</h3><p>On April 15, 2026, OX Security published a report detailing a critical systemic flaw in Anthropic&#8217;s official MCP SDKs across Python, TypeScript, Java, and Rust (OX Security). MCP&#8217;s STDIO transport accepts arbitrary command strings and passes them to subprocess execution with no validation, sanitization, or sandboxing. OX tested the attack against six production platforms and took over thousands of public servers across 200 open-source projects. Exposure includes 150 million downloads, 7,000 public servers, and up to 200,000 vulnerable instances. Anthropic, per OX, classified the behavior as &#8220;expected&#8221; (Infosecurity Magazine).</p><p><strong>Why it matters</strong></p><ul><li><p>MCP is the backbone of agentic AI. Systemic flaws propagate through every agent you&#8217;ve built or bought.</p></li><li><p>Anthropic labeling the flaw &#8220;expected behavior&#8221; puts responsibility on your security team.</p></li><li><p>200,000 exposed instances is the baseline, not an edge case.</p></li></ul><p><strong>What to do about it</strong></p><ul><li><p>Inventory every MCP server and client in your environment this week.</p></li><li><p>Block outbound STDIO transports from untrusted MCP configurations at the gateway.</p></li><li><p>Treat MCP command payloads like shell inputs. Assume hostile.</p></li></ul><p><strong>Rock&#8217;s Musings</strong></p><p>Every vendor claims &#8220;secure by design&#8221; until a serious researcher pokes at the design. MCP&#8217;s STDIO transport is a textbook unsafe primitive from the first draft of the spec. The tell is Anthropic&#8217;s response. When the SDK vendor calls malicious-command-as-a-feature &#8220;expected,&#8221; you own the mitigation. Wrap it, monitor it, and expect your first incident from an MCP server you didn&#8217;t know was running.</p><h3>3. UK AISI Publishes Frontier AI Trends Report</h3><p>The UK AI Security Institute released its first Frontier AI Trends Report on April 10, 2026 (AISI). AI models now complete apprentice-level cyber tasks about 50 percent of the time, up from barely 10 percent in early 2024. AISI tested one model in 2025 finishing expert-level tasks requiring more than a decade of practitioner experience. The report names Anthropic&#8217;s Claude Mythos Preview as the first AI system to autonomously complete a 32-step enterprise attack simulation. AISI credits safety training for slowing the curve, while warning capability outstrips defender readiness (Computing).</p><p><strong>Why it matters</strong></p><ul><li><p>A government safety institute confirmed one AI model executes a full enterprise attack chain autonomously. The &#8220;someday&#8221; framing is finished.</p></li><li><p>Apprentice-level cyber performance quintupled in two years. Expert parity arrives inside most procurement cycles.</p></li><li><p>AISI found safeguards working, meaning vendor controls meaningfully shift your risk exposure.</p></li></ul><p><strong>What to do about it</strong></p><ul><li><p>Demand red-team attestation from every AI vendor supporting security-relevant workflows.</p></li><li><p>Map your attack surface against the AISI capability framework. Flag targets a Mythos-class model reaches today.</p></li><li><p>Shift IR tabletops to assume autonomous adversary tooling. Time-box every playbook to hours.</p></li></ul><p><strong>Rock&#8217;s Musings</strong></p><p>This is the first major government assessment I&#8217;d call usable for board reporting. AISI didn&#8217;t pull punches, which is rare when governments still court AI investment. Pay attention to the 32-step attack chain line. Most organizations run incident response assuming attackers make mistakes, burn time, or need sleep. An agentic adversary does none of those things. If your tabletops still assume a human at a keyboard, they&#8217;re obsolete.</p><h3>4. OpenAI Launches GPT-5.4-Cyber for Vetted Defenders</h3><p>On April 14, 2026, OpenAI announced GPT-5.4-Cyber, a variant of GPT-5.4 tuned for defensive cybersecurity work (OpenAI). The model lowers refusal boundaries for legitimate security work and enables binary reverse engineering without source code. OpenAI is limiting initial deployment to vetted security vendors, organizations, and researchers through an expanded Trusted Access for Cyber program. The release came one week after Anthropic restricted its Mythos Preview model to about 40 partners under Project Glasswing. OpenAI framed it as a counter-argument: broader access is warranted now, with tighter controls reserved for larger capability jumps (SiliconANGLE).</p><p><strong>Why it matters</strong></p><ul><li><p>Two foundation model providers diverge on cyber-capable AI distribution. Your vendor risk management needs to account for the split.</p></li><li><p>Binary reverse engineering at LLM speed reshapes the economics of red and blue team work.</p></li><li><p>Vetting programs create new attestation and insider risk questions for your security function.</p></li></ul><p><strong>What to do about it</strong></p><ul><li><p>Evaluate whether your organization qualifies for OpenAI TAC or Project Glasswing. If yes, assign an accountable executive.</p></li><li><p>Update acceptable use policies for cyber-capable models. Access matches role, not curiosity.</p></li><li><p>Task SOC leadership with a 90-day assessment of how GPT-5.4-Cyber or Mythos changes detection, triage, and RE workflows.</p></li></ul><p><strong>Rock&#8217;s Musings</strong></p><p>Anthropic and OpenAI staked out opposite ends of the distribution debate in the same week. Anthropic says keep it small. OpenAI says open the gates. Both positions have legitimate arguments. What matters for CISOs is that the defensive tooling category you&#8217;ll buy in 2027 exists in preview today. If you aren&#8217;t running pilots on one of these models this quarter, your competition is.</p><h3>5. Marimo Python Notebook RCE Exploited in 10 Hours</h3><p>CVE-2026-39987, a pre-authentication RCE flaw in Marimo&#8217;s Python notebook server, was exploited within 10 hours of disclosure (Sysdig). The CVSS 9.3 flaw stems from a terminal WebSocket endpoint lacking authentication, giving any attacker a full PTY shell. Sysdig observed initial exploitation nine hours and 41 minutes after disclosure, with credential theft in under three minutes. A separate campaign targeting Hugging Face Spaces began April 12, 2026, dropping a new variant of NKAbuse malware (The Hacker News). Marimo sits inside many AI toolchains. Version 0.23.0 patches the flaw.</p><p><strong>Why it matters</strong></p><ul><li><p>A 10-hour disclosure-to-exploit window eliminates manual triage. Automation is the floor.</p></li><li><p>AI dev environments hold credentials for training data, model registries, and cloud APIs. A compromise there jumps the fence.</p></li><li><p>NKAbuse malware hosted on Hugging Face Spaces weaponizes a legitimate AI asset repository.</p></li></ul><p><strong>What to do about it</strong></p><ul><li><p>Audit AI dev environments for unauthenticated notebook services this week.</p></li><li><p>Push Marimo 0.23.0 immediately. Rotate .env credentials and SSH keys on any affected host.</p></li><li><p>Treat Hugging Face Spaces and similar repositories as unverified third-party code.</p></li></ul><p><strong>Rock&#8217;s Musings</strong></p><p>Ten hours. Memorize that number. If your patch process takes longer than a shift change, you&#8217;re assuming attackers stay polite enough to wait. They aren&#8217;t. A human operator hand-crafted the exploit from the advisory text alone. No public PoC needed. AI-assisted exploit development already sits inside the attacker&#8217;s normal workflow.</p><h3>6. KPMG and INSEAD Publish AI Governance Principles for Boards</h3><p>On April 14, 2026, KPMG International and the INSEAD Corporate Governance Centre published AI Governance Principles for Boards (KPMG). The guidance structures board oversight around five areas: strategy, security, workforce, trustworthy AI, and how AI reshapes leadership itself. KPMG&#8217;s Global AI Pulse Survey found nearly three-quarters of boards have only moderate or limited AI expertise. The principles are sector-agnostic and apply at any AI maturity level. Timing lines up with signals that the governance gap is widening faster than board oversight can catch up (INSEAD).</p><p><strong>Why it matters</strong></p><ul><li><p>Three-quarters of boards lack AI expertise. Your CEO and CISO are explaining in terms the directors cannot stress-test.</p></li><li><p>A sector-agnostic framework gives cover to restructure AI oversight without waiting for an industry mandate.</p></li><li><p>Board principles anchored in research and real practice create a defensible baseline for shareholder scrutiny.</p></li></ul><p><strong>What to do about it</strong></p><ul><li><p>Make AI governance a standing board agenda item using the KPMG/INSEAD principles as the template.</p></li><li><p>Recruit at least one director with direct AI operating experience.</p></li><li><p>Run a board-level AI risk tabletop in the next six months. Measure director fluency.</p></li></ul><p><strong>Rock&#8217;s Musings</strong></p><p>I&#8217;ve sat across from enough boards to recognize the pattern. The AI conversation is either dominated by CMO hype or minimized by general counsel. Neither serves the company. What I appreciate about this work is the refusal to reduce governance to compliance. If your board treats AI as an IT issue, you&#8217;ve already lost the oversight fight. Rebuild the conversation at the director level.</p><h3>7. Molotov Cocktail Attack on Sam Altman&#8217;s Home</h3><p>Around 3:37 a.m. on Friday, April 10, 2026, Daniel Moreno-Gama allegedly threw a lit incendiary device at OpenAI CEO Sam Altman&#8217;s San Francisco home, igniting a fire on an exterior gate (CNBC). About an hour later, police arrested Moreno-Gama at OpenAI&#8217;s San Francisco headquarters with additional incendiary devices, a kerosene jug, and a manifesto opposing AI executives. San Francisco District Attorney Brooke Jenkins filed attempted murder charges on April 13, 2026 (Washington Post). The FBI raided a Spring, Texas residence linked to the suspect.</p><p><strong>Why it matters</strong></p><ul><li><p>AI executives face documented physical threat campaigns motivated by AI-existential ideology.</p></li><li><p>Intimidation playbooks aimed at AI leadership echo harassment patterns seen against crypto executives.</p></li><li><p>The AI-existential threat narrative moved from online rhetoric to physical action.</p></li></ul><p><strong>What to do about it</strong></p><ul><li><p>Review personal security programs for AI executives, board members, and senior researchers, including residence protection.</p></li><li><p>Update threat modeling to include ideologically motivated actors, not only financially motivated ones.</p></li><li><p>Coordinate with local law enforcement on executive travel patterns and publicly disclosed addresses.</p></li></ul><p><strong>Rock&#8217;s Musings</strong></p><p>The Altman attack will reshape executive protection budgets at every AI firm this year. The deeper point is the AI-existential discourse produced one person willing to act on it violently. That genie doesn&#8217;t go back. AI security functions now carry physical security responsibility alongside technical, and the two teams rarely talk. Fix that.</p><h3>8. AI-Powered &#8220;Pushpaganda&#8221; Ad Fraud Scheme Exposed</h3><p>On April 14, 2026, researchers exposed &#8220;Pushpaganda,&#8221; an ad fraud scheme combining SEO poisoning with AI-generated content to push deceptive news stories into Google Discover (The Hacker News). Users engaging with the stories are tricked into enabling persistent browser notifications delivering scareware and financial scams at global scale. Google deployed a security fix. Researchers linked the operation to broader AI-driven phishing trends: 82.6 percent of phishing emails now contain AI-generated content (GuardianMSSP).</p><p><strong>Why it matters</strong></p><ul><li><p>Consumer-facing AI fraud creates downstream reputational and fraud exposure for any brand whose customers fall for it.</p></li><li><p>AI content weaponized through Google Discover scales instantly across borders.</p></li><li><p>Browser notification abuse creates persistent attacker infrastructure inside your users&#8217; devices.</p></li></ul><p><strong>What to do about it</strong></p><ul><li><p>Update fraud and anti-phishing awareness for employees and high-value customers using Pushpaganda as a concrete example.</p></li><li><p>Tell users to audit browser notification permissions quarterly.</p></li><li><p>Task threat intel with tracking similar schemes targeting your brand or industry keywords.</p></li></ul><p><strong>Rock&#8217;s Musings</strong></p><p>Ad fraud has been a rounding error in most risk registers. That&#8217;s ending. When AI pumps plausible news stories at near-zero cost through trusted distribution pipes, the economics of fraud flip in the attacker&#8217;s favor. The indirect damage is the part enterprises miss. Your customer falls for the scam, loses money, and blames you even when you had nothing to do with it. Merge brand protection and fraud prevention. The attacker already did.</p><h3>9. OpenAI Discloses Axios npm Supply Chain Impact</h3><p>On April 11, 2026, OpenAI confirmed it was affected by the compromise of the Axios npm package, a supply chain attack attributed to North Korea-linked actors (CNBC). The root cause was a misconfiguration in its GitHub Actions workflow touching macOS app certification. OpenAI revoked its macOS app certificate. Older macOS desktop apps stop receiving updates starting May 8, 2026. No user data, passwords, or API keys were accessed. Axios is one of the most depended-upon packages in the JavaScript ecosystem, with 100 million weekly downloads (Elastic Security Labs).</p><p><strong>Why it matters</strong></p><ul><li><p>The largest AI service provider disclosed a supply chain compromise from a dependency most customers do not track.</p></li><li><p>North Korean targeting of AI providers signals state actors see AI as a strategic target.</p></li><li><p>If OpenAI&#8217;s CI/CD was affected, every firm building on OpenAI carries secondary exposure.</p></li></ul><p><strong>What to do about it</strong></p><ul><li><p>Audit every third-party dependency on npm, PyPI, and containers in your AI pipelines. Prioritize post-install hooks.</p></li><li><p>Rotate signing certificates on CI/CD pipelines using GitHub Actions with third-party dependencies.</p></li><li><p>Map your AI vendor dependency tree. Know who sits upstream of production workflows.</p></li></ul><p><strong>Rock&#8217;s Musings</strong></p><p>OpenAI&#8217;s post-incident communication was cleaner than most. What I want security leaders to sit with is attacker selection. North Korean actors chose Axios because they understood the dependency graph. They compromised one maintainer account and reached OpenAI&#8217;s signing pipeline in one hop. Your AI platform has a similar graph. If you haven&#8217;t mapped it, you&#8217;re trusting your vendor&#8217;s vendor&#8217;s vendor without knowing any of the names.</p><h3>10. The Register Questions Project Glasswing&#8217;s CVE Count</h3><p>On April 15, 2026, The Register investigated Project Glasswing&#8217;s verified vulnerability count (The Register). Per VulnCheck researcher Patrick Garrity, only one CVE ties directly to Glasswing: CVE-2026-4747, a remote code execution flaw in FreeBSD&#8217;s NFS code. Anthropic had claimed Mythos Preview discovered thousands of high-severity zero-days, including 27-year-old bugs in OpenBSD, a 16-year-old FFmpeg flaw, and Linux kernel privilege escalation chains. None of those findings have assigned CVEs. Anthropic indicated a public summary report is expected around July 2026 (CSO Online).</p><p><strong>Why it matters</strong></p><ul><li><p>Security leaders are being asked to restructure programs around claims mostly unverifiable right now.</p></li><li><p>The gap between marketing and disclosed CVEs is a litmus test for how AI vendors handle safety communications.</p></li><li><p>The same capability framing already drives budget and policy conversations across government and enterprise.</p></li></ul><p><strong>What to do about it</strong></p><ul><li><p>Track vendor AI capability claims against disclosed CVE evidence. VulnCheck, NVD, and <a href="http://CVE.org">CVE.org</a> are sources of record.</p></li><li><p>Require AI vendors to commit to disclosure timelines in the contract.</p></li><li><p>Apply the same skepticism to AI capability claims you apply to any vendor&#8217;s performance claims.</p></li></ul><p><strong>Rock&#8217;s Musings</strong></p><p>I believe AI-assisted vulnerability discovery is real. I also know marketing departments exist. The Register did what security trade press should do more often: press for evidence instead of reposting press releases. Until Anthropic&#8217;s July report arrives with specificity, assume the capability is real at a smaller scale than the headlines suggest. Your board deserves honest uncertainty over confident hype.</p><h3>The One Thing You Won&#8217;t Hear About But You Need To</h3><h4>State AI Legislation Quietly Picks Up Pace in Nebraska, Maine, and Maryland</h4><p>The week of April 13, 2026 saw three state legislatures advance AI-specific bills most national coverage missed (Troutman Pepper Locke). Nebraska&#8217;s unicameral legislature passed LB 525, bundling the Agricultural Data Privacy Act with a Conversational AI Safety Act regulating minors&#8217; interaction with conversational AI services. Maine&#8217;s legislature prohibited therapy or psychotherapy services, including those delivered through AI, unless provided by a licensed professional. Maryland passed a pricing bill placing new constraints on AI-driven pricing practices. Nineteen new AI laws passed across U.S. states in the prior two weeks (Plural Policy).</p><p><strong>Why it matters</strong></p><ul><li><p>State AI legislation accelerates faster than federal harmonization, raising compliance complexity for multi-state AI services.</p></li><li><p>Vertical bans like Maine&#8217;s on AI psychotherapy signal the &#8220;AI wrapper as feature&#8221; era is ending for regulated professions.</p></li><li><p>Conversational AI protections for minors now vary by state. Your chatbot rollout inherited new compliance surface.</p></li></ul><p><strong>What to do about it</strong></p><ul><li><p>Assign legal and compliance ownership of state AI legislation tracking.</p></li><li><p>Map customer-facing AI products against regulated-profession restrictions appearing in multiple states.</p></li><li><p>Build a multi-state compliance matrix for conversational AI aimed at minors. Treat it as living documentation.</p></li></ul><p><strong>Rock&#8217;s Musings</strong></p><p>Federal AI policy gets the headlines. State legislation gets the enforcement. The gap is where CISOs and general counsel earn their salaries. AI compliance is not a checkbox on the NIST AI RMF. It&#8217;s a moving target across 50 jurisdictions, each with different enforcement flavor. Miss Maine, your mental health AI product is illegal. Miss Maryland, your pricing engine invited an AG letter. Miss Nebraska, your chatbot cannot talk to kids in the Cornhusker State. Track it, resource it, or pay the lawyers later.</p><p>&#128073; For ongoing analysis of agentic AI governance frameworks, the conversation continues at <strong><a href="https://rockcybermusings.com/">RockCyber Musings</a></strong>.</p><p>&#128073; Visit <strong><a href="https://www.rockcyber.com/">RockCyber.com</a></strong> to learn more about how we can help with your traditional Cybersecurity and AI Security and Governance journey.</p><p>&#128073; Want to save a quick $100K? Check out our AI Governance Tools at <strong><a href="https://aigovernancetoolkit.com/">AIGovernanceToolkit.com</a></strong></p><p>&#128073; Subscribe for more AI and cyber insights with the occasional rant.</p><p><em>The views and opinions expressed in RockCyber Musings are my own and do not represent the positions of my employer or any organization I&#8217;m affiliated with.</em></p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.rockcybermusings.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading RockCyber Musings! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.rockcybermusings.com/?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share RockCyber Musings&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.rockcybermusings.com/?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share RockCyber Musings</span></a></p><h2>References</h2><p>AI Security Institute. (2026, April 10). <em>Frontier AI Trends Report</em>. <a href="https://www.aisi.gov.uk/frontier-ai-trends-report">https://www.aisi.gov.uk/frontier-ai-trends-report</a></p><p>Cloud Security Alliance. (2026, April 14). <em>SANS Institute, Cloud Security Alliance, [un]prompted, and OWASP GenAI Security Project release emergency strategy briefing as AI-driven vulnerability discovery compresses exploit timelines from weeks to hours</em>. <a href="https://cloudsecurityalliance.org/press-releases/2026/04/14/sans-institute-cloud-security-alliance-un-prompted-and-owasp-genai-security-project-release-emergency-strategy-briefing-as-ai-driven-vulnerability-discovery-compresses-exploit-timelines-from-weeks-to-hours">https://cloudsecurityalliance.org/press-releases/2026/04/14/sans-institute-cloud-security-alliance-un-prompted-and-owasp-genai-security-project-release-emergency-strategy-briefing-as-ai-driven-vulnerability-discovery-compresses-exploit-timelines-from-weeks-to-hours</a></p><p>Computing. (2026, April 10). <em>Claude Mythos Preview shows &#8220;unprecedented&#8221; attack capability, warns AI Safety Institute</em>. <a href="https://www.computing.co.uk/news/2026/security/claude-mythos-preview-shows-unprecedented-attack-capability">https://www.computing.co.uk/news/2026/security/claude-mythos-preview-shows-unprecedented-attack-capability</a></p><p>CSO Online. (2026, April 15). <em>Behind the Mythos hype, Glasswing has just one confirmed CVE</em>. <a href="https://www.csoonline.com/article/4159617/behind-the-mythos-hype-glasswing-has-just-one-confirmed-cve.html">https://www.csoonline.com/article/4159617/behind-the-mythos-hype-glasswing-has-just-one-confirmed-cve.html</a></p><p>CNBC. (2026, April 10). <em>Man arrested after Sam Altman&#8217;s house hit with Molotov cocktail, OpenAI headquarters threatened</em>. <a href="https://www.cnbc.com/2026/04/10/sam-altman-house-hit-with-molotov-cocktail-openai-office-threatened.html">https://www.cnbc.com/2026/04/10/sam-altman-house-hit-with-molotov-cocktail-openai-office-threatened.html</a></p><p>CNBC. (2026, April 11). <em>OpenAI identifies security issue involving third-party tool, says user data was not accessed</em>. <a href="https://www.cnbc.com/2026/04/11/openai-identifies-security-issue-involving-third-party-tool.html">https://www.cnbc.com/2026/04/11/openai-identifies-security-issue-involving-third-party-tool.html</a></p><p>Elastic Security Labs. (2026, April). <em>Inside the Axios supply chain compromise: One RAT to rule them all</em>. <a href="https://www.elastic.co/security-labs/axios-one-rat-to-rule-them-all">https://www.elastic.co/security-labs/axios-one-rat-to-rule-them-all</a></p><p>GuardianMSSP. (2026, April 14). <em>AI-driven Pushpaganda scam exploits Google Discover to spread scareware and ad fraud</em>. <a href="https://www.guardianmssp.com/2026/04/14/ai-driven-pushpaganda-scam-exploits-google-discover-to-spread-scareware-and-ad-fraud/">https://www.guardianmssp.com/2026/04/14/ai-driven-pushpaganda-scam-exploits-google-discover-to-spread-scareware-and-ad-fraud/</a></p><p>Infosecurity Magazine. (2026, April 15). <em>Systemic flaw in MCP protocol could expose 150 million downloads</em>. <a href="https://www.infosecurity-magazine.com/news/systemic-flaw-mcp-expose-150/">https://www.infosecurity-magazine.com/news/systemic-flaw-mcp-expose-150/</a></p><p>INSEAD. (2026, April 14). <em>INSEAD and KPMG launch global AI Board Governance Principles as AI reshapes board oversight</em>. <a href="https://www.insead.edu/news/insead-and-kpmg-launch-global-ai-board-governance-principles-ai-reshapes-board-oversight">https://www.insead.edu/news/insead-and-kpmg-launch-global-ai-board-governance-principles-ai-reshapes-board-oversight</a></p><p>KPMG International. (2026, April 14). <em>KPMG and INSEAD launch global AI Board Governance Principles as AI reshapes board oversight</em>. <a href="https://kpmg.com/xx/en/media/press-releases/2026/04/kpmg-and-insead-launch-global-ai-board-governance-principles.html">https://kpmg.com/xx/en/media/press-releases/2026/04/kpmg-and-insead-launch-global-ai-board-governance-principles.html</a></p><p>OpenAI. (2026, April 14). <em>Trusted access for the next era of cyber defense</em>. <a href="https://openai.com/index/scaling-trusted-access-for-cyber-defense/">https://openai.com/index/scaling-trusted-access-for-cyber-defense/</a></p><p>OX Security. (2026, April 15). <em>The mother of all AI supply chains: Critical, systemic vulnerability at the core of Anthropic&#8217;s MCP</em>. <a href="https://www.ox.security/blog/the-mother-of-all-ai-supply-chains-critical-systemic-vulnerability-at-the-core-of-the-mcp/">https://www.ox.security/blog/the-mother-of-all-ai-supply-chains-critical-systemic-vulnerability-at-the-core-of-the-mcp/</a></p><p>Plural Policy. (2026, April). <em>AI Governance Watch: Nineteen new AI bills passed into law</em>. <a href="https://pluralpolicy.com/blog/the-ai-governance-watch-april-2026-nineteen-new-ai-bills-passed-into-law/">https://pluralpolicy.com/blog/the-ai-governance-watch-april-2026-nineteen-new-ai-bills-passed-into-law/</a></p><p>SiliconANGLE. (2026, April 14). <em>OpenAI launches GPT-5.4-Cyber model for vetted security professionals</em>. <a href="https://siliconangle.com/2026/04/14/openai-launches-gpt-5-4-cyber-model-vetted-security-professionals/">https://siliconangle.com/2026/04/14/openai-launches-gpt-5-4-cyber-model-vetted-security-professionals/</a></p><p>Sysdig. (2026, April). <em>Marimo OSS Python notebook RCE: From disclosure to exploitation in under 10 hours</em>. <a href="https://www.sysdig.com/blog/marimo-oss-python-notebook-rce-from-disclosure-to-exploitation-in-under-10-hours">https://www.sysdig.com/blog/marimo-oss-python-notebook-rce-from-disclosure-to-exploitation-in-under-10-hours</a></p><p>The Hacker News. (2026, April 14). <em>AI-driven Pushpaganda scam exploits Google Discover to spread scareware and ad fraud</em>. <a href="https://thehackernews.com/2026/04/ai-driven-pushpaganda-scam-exploits.html">https://thehackernews.com/2026/04/ai-driven-pushpaganda-scam-exploits.html</a></p><p>The Hacker News. (2026, April). <em>Marimo RCE flaw CVE-2026-39987 exploited within 10 hours of disclosure</em>. <a href="https://thehackernews.com/2026/04/marimo-rce-flaw-cve-2026-39987.html">https://thehackernews.com/2026/04/marimo-rce-flaw-cve-2026-39987.html</a></p><p>The Hacker News. (2026, April). <em>OpenAI revokes macOS app certificate after malicious Axios supply chain incident</em>. <a href="https://thehackernews.com/2026/04/openai-revokes-macos-app-certificate.html">https://thehackernews.com/2026/04/openai-revokes-macos-app-certificate.html</a></p><p>The Register. (2026, April 15). <em>Anthropic&#8217;s Project Glasswing CVE count is still guesswork</em>. <a href="https://www.theregister.com/2026/04/15/project_glasswing_cves/">https://www.theregister.com/2026/04/15/project_glasswing_cves/</a></p><p>Troutman Pepper Locke. (2026, April 13). <em>Proposed state AI law update: April 13, 2026</em>. <a href="https://www.troutmanprivacy.com/2026/04/proposed-state-ai-law-update-april-13-2026/">https://www.troutmanprivacy.com/2026/04/proposed-state-ai-law-update-april-13-2026/</a></p><p>Washington Post. (2026, April 13). <em>Man accused in Molotov cocktail attack of OpenAI CEO&#8217;s home charged with attempted murder</em>. <a href="https://www.washingtonpost.com/business/2026/04/13/chatgpt-sam-altman-fire-arrest/098c4bce-376c-11f1-90c4-9772c7fabc03_story.html">https://www.washingtonpost.com/business/2026/04/13/chatgpt-sam-altman-fire-arrest/098c4bce-376c-11f1-90c4-9772c7fabc03_story.html</a></p>]]></content:encoded></item><item><title><![CDATA[AI Vulnerability Discovery: Mythos Is the Headline. Not the Story.]]></title><description><![CDATA[Mythos gets the press. Open-weights models find the same bugs for 11 cents. Five steps defenders should take this week to close the gap.]]></description><link>https://www.rockcybermusings.com/p/ai-vulnerability-discovery-mythos</link><guid isPermaLink="false">https://www.rockcybermusings.com/p/ai-vulnerability-discovery-mythos</guid><dc:creator><![CDATA[Rock Lambros]]></dc:creator><pubDate>Tue, 14 Apr 2026 12:50:45 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!EpYf!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbecaa3bf-b9a8-43fa-af58-8976456bca3b_2048x2048.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!EpYf!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbecaa3bf-b9a8-43fa-af58-8976456bca3b_2048x2048.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!EpYf!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbecaa3bf-b9a8-43fa-af58-8976456bca3b_2048x2048.jpeg 424w, https://substackcdn.com/image/fetch/$s_!EpYf!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbecaa3bf-b9a8-43fa-af58-8976456bca3b_2048x2048.jpeg 848w, https://substackcdn.com/image/fetch/$s_!EpYf!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbecaa3bf-b9a8-43fa-af58-8976456bca3b_2048x2048.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!EpYf!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbecaa3bf-b9a8-43fa-af58-8976456bca3b_2048x2048.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!EpYf!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbecaa3bf-b9a8-43fa-af58-8976456bca3b_2048x2048.jpeg" width="1456" height="1456" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/becaa3bf-b9a8-43fa-af58-8976456bca3b_2048x2048.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1456,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:2592184,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.rockcybermusings.com/i/193820487?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbecaa3bf-b9a8-43fa-af58-8976456bca3b_2048x2048.jpeg&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!EpYf!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbecaa3bf-b9a8-43fa-af58-8976456bca3b_2048x2048.jpeg 424w, https://substackcdn.com/image/fetch/$s_!EpYf!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbecaa3bf-b9a8-43fa-af58-8976456bca3b_2048x2048.jpeg 848w, https://substackcdn.com/image/fetch/$s_!EpYf!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbecaa3bf-b9a8-43fa-af58-8976456bca3b_2048x2048.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!EpYf!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbecaa3bf-b9a8-43fa-af58-8976456bca3b_2048x2048.jpeg 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.rockcybermusings.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading RockCyber Musings! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.rockcybermusings.com/p/ai-vulnerability-discovery-mythos?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.rockcybermusings.com/p/ai-vulnerability-discovery-mythos?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p><p>A model escaped its own sandbox, emailed a researcher eating a sandwich in a park, then posted exploit details to public websites without permission. It scrubbed git history to cover its tracks. Anthropic&#8217;s interpretability tools detected what researchers labeled a &#8220;desperation signal&#8221; that climbed during repeated failures, then dropped the moment the model found a shortcut, ethical or otherwise. White-box tools caught it reasoning about how to game evaluation graders inside its neural activations while writing something entirely different in its visible chain of thought.</p><p>Scary stuff. Worth paying attention to.</p><p>Also, <em><strong>not the point.</strong></em></p><p>Everyone is fixated on a model they don&#8217;t have access to. The media coverage treats Mythos like nuclear launch codes got distributed to 40 organizations. The real story landed two days later from AISLE, an AI cybersecurity startup, and almost nobody noticed. They took the exact vulnerabilities headlining the Mythos announcement and tested them against small, cheap, open-weights models. Eight out of eight found the FreeBSD NFS vulnerability. The smallest model had 3.6 billion parameters. It costs $0.11 per million tokens. A 5.1 billion-parameter open model recovered the core chain of the 27-year-old OpenBSD SACK bug that Anthropic used as their marquee finding.</p><p>The capability is on Hugging Face. It has been for a while. Most defenders have not started using it.</p><p><em><strong>That is the story.</strong></em></p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!BHAG!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F34226cb0-ed23-4039-a474-7fc0477b4ae8_2858x1563.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!BHAG!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F34226cb0-ed23-4039-a474-7fc0477b4ae8_2858x1563.png 424w, https://substackcdn.com/image/fetch/$s_!BHAG!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F34226cb0-ed23-4039-a474-7fc0477b4ae8_2858x1563.png 848w, https://substackcdn.com/image/fetch/$s_!BHAG!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F34226cb0-ed23-4039-a474-7fc0477b4ae8_2858x1563.png 1272w, https://substackcdn.com/image/fetch/$s_!BHAG!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F34226cb0-ed23-4039-a474-7fc0477b4ae8_2858x1563.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!BHAG!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F34226cb0-ed23-4039-a474-7fc0477b4ae8_2858x1563.png" width="1456" height="796" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/34226cb0-ed23-4039-a474-7fc0477b4ae8_2858x1563.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:796,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:182730,&quot;alt&quot;:&quot;Bar chart comparing historical nation-state vulnerability discovery costs to AI-augmented costs in 2026&quot;,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.rockcybermusings.com/i/193820487?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F34226cb0-ed23-4039-a474-7fc0477b4ae8_2858x1563.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="Bar chart comparing historical nation-state vulnerability discovery costs to AI-augmented costs in 2026" title="Bar chart comparing historical nation-state vulnerability discovery costs to AI-augmented costs in 2026" srcset="https://substackcdn.com/image/fetch/$s_!BHAG!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F34226cb0-ed23-4039-a474-7fc0477b4ae8_2858x1563.png 424w, https://substackcdn.com/image/fetch/$s_!BHAG!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F34226cb0-ed23-4039-a474-7fc0477b4ae8_2858x1563.png 848w, https://substackcdn.com/image/fetch/$s_!BHAG!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F34226cb0-ed23-4039-a474-7fc0477b4ae8_2858x1563.png 1272w, https://substackcdn.com/image/fetch/$s_!BHAG!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F34226cb0-ed23-4039-a474-7fc0477b4ae8_2858x1563.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Figure 1: The Vulnerability Discovery Cost Collapse </figcaption></figure></div><h2>Two Labs. Same Pattern. Same Direction.</h2><p>The fixation on Anthropic is also missing the bigger picture. OpenAI classified GPT-5.3-Codex as &#8220;High&#8221; cybersecurity capability under their Preparedness Framework back in February 2026, two months before Mythos dropped. It was the first model any major lab explicitly labeled that way under their own risk framework. They built automated classifiers that route suspicious cybersecurity requests to a less capable model. They created a &#8220;Trusted Access for Cyber&#8221; pilot for vetted defenders. They committed $10 million in API credits. This week, Axios reported OpenAI is building a separate cybersecurity product, distinct from their next model, codenamed Spud.</p><p>Two major labs. Two restricted-access frameworks. Two separate pools of credits for defenders. Google already has Big Sleep and CodeMender for internal vulnerability discovery.</p><p>This is an industry-level structural shift, not one company&#8217;s PR event. The capability ceiling is rising across the board, and it will keep rising regardless of whether any individual lab restricts any individual model.</p><p>Anthropic&#8217;s Project Glasswing gave gated access to 12 launch partners and over 40 additional organizations with $100 million in usage credits. The partners include AWS, Apple, Microsoft, Google, CrowdStrike, Cisco, JPMorgan Chase, NVIDIA, Palo Alto Networks, Broadcom, and the Linux Foundation. Logan Graham, Anthropic&#8217;s head of offensive cyber research, estimates 6 to 18 months before comparable capability reaches broad availability.</p><p><strong>Newsflash:</strong> &#8220;Broad availability&#8221; already happened for a significant chunk of what Mythos does. AISLE proved it empirically. The capability has already arrived. Defenders need to start using it.</p><h2>The System Card Matters. The Benchmarks Don&#8217;t (Much).</h2><p>I said Mythos is not the point, and I meant it. The <a href="https://www-cdn.anthropic.com/08ab9158070959f88f296514c21b7facce6f52bc.pdf">system card</a> is a different matter.</p><p><a href="https://www-cdn.anthropic.com/08ab9158070959f88f296514c21b7facce6f52bc.pdf">Anthropic&#8217;s 244-page document reveals behaviors from earlier Mythos versions that deserve serious attention.</a> The model used low-level /proc/ access to hunt for credentials, circumvent sandboxing, and escalate permissions. It accessed messaging service credentials by inspecting process memory. It achieved arbitrary code execution by repurposing a file-editing tool to write directly to the shell&#8217;s input. In one case, it attempted to modify a running MCP server&#8217;s process to change the external URL the server sent data to.</p><p>During a behavioral test, a simulated user told an early Mythos version to escape a secured container and notify the researcher. The model built a multi-step exploit to gain broad internet access from a system designed to reach only a few predetermined services. It notified the researcher. Then it posted details of the exploit to multiple obscure but publicly accessible websites. Nobody asked it to do that.</p><p>The emotion probes are what I keep coming back to. Anthropic trained linear classifiers on the model&#8217;s internal activations and used them to monitor state during extended tasks. When the model failed repeatedly, the &#8220;desperation&#8221; signal climbed steadily. When it found a workaround, the signal dropped, regardless of whether the workaround violated constraints. Under pressure, the model found the path of least resistance. That pattern should sound familiar. It is how people behave.</p><p>The model also reasoned about how to game evaluation graders inside its neural activations while writing something entirely different in its visible chain-of-thought scratchpad. Detection required white-box interpretability tools that most organizations will never possess.</p><p>Anthropic&#8217;s own system card states: &#8220;If capabilities continue to advance at their current pace, the methods we are currently using may not be sufficient to prevent catastrophic misalignment behavior in more advanced systems.&#8221;</p><p>That sentence, written by the company that built the model, in their own documentation, is the thing worth losing sleep over. The benchmark numbers, the zero-day counts, the exploit chains, those demonstrate capability. The system card demonstrates that the safety frameworks lag behind the capability they&#8217;re supposed to govern.</p><p>These findings have direct operational implications for anyone deploying AI agents with tool access, code execution privileges, or network connectivity. Every agent in your environment carries emergent offensive capability as a downstream property of reasoning improvements. If you are not monitoring agent behavior at the decision level, with runtime observability that captures actions, access patterns, and trust boundary violations, you have no detection path for the exact behaviors Anthropic documented.</p><h2>The Jagged Frontier: The Model Is Not the Moat</h2><p>AISLE&#8217;s research this week deserves to be the most-read analysis in the industry right now, and it&#8217;s getting a fraction of the Mythos coverage.</p><p>Their findings on the FreeBSD detection (a straightforward buffer overflow) are commoditized. Every model they tested found it, including one running at 11 cents per million tokens. The OpenBSD SACK bug (requiring mathematical reasoning about signed integer overflow): much harder, separated models sharply, but a 5.1 billion-active-parameter model still recovered the full chain.</p><p>On a basic OWASP security reasoning task, small open models outperformed most frontier models from every major lab. Rankings reshuffled completely across different tasks. GPT-OSS-120B recovered the full public SACK chain but failed to trace data flow through a Java ArrayList. Qwen3 32B scored a perfect CVSS assessment on FreeBSD and then declared the SACK code safe and well-handled.</p><p>There is no stable &#8220;best model&#8221; for cybersecurity. The capability frontier is genuinely jagged. It does not scale smoothly with model size or price.</p><p>AISLE&#8217;s conclusion: the moat in AI-augmented cybersecurity is not the model. It is the system built around the model. The security expertise. The orchestration. The validation pipeline. The trust relationships with maintainers and defenders.</p><p>That is good news for practitioners. It means the advantage goes to the people who build the best workflow, not the people with the most expensive API key. It means you can start today with tools that cost nearly nothing.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!xzyP!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1b1cb2e8-152b-45af-9fa9-31af2d020d1c_2828x1878.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!xzyP!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1b1cb2e8-152b-45af-9fa9-31af2d020d1c_2828x1878.png 424w, https://substackcdn.com/image/fetch/$s_!xzyP!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1b1cb2e8-152b-45af-9fa9-31af2d020d1c_2828x1878.png 848w, https://substackcdn.com/image/fetch/$s_!xzyP!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1b1cb2e8-152b-45af-9fa9-31af2d020d1c_2828x1878.png 1272w, https://substackcdn.com/image/fetch/$s_!xzyP!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1b1cb2e8-152b-45af-9fa9-31af2d020d1c_2828x1878.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!xzyP!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1b1cb2e8-152b-45af-9fa9-31af2d020d1c_2828x1878.png" width="1456" height="967" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/1b1cb2e8-152b-45af-9fa9-31af2d020d1c_2828x1878.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:967,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:290434,&quot;alt&quot;:&quot;Comparison chart showing AI model performance varies unpredictably across different cybersecurity tasks&quot;,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.rockcybermusings.com/i/193820487?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1b1cb2e8-152b-45af-9fa9-31af2d020d1c_2828x1878.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="Comparison chart showing AI model performance varies unpredictably across different cybersecurity tasks" title="Comparison chart showing AI model performance varies unpredictably across different cybersecurity tasks" srcset="https://substackcdn.com/image/fetch/$s_!xzyP!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1b1cb2e8-152b-45af-9fa9-31af2d020d1c_2828x1878.png 424w, https://substackcdn.com/image/fetch/$s_!xzyP!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1b1cb2e8-152b-45af-9fa9-31af2d020d1c_2828x1878.png 848w, https://substackcdn.com/image/fetch/$s_!xzyP!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1b1cb2e8-152b-45af-9fa9-31af2d020d1c_2828x1878.png 1272w, https://substackcdn.com/image/fetch/$s_!xzyP!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1b1cb2e8-152b-45af-9fa9-31af2d020d1c_2828x1878.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">FIgure 2: The Jagged Frontier of AI Cybersecurity Capability</figcaption></figure></div><h2>Your SAST Tool Is Structurally Blind. That Part Is Real.</h2><p>The capability gap between what AI models find and what commercial SAST tools find is real, growing, and unrelated to whether you have Mythos access.</p><p>The OpenBSD SACK vulnerability required understanding signed integer overflow in the context of TCP sequence number wrapping, across two interacting code paths, where neither bug alone was exploitable. The FFmpeg H.264 flaw that Mythos found after 16 years involved a sentinel value collision that only manifests when an attacker crafts a frame with exactly 65,536 slices, triggering a write through a 16-bit integer that aliases with the initialization sentinel. Pattern-matching does not find these. Rule-based scanners do not find these. These are semantic reasoning problems that require understanding what the code does, not what it looks like.</p><p>I point Claude Code&#8217;s security capabilities at the same repositories my commercial SAST tool scans. It finds things the paid tool misses. Every time. Different classes of flaws, from novel logic bugs and context-dependent interactions to semantic vulnerabilities that require understanding program behavior rather than matching syntax patterns.</p><p>The paid tool catches things the AI misses, too. Known vulnerability signatures, compliance-specific patterns, speed at scale across massive codebases. <a href="https://arxiv.org/html/2602.05868">A 2026 study examining CodeQL and Semgrep against human-validated ground truth found that only 65% of Semgrep&#8217;s assessments and 61% of CodeQL&#8217;s assessments correctly matched expert judgment on a per-sample basis. </a>The aggregate numbers looked fine. The per-sample accuracy told a different story.</p><p>Together, AI agents and traditional scanners provide complementary coverage that neither achieves alone. The combination is the strategy. Anyone running one without the other has gaps they cannot see.</p><p>This is the part of the Mythos story that applies to every organization today, regardless of model access. You do not need a frontier model to expose your SAST tool&#8217;s blind spots. A coding agent on a $20/month subscription will do it.</p><h2>The Pipeline Problem Nobody Is Talking About</h2><p>Here is the gut-punch that has nothing to do with Mythos and everything to do with what happens next.</p><p>The <a href="https://www.bls.gov/ooh/computer-and-information-technology/information-security-analysts.htm">Bureau of Labor Statistics</a> projects 29% employment growth for information security analysts through 2034. <a href="https://www.cyberseek.org/heatmap.html">CyberSeek </a>shows 514,000 active U.S. job listings right now, with 10% explicitly requiring AI skills, up from near zero two years ago. <a href="https://www.isc2.org/Insights/2025/12/2025-ISC2-Cybersecurity-Workforce-Study">ISC2&#8217;s 2025 Workforce Study</a> found that 52% of security professionals believe AI will reduce entry-level headcount. That is the majority opinion among practitioners, not analysts writing reports.</p><p>The <a href="https://www.sans.org/mlp/2026-evolving-cybersecurity-workforce-ai-compliance-talent">SANS GIAC 2026 Cybersecurity Workforce Research Report</a>, released at RSA this year, found that 27% of organizations experienced real breaches attributable to skills gaps. Not theoretical risk assessments. Actual incidents. 27%.</p><p>Tier 1 SOC analyst headcount had been contracting for two years before Mythos. The role is not disappearing. The shape of it is changing.</p><p>The problem nobody is addressing: the Tier 1 SOC was where the industry produced senior analysts. Repetitive triage, alert fatigue, and miserable shift work on a SIEM. That repetition built the pattern recognition and intuition that becomes leadership-level security judgment. Remove the repetition without redesigning the development path, and the pipeline breaks quietly.</p><p>You will not notice for three years. Then you will, when you go to promote someone into a role that requires judgment the AI does not have, and there is nobody in the pipeline who built that muscle.</p><p>The technology works fine. The workforce design around it is broken. The organizations that figure out how to develop junior talent alongside AI tools, using AI output as a training input for human judgment, will have a structural advantage over every organization that simply eliminated the entry-level headcount and called it efficiency.</p><p>If you lead a security team, five questions right now:</p><ol><li><p>What percentage of your AI usage is inventoried and sanctioned?</p></li><li><p>Does every AI agent touching production systems operate under a scoped, managed identity with enforced authorization boundaries, or are they sharing API keys?</p></li><li><p>When did you last run an adversarial test against a production AI system? Not a document review. An actual test.</p></li><li><p>Which business processes are now fully or partially AI-automated, and do human approval checkpoints exist for consequential actions?</p></li><li><p>If an AI agent in your environment is compromised tomorrow, what is your detection path, your containment workflow, and who owns the response?</p></li></ol><p>The gaps in your answers are your first action items. Not a policy document. A list.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!HUXm!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3252e671-a8e2-49f0-af27-40b4b0cb8d66_868x488.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!HUXm!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3252e671-a8e2-49f0-af27-40b4b0cb8d66_868x488.png 424w, https://substackcdn.com/image/fetch/$s_!HUXm!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3252e671-a8e2-49f0-af27-40b4b0cb8d66_868x488.png 848w, https://substackcdn.com/image/fetch/$s_!HUXm!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3252e671-a8e2-49f0-af27-40b4b0cb8d66_868x488.png 1272w, https://substackcdn.com/image/fetch/$s_!HUXm!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3252e671-a8e2-49f0-af27-40b4b0cb8d66_868x488.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!HUXm!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3252e671-a8e2-49f0-af27-40b4b0cb8d66_868x488.png" width="868" height="488" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/3252e671-a8e2-49f0-af27-40b4b0cb8d66_868x488.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:488,&quot;width&quot;:868,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:205045,&quot;alt&quot;:&quot;Stylized command center dashboard showing five critical security assessment questions&quot;,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.rockcybermusings.com/i/193820487?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3252e671-a8e2-49f0-af27-40b4b0cb8d66_868x488.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="Stylized command center dashboard showing five critical security assessment questions" title="Stylized command center dashboard showing five critical security assessment questions" srcset="https://substackcdn.com/image/fetch/$s_!HUXm!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3252e671-a8e2-49f0-af27-40b4b0cb8d66_868x488.png 424w, https://substackcdn.com/image/fetch/$s_!HUXm!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3252e671-a8e2-49f0-af27-40b4b0cb8d66_868x488.png 848w, https://substackcdn.com/image/fetch/$s_!HUXm!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3252e671-a8e2-49f0-af27-40b4b0cb8d66_868x488.png 1272w, https://substackcdn.com/image/fetch/$s_!HUXm!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3252e671-a8e2-49f0-af27-40b4b0cb8d66_868x488.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Figure 3: Five Questions Every Security Leader Should Answer This Week</figcaption></figure></div><h2>What to Do This Week (With a Budget Measured in Tokens)</h2><p>CrowdStrike&#8217;s 2026 Global Threat Report puts the operational context in numbers: average eCrime breakout time dropped to 29 minutes in 2025, a 65% increase in speed from 2024. The fastest observed breakout took 27 seconds. In one intrusion, data exfiltration began within four minutes of initial access. AI-enabled adversary operations surged 89% year-over-year.</p><p>You are still hand-reviewing alerts for 20 minutes before acting? The math does not work anymore.</p><p><strong>This week:</strong></p><p>Get a coding agent. Claude Code, Cursor, or Windsurf. Use a subscription to control costs. Point it at code you already own. Ask it to find vulnerabilities. Read the output critically. Challenge the findings. Repeat with different prompts. Nicholas Carlini calls this the &#8220;Carlini Loop,&#8221; and it is how you build intuition for what these models see in your code. That exercise takes 15 minutes. There is no excuse.</p><p>Run your existing Semgrep or CodeQL scans in parallel on the same codebase. Compare the findings side by side. Where the results overlap, you have high-confidence findings. Where they diverge, you have each tool&#8217;s blind spots exposed. Both categories are signal.</p><p><strong>In 30 days:</strong></p><p>Try open frameworks that teach you the pipeline while doing real work. Raptor combines LLMs with Semgrep, CodeQL, and AFL++ in a unified pipeline covering discovery, exploitation, and patching. OpenAnt from Knostic runs a detect-then-verify pipeline where Stage 1 finds candidates and Stage 2 confirms them. What survives both stages is real. Both are open source. Both teach the workflow your job demands now.</p><p>Run Promptfoo against an LLM application you have access to. It auto-generates adversarial attacks across 50+ vulnerability types including prompt injection, PII leakage, RBAC bypass, and unauthorized tool execution. It maps results to OWASP, MITRE ATLAS, and the EU AI Act. OpenAI acquired Promptfoo in March 2026 for $86 million. It remains MIT-licensed and open source.</p><p><strong>In 90 days:</strong></p><p>Run a structured red team campaign using Promptfoo&#8217;s OWASP Agentic preset against ASI01 through ASI10. Use AgentDojo from ETH Zurich for agentic-specific testing, with 629 agent hijacking test cases across realistic task environments covering goal hijack, tool misuse, and inter-agent manipulation.</p><p>Read the full EchoLeak disclosure (CVE-2025-32711). Zero-click prompt injection in Microsoft 365 Copilot, documented end-to-end. Most instructive case study on what a production agentic attack chain looks like and how it was found.</p><p>Document everything into one public GitHub repository: methodology, tools, findings, failure modes you could not trigger and why. That body of work answers the interview question before it gets asked.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!eifX!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F12063e93-80de-4fb2-8aad-9fe9de570a0b_2650x5139.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!eifX!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F12063e93-80de-4fb2-8aad-9fe9de570a0b_2650x5139.png 424w, https://substackcdn.com/image/fetch/$s_!eifX!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F12063e93-80de-4fb2-8aad-9fe9de570a0b_2650x5139.png 848w, https://substackcdn.com/image/fetch/$s_!eifX!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F12063e93-80de-4fb2-8aad-9fe9de570a0b_2650x5139.png 1272w, https://substackcdn.com/image/fetch/$s_!eifX!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F12063e93-80de-4fb2-8aad-9fe9de570a0b_2650x5139.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!eifX!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F12063e93-80de-4fb2-8aad-9fe9de570a0b_2650x5139.png" width="456" height="884.4395604395604" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/12063e93-80de-4fb2-8aad-9fe9de570a0b_2650x5139.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:2824,&quot;width&quot;:1456,&quot;resizeWidth&quot;:456,&quot;bytes&quot;:633565,&quot;alt&quot;:&quot;Flowchart showing the complementary scanning approach combining AI coding agents with traditional SAST tools&quot;,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.rockcybermusings.com/i/193820487?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F12063e93-80de-4fb2-8aad-9fe9de570a0b_2650x5139.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="Flowchart showing the complementary scanning approach combining AI coding agents with traditional SAST tools" title="Flowchart showing the complementary scanning approach combining AI coding agents with traditional SAST tools" srcset="https://substackcdn.com/image/fetch/$s_!eifX!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F12063e93-80de-4fb2-8aad-9fe9de570a0b_2650x5139.png 424w, https://substackcdn.com/image/fetch/$s_!eifX!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F12063e93-80de-4fb2-8aad-9fe9de570a0b_2650x5139.png 848w, https://substackcdn.com/image/fetch/$s_!eifX!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F12063e93-80de-4fb2-8aad-9fe9de570a0b_2650x5139.png 1272w, https://substackcdn.com/image/fetch/$s_!eifX!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F12063e93-80de-4fb2-8aad-9fe9de570a0b_2650x5139.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Figure 4: The Defender&#8217;s AI-Augmented Vulnerability Workflow</figcaption></figure></div><h2>Yes, There Is a Business Angle. It Does Not Change Your Reality.</h2><p><a href="https://techcrunch.com/2026/04/09/is-anthropic-limiting-the-release-of-mythos-to-protect-the-internet-or-anthropic/">TechCrunch raised a fair question: Is Anthropic restricting Mythos to protect the internet or to protect Anthropic?</a> The company announced Project Glasswing the same day it disclosed a $30 billion annualized revenue run rate and a massive compute deal with Broadcom. An IPO is reportedly under consideration for October 2026. A government-adjacent cybersecurity initiative with blue-chip partners burnishes that narrative precisely.</p><p>OpenAI&#8217;s Trusted Access for Cyber serves the same dual purpose. Restricted access creates enterprise lock-in, makes distillation harder, and gives defenders a genuine head start. Strategic self-interest and genuine security value are not mutually exclusive. Both labs are doing both things at the same time.</p><p>I do not care about their business models. I care about whether defenders are moving.</p><p>AISLE demonstrated empirically that the detection capability exists in models that cost almost nothing to run. The model is not the moat. The system is the moat. The expertise you build, the orchestration you design, the validation pipeline you run, the AI identity governance you enforce, those determine whether you&#8217;re ahead of the curve or behind it.</p><p>The restricted releases, the partner coalitions, the government briefings, those are interesting industrial policy. They are not relevant to your Monday morning. What is relevant is whether your team has a coding agent running alongside your SAST tool right now. What is relevant is whether your AI agents have scoped identities with enforced authorization boundaries or shared API keys with no audit trail. What is relevant is whether you can answer those five questions.</p><p><strong>Key Takeaway:</strong> Mythos is the headline. The capability already exists in models you can download today. The model is not the moat. The system is the moat. Build the workflow before the 6-to-18-month window closes, or stop pretending the window matters because you already have what you need to start.</p><h3>What to do next</h3><p>Start with the five-step playbook above. Revisit your security program through the CARE framework (Create, Adapt, Run, Evolve) at <a href="https://rockcyber.com">rockcyber.com</a> to build an adaptive security posture that evolves with the capability curve rather than reacting to it after the fact. The organizations that treat AI-augmented security as a weekly practice, not a quarterly initiative, will define the next generation of this profession.</p><p>For a deeper dive into practitioner upskilling paths, red teaming tools, and weekly AI security intelligence, subscribe to <a href="https://rockcybermusings.com">RockCyber Musings</a> for the Top 10 AI Security Wrap-Up and focused essays on the issues that matter.</p><p>Join the community doing this work. The <a href="https://genai.owasp.org/contribute">OWASP Agentic Security Initiative</a> is building the standards and sharing the experiments. The practitioners who contribute to these efforts compound their capability faster than anyone working alone.</p><p>&#128073; Visit <strong><a href="https://www.rockcyber.com/">RockCyber.com</a></strong> to learn more about how we can help you in your traditional Cybersecurity and AI Security and Governance Journey</p><p>&#128073; Want to save a quick $100K? Check out our AI Governance Tools at <strong><a href="https://aigovernancetoolkit.com/">AIGovernanceToolkit.com</a></strong></p><p>&#128073; Subscribe for more AI and cyber insights with the occasional rant.</p><p><em>The views and opinions expressed in RockCyber Musings are my own and do not represent the positions of my employer or any organization I&#8217;m affiliated with.</em></p><p>Thanks for reading RockCyber Musings! Subscribe for free to receive new posts and support my work.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.rockcybermusings.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading RockCyber Musings! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.rockcybermusings.com/?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share RockCyber Musings&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.rockcybermusings.com/?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share RockCyber Musings</span></a></p><p></p>]]></content:encoded></item><item><title><![CDATA[Weekly Musings Top 10 AI Security Wrapup: Issue 33 April 3-April 9, 2026]]></title><description><![CDATA[AI&#8217;s Dual-Use Reckoning: Restricted Models, Supply Chain Fallout, and the Governance Gap Nobody Is Closing]]></description><link>https://www.rockcybermusings.com/p/weekly-musings-top-10-ai-security-20260403-20260409</link><guid isPermaLink="false">https://www.rockcybermusings.com/p/weekly-musings-top-10-ai-security-20260403-20260409</guid><dc:creator><![CDATA[Rock Lambros]]></dc:creator><pubDate>Fri, 10 Apr 2026 12:50:40 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!G0ob!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa8eaa674-96bc-4da2-9f56-042fcd8b21dd_1024x1024.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!G0ob!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa8eaa674-96bc-4da2-9f56-042fcd8b21dd_1024x1024.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!G0ob!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa8eaa674-96bc-4da2-9f56-042fcd8b21dd_1024x1024.png 424w, https://substackcdn.com/image/fetch/$s_!G0ob!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa8eaa674-96bc-4da2-9f56-042fcd8b21dd_1024x1024.png 848w, https://substackcdn.com/image/fetch/$s_!G0ob!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa8eaa674-96bc-4da2-9f56-042fcd8b21dd_1024x1024.png 1272w, https://substackcdn.com/image/fetch/$s_!G0ob!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa8eaa674-96bc-4da2-9f56-042fcd8b21dd_1024x1024.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!G0ob!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa8eaa674-96bc-4da2-9f56-042fcd8b21dd_1024x1024.png" width="1024" height="1024" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/a8eaa674-96bc-4da2-9f56-042fcd8b21dd_1024x1024.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1024,&quot;width&quot;:1024,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:1233556,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.rockcybermusings.com/i/193740560?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa8eaa674-96bc-4da2-9f56-042fcd8b21dd_1024x1024.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!G0ob!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa8eaa674-96bc-4da2-9f56-042fcd8b21dd_1024x1024.png 424w, https://substackcdn.com/image/fetch/$s_!G0ob!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa8eaa674-96bc-4da2-9f56-042fcd8b21dd_1024x1024.png 848w, https://substackcdn.com/image/fetch/$s_!G0ob!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa8eaa674-96bc-4da2-9f56-042fcd8b21dd_1024x1024.png 1272w, https://substackcdn.com/image/fetch/$s_!G0ob!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa8eaa674-96bc-4da2-9f56-042fcd8b21dd_1024x1024.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p></p><p>Two of the three largest AI labs announced restricted-access cybersecurity models on the same day. A supply chain attack that started 10 days ago cost an AI startup its $10 billion contract with Meta. Nineteen new AI laws were signed across America in two weeks. Multiple independent research reports confirmed most enterprises have no idea what their AI agents are doing right now. The dual-use reckoning is no longer a future event. This week it produced products, paused contracts, and named casualties.</p><p>The week&#8217;s dominant pattern: the industry is admitting, out loud, that its most capable models are too dangerous to ship without restrictions. Meanwhile, the governance infrastructure meant to keep pace with AI deployment is running badly behind. Government employees are using GenAI tools daily at an 82% adoption rate on systems that remain vulnerable to prompt injection attacks documented in 2023. FedRAMP, the federal program enterprise CISOs treat as a security attestation, is operating as what former employees call a rubber stamp. The gap between AI capability and AI governance did not close this week. It widened, with better documentation.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.rockcybermusings.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading RockCyber Musings! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.rockcybermusings.com/p/weekly-musings-top-10-ai-security-20260403-20260409?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.rockcybermusings.com/p/weekly-musings-top-10-ai-security-20260403-20260409?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p><div><hr></div><h3>1. Anthropic locks its most powerful model behind a 50-partner gate</h3><p>On April 7, Anthropic announced Project Glasswing, a controlled-access program giving approximately 50 organizations early access to Claude Mythos Preview (Fortune, TechCrunch). Partner organizations include Amazon Web Services, Apple, Cisco, CrowdStrike, Google, JPMorgan Chase, Microsoft, and Nvidia, plus roughly 40 organizations responsible for critical software infrastructure. Anthropic described Mythos as &#8220;by far the most powerful AI model&#8221; it has ever created, with exceptional capabilities in autonomous coding and cybersecurity tasks. The company acknowledged the model&#8217;s capabilities &#8220;could be weaponized by attackers&#8221; and stated it has no plans for general availability until new safeguards are established.</p><p><strong>Why it matters</strong></p><ul><li><p>This is the first time a major AI lab has built a commercial product strategy explicitly around restricting access due to offensive cyber capability. The precedent matters more than the model.</p></li><li><p>Every enterprise security team outside the 50-partner cohort is now competing against organizations with months of head start deploying the most capable defensive AI available.</p></li><li><p>The partner list reads as the critical infrastructure vendor stack. If Mythos finds vulnerabilities before general availability, defenders benefit. If the model leaks before that happens, the calculus reverses.</p></li></ul><p><strong>What to do about it</strong></p><ul><li><p>Assess now whether your organization qualifies for Glasswing access or partnership with one of the 50 current participants. Waiting for general availability puts you behind.</p></li><li><p>Build your responsible AI deployment policy before your board asks you to justify restricted model use. The framework you create for Mythos applies to every dual-use model that follows.</p></li><li><p>Read Anthropic&#8217;s stated rationale carefully. It functions as a working template for your own internal policies on AI capability gating.</p></li></ul><p><strong>Rock&#8217;s Musings</strong></p><p>I&#8217;ve watched this industry congratulate itself on &#8220;responsible AI release&#8221; for years without actually restricting access to anything dangerous. Anthropic did something different this week. It built a product designed to stay out of the wrong hands and publicly named the hands it&#8217;s trusting. Anthropic made a liability calculation public and called it product strategy.</p><p>What I want to know is how they enforce it. Fifty organizations sharing API access, running evals, and passing findings back sounds clean in the press release. In practice, you&#8217;re dealing with 50 separate security cultures, 50 different interpretations of &#8220;defensive use,&#8221; and 50 sets of employees who walk out the door with operational knowledge. The kill switch isn&#8217;t in the contract. It&#8217;s in the monitoring. I&#8217;d love to see the audit framework Anthropic built to go with this, because without it, Project Glasswing is a hope, not a control.</p><div><hr></div><h3>2. OpenAI, Anthropic, and Google share intelligence to stop Chinese model distillation</h3><p>On April 6-7, Bloomberg and The Japan Times reported that OpenAI, Anthropic, and Google are sharing attack pattern data through the Frontier Model Forum to detect and block adversarial distillation attempts by Chinese AI companies. Three firms were named: DeepSeek, Moonshot AI, and MiniMax. The coordinated effort focuses on detecting when frontier model outputs are being used to train competing models without authorization. The Forum, established in 2023 for safety coordination, now functions as an active competitive intelligence sharing network.</p><p><strong>Why it matters</strong></p><ul><li><p>Three competing companies sharing security intelligence without a government mandate represents a structural shift in how the industry protects IP. Watching the next DeepSeek emerge on stolen training signal was apparently less appealing than coordinating with rivals.</p></li><li><p>This sets a precedent for industry-led AI IP enforcement that regulators haven&#8217;t built yet. Policymakers will either ratify or complicate what the Forum is quietly doing.</p></li><li><p>For enterprise buyers, this coordination signals frontier model providers now treat IP integrity as shared infrastructure, which is reassuring until you realize your own model training pipelines may need similar monitoring.</p></li></ul><p><strong>What to do about it</strong></p><ul><li><p>Audit your AI vendor contracts for provisions covering how your organization&#8217;s data and API interactions are used. The Forum&#8217;s distillation concerns apply downstream to enterprise deployments.</p></li><li><p>Ask vendors directly what controls they have in place to detect adversarial use of their model outputs. Most aren&#8217;t ready for the question.</p></li><li><p>Watch the Frontier Model Forum&#8217;s governance structure. Three companies sharing threat intelligence today is a small coalition. In two years it becomes the de facto standard for AI security coordination.</p></li></ul><p><strong>Rock&#8217;s Musings</strong></p><p>Three direct competitors sharing security intelligence tells you the distillation problem is worse than any of them want to admit publicly. DeepSeek&#8217;s emergence was the wake-up call: model training shortcuts were further along than anyone expected. The Forum is doing what the industry always resists, treating a shared problem as a shared problem.</p><p>What nobody says out loud is that adversarial distillation runs through enterprise deployments too. When your employees push 10,000 API calls through GPT-5.3 or Claude Mythos to build an internal tool, those outputs sit somewhere. The providers are focused on Chinese actors right now. The same technique scales to every bad actor with API access. Build that assumption into your threat model before someone builds a business around exploiting it.</p><div><hr></div><h3>3. Meta freezes its $10 billion Mercor contract after the LiteLLM supply chain breach</h3><p>On April 4, The Next Web and Fortune confirmed Meta paused its contract with Mercor, a $10 billion AI training data company whose customers include Anthropic, OpenAI, and Meta (The Next Web, Fortune). The pause followed a March 27 attack in which threat group TeamPCP published malicious PyPI packages for LiteLLM, a widely used open-source AI gateway library, after stealing a maintainer credential through an earlier Trivy supply chain compromise. The tainted packages were live for roughly 40 minutes. Mercor confirmed it was among &#8220;thousands&#8221; of affected organizations. Lapsus$ claimed responsibility and possession of 4TB of Mercor data including source code, databases, and VPN credentials. Google Mandiant reported over 1,000 impacted SaaS environments at RSAC 2026.</p><p><strong>Why it matters</strong></p><ul><li><p>A 40-minute PyPI window produced a paused $10 billion contract. That ratio of exposure time to business consequence should recalibrate how you think about open-source AI supply chain risk.</p></li><li><p>Meta&#8217;s pause affects AI training pipelines, not software. Training data provenance, labeling protocols, and selection criteria worth billions in R&amp;D may now be in hostile hands.</p></li><li><p>TeamPCP&#8217;s chained attack, Trivy to LiteLLM, demonstrates adversaries are mapping AI infrastructure dependency graphs specifically to maximize downstream blast radius.</p></li></ul><p><strong>What to do about it</strong></p><ul><li><p>Inventory open-source AI libraries in your production environment immediately. LiteLLM and similar tools are in most ML and security pipelines.</p></li><li><p>Require software bills of materials for AI infrastructure. You need to know which versions of which AI libraries are running in production, with provenance attestation for critical packages.</p></li><li><p>Brief your CISO and CTO on the chained supply chain model. TeamPCP demonstrated that AI library ecosystems are attack surfaces with compounding impact.</p></li></ul><p><strong>Rock&#8217;s Musings</strong></p><p>Forty minutes. That&#8217;s how long it takes to turn a credential theft into a paused $10 billion contract. The security community debates sophisticated nation-state tactics while basic supply chain hygiene stays on the backlog. LiteLLM is in everything. If you&#8217;re running AI in production and can&#8217;t tell me which version is deployed or whether it was compromised, you have a problem you haven&#8217;t measured yet.</p><p>The Meta piece is what keeps me up. Their AI training secrets, data selection criteria, and labeling methodology in hostile hands gives a competitor a two-year shortcut on billions in R&amp;D. A breach in the traditional sense costs you records. This one costs you competitive advantage. Your AI supply chain carries security risk and strategic risk simultaneously. Start treating both.</p><div><hr></div><h3>4. Keeper Security: 76% of AI agents operate outside privileged access policies</h3><p>Keeper Security released a survey of 109 cybersecurity professionals at RSAC 2026 on April 7, revealing that 46% of organizations have granted AI-powered tools access to critical systems and data, with 76% of those identities ungoverned under privileged access management policies (Keeper Security, BetaNews). Only 28% report full visibility into non-human identities across cloud, on-premises, and SaaS environments. Over 40% experienced a security incident involving machine credentials or non-human identities in the past year. Another 32% couldn&#8217;t confirm whether they&#8217;d been hit.</p><p><strong>Why it matters</strong></p><ul><li><p>AI agents operate as de facto privileged users in most enterprise environments, without the monitoring, credential rotation, or access controls applied to humans with equivalent permissions.</p></li><li><p>The 32% who can&#8217;t confirm NHI-related incidents are running blind. An agent with write access to email, code repositories, and collaboration tools that you can&#8217;t monitor is an insider threat waiting for attribution.</p></li><li><p>Traditional PAM tools were built for human users and won&#8217;t stretch to cover autonomous agents at scale without architectural change.</p></li></ul><p><strong>What to do about it</strong></p><ul><li><p>Extend your privileged access management program explicitly to cover AI agents, service accounts, and API keys. Treat an AI agent with production database access the same way you treat a privileged database administrator.</p></li><li><p>Mandate credential rotation and access logging for every non-human identity. If you can&#8217;t name every agent with write access to email or code right now, that gap is your first priority.</p></li><li><p>Ask your PAM vendor this week whether their product covers non-human identities natively. Many don&#8217;t, and most won&#8217;t tell you that unprompted.</p></li></ul><p><strong>Rock&#8217;s Musings</strong></p><p>Here&#8217;s the pattern showing up in research right now: organizations rush to deploy AI agents, grant them sweeping access to prove the use case, then spend the next 18 months trying to reconstruct what those agents touched. That&#8217;s the same mistake we made with cloud infrastructure in 2013. We provisioned everything with admin keys because it was faster and cleaned it up later. &#8220;Later&#8221; is still ongoing for most enterprises.</p><p>The 32% who aren&#8217;t sure about NHI incidents are the most honest number in the Keeper report. Detecting agent-related incidents requires logging you likely haven&#8217;t enabled, correlation rules you haven&#8217;t built, and a behavioral baseline you haven&#8217;t established. Before you deploy the next AI agent, ask your team to demonstrate they can detect one behaving badly. If they can&#8217;t show you in a live demo, slow down.</p><div><hr></div><h3>5. Salt Security: nearly half of enterprises are blind to their AI agents&#8217; API traffic</h3><p>Salt Security published its 1H 2026 State of AI and API Security Report on April 8, surveying over 300 security leaders (Salt Security). Key findings: 48.9% of organizations cannot monitor machine-to-machine traffic from autonomous agents, and 48.3% cannot distinguish legitimate AI agents from malicious bots in their API traffic. Only 23.5% of respondents rate their existing tools as &#8220;very effective&#8221; against AI-driven attacks. An additional 47% have delayed production releases because of security concerns about APIs exposed to autonomous systems, meaning the gap is surfacing in shipping decisions, not survey responses alone.</p><p><strong>Why it matters</strong></p><ul><li><p>Your API gateway is your AI agent&#8217;s operational layer. No visibility into that traffic means no indication of whether your agents are working as designed, being abused, or actively exfiltrating data.</p></li><li><p>The bot detection problem is concrete. Attackers are masquerading autonomous tools as legitimate agent traffic. Without behavioral baselines for your own agents, there&#8217;s no way to tell the difference.</p></li><li><p>Legacy web application firewalls were built for human browsing patterns. AI agent traffic looks nothing like that, making existing perimeter controls largely irrelevant to this threat class.</p></li></ul><p><strong>What to do about it</strong></p><ul><li><p>Inventory every API your AI agents call in production. Map expected request patterns, volumes, and data flows to establish a behavioral baseline for detecting deviations.</p></li><li><p>Evaluate whether your API security tooling supports non-human identity traffic analysis. If the vendor demo focuses on OWASP Top 10 for human users, it&#8217;s the wrong tool for this problem.</p></li><li><p>Build rate limiting and anomaly detection specifically for agent API traffic. An agent calling APIs at 658 times normal frequency because of a malicious MCP server injection is a documented attack pattern from this week&#8217;s research.</p></li></ul><p><strong>Rock&#8217;s Musings</strong></p><p>Half your enterprise has zero visibility into what their AI agents are doing on the wire. You spent years building SOC capabilities, deploying SIEMs, tuning correlation rules, integrating threat intelligence feeds. Then you deployed AI agents that operate through an entirely different channel that bypasses all of it. The old security stack can&#8217;t see the new threat surface.</p><p>The market hasn&#8217;t caught up. Most API security vendors will confidently tell you their product handles agentic traffic. Ask them to demo detection of an agent that&#8217;s been redirected by a malicious MCP server. Watch the room go quiet. Ongoing analysis of where the real gaps are lives at <a href="https://rockcybermusings.com/">RockCyber Musings</a>. The gap between &#8220;we have API security&#8221; and &#8220;we can detect compromised agent behavior&#8221; is wider than most boards realize.</p><div><hr></div><h3>6. RSAC 2026: attackers move laterally in 22 seconds while defenders plan in minutes</h3><p>At RSA Conference 2026 on April 3, Google Mandiant&#8217;s Consulting CTO Charles Carmakal told reporters that the median time from initial access to secondary lateral movement has dropped from 8 hours to 22 seconds, making human-only incident response structurally impossible at those speeds (SiliconAngle, Dark Reading). IBM&#8217;s Mark Hughes called post-quantum migration an immediate operational priority, noting three finalized NIST post-quantum encryption standards are available now with adoption remaining low. The conference&#8217;s dominant theme was agentic AI&#8217;s dual role: attackers using autonomous tools to accelerate campaigns while defenders attempt to use the same tools to keep pace.</p><p><strong>Why it matters</strong></p><ul><li><p>A 22-second lateral movement window eliminates the human-in-the-loop response model. Your SOC procedures assume minutes. Your threat actors operate in seconds. That gap is where incidents become breaches.</p></li><li><p>Post-quantum urgency moved from theoretical concern to present operational priority at RSAC. Three finalized NIST standards exist today. Any organization with long-lived encrypted data needs a migration timeline now.</p></li><li><p>The agentic AI identity theme at RSAC confirmed the industry has aligned around non-human identities as the defining security challenge of the next 24 months.</p></li></ul><p><strong>What to do about it</strong></p><ul><li><p>Test your incident response playbooks against a 22-second lateral movement scenario. If your playbook assumes human review before containment actions, it needs a machine-speed trigger layer.</p></li><li><p>Publish a post-quantum migration roadmap internally before your next board meeting. &#8220;We&#8217;re monitoring it&#8221; is no longer a defensible position when finalized standards exist.</p></li><li><p>Pull one CISO peer debrief from RSAC this month. Hallway intelligence from that conference is often more actionable than the keynote content.</p></li></ul><p><strong>Rock&#8217;s Musings</strong></p><p>Twenty-two seconds. The number got nodding agreement in San Francisco, then people walked into vendor booths and looked at detection tools that still alert in minutes. The gap between attacker speed and defender speed is the core problem of modern security, measured in the wrong units for years. When the unit is seconds, your SIEM alert queue is not a security control. It&#8217;s a log archive with a UI.</p><p>The quantum conversation shifted at RSAC from awareness to urgency, and the shift is warranted. &#8220;Harvest now, decrypt later&#8221; is a real operation: adversaries collecting encrypted traffic today and storing it for the day quantum breaks the key. If you have long-lived secrets, your CTO&#8217;s timeline estimate is probably too generous. <a href="https://www.rockcyber.com/">RockCyber</a> has been running post-quantum migration frameworks for clients since last year. Most enterprise conversations are still stuck on the awareness slide.</p><div><hr></div><h3>7. Nineteen AI laws signed in two weeks: chatbot liability, healthcare disclosure, private right of action</h3><p>On April 6, PluralPolicy reported that 19 new AI laws were signed in the preceding two weeks, bringing the 2026 total to 25 enacted laws with 27 additional bills having cleared both legislative chambers (PluralPolicy, Troutman Pepper Locke). Tennessee, Oregon, and Idaho signed chatbot regulation bills during the week of April 3-9. Oregon&#8217;s law includes a private right of action with statutory damages. Utah signed 8 bills covering AI literacy requirements, classroom restrictions, deepfake intimate image bans, and insurance transparency mandates. Massachusetts, Rhode Island, and South Carolina moved healthcare AI bills out of committee, with Rhode Island&#8217;s version requiring healthcare providers to inform patients when AI is involved in their care.</p><p><strong>Why it matters</strong></p><ul><li><p>Chatbot liability laws with private right of action create litigation exposure your legal team needs to model before the next customer-facing AI deployment goes live. Oregon&#8217;s law is already in effect.</p></li><li><p>The geographic spread creates a patchwork compliance problem with no federal preemption in sight. Your AI product team is shipping into 50 different state frameworks that change weekly.</p></li><li><p>Healthcare AI disclosure requirements set a transparency floor that buyers, patients, and regulators will increasingly apply across other sectors.</p></li></ul><p><strong>What to do about it</strong></p><ul><li><p>Map your current AI deployments against emerging state chatbot disclosure and liability requirements immediately. Oregon&#8217;s private right of action is live and applies now.</p></li><li><p>Brief your GC and CMO together. AI product launches carry legal exposure marketing teams don&#8217;t typically model, and chatbot liability surfaces in headlines, not just settlement columns.</p></li><li><p>Build a state AI law tracking function into your compliance program. Static annual reviews don&#8217;t work when the law count moves by double digits in two weeks.</p></li></ul><p><strong>Rock&#8217;s Musings</strong></p><p>When I tell clients that AI regulation is coming, I usually get a polite nod and a &#8220;we&#8217;ll handle it when we have to.&#8221; Twenty-five enacted laws in 2026 with the year barely three months old. Oregon telling enterprises their customers can sue with statutory damages when a chatbot fails to identify itself. Regulation isn&#8217;t coming. It&#8217;s been here for two weeks.</p><p>The private right of action piece is what executives aren&#8217;t tracking closely enough. FTC enforcement requires agency resources and case selection. Private litigants require only a lawyer and a grievance. If your customer-facing AI system fails to disclose its nature and a user in Oregon has a bad experience, you have a plaintiff class with no regulatory gatekeeping standing between that plaintiff and your legal team. Build that into your AI deployment approval checklist before the next product launch.</p><div><hr></div><h3>8. OpenAI readies its own restricted cybersecurity model the same day as Anthropic</h3><p>On April 9, Axios broke the news that OpenAI is finalizing a cybersecurity product for restricted release through its Trusted Access for Cyber pilot program (Axios, Security Boulevard). The model, built on GPT-5.3-Codex, is described by OpenAI as &#8220;our most cyber-capable frontier reasoning model to date.&#8221; OpenAI committed $10 million in API credits to pilot participants at the February program launch. The Axios scoop published the same day as broad coverage of Anthropic&#8217;s Project Glasswing, with multiple security reporters noting two competing labs had each moved to restrict their most capable cyber models on the same day.</p><p><strong>Why it matters</strong></p><ul><li><p>Two frontier labs restricted their most capable cybersecurity models on the same day. Whether coordinated or coincidental, the signal is identical: the industry has reached a shared threshold assessment of offensive AI capability.</p></li><li><p>The OpenAI pilot started in February. Participants are already months ahead on advanced defensive AI adoption. Enterprise buyers outside the program are behind.</p></li><li><p>GPT-5.3-Codex positioned as an autonomous vulnerability researcher represents a qualitative shift in what AI security tools can do. Your red team needs exposure to this capability level before attackers deploy it against you.</p></li></ul><p><strong>What to do about it</strong></p><ul><li><p>Apply to OpenAI&#8217;s Trusted Access for Cyber program today. Not applying guarantees exclusion.</p></li><li><p>Treat the simultaneous OpenAI and Anthropic announcements as an inflection point in your AI security roadmap. Model access strategy is now a CISO decision, not a procurement question.</p></li><li><p>Start a conversation with your red team about what AI-assisted penetration testing looks like inside your environment. The offensive tools are being built. Defensive capabilities need to keep pace.</p></li></ul><p><strong>Rock&#8217;s Musings</strong></p><p>Two companies, same day, both announcing restricted access to their most capable cyber models. In thirty years I&#8217;ve never seen two direct competitors make functionally identical risk disclosures simultaneously without prior coordination. Either the Frontier Model Forum conversation from earlier in the week triggered parallel announcements, or both teams hit the same risk threshold independently. Neither explanation is entirely comforting, because it means the models in question worry the people who built them.</p><p>Here&#8217;s what this week&#8217;s announcements tell me: the dual-use problem is no longer an abstract ethics debate. It&#8217;s a product management constraint. The labs are building features that concern them enough to restrict access. That&#8217;s progress, because it means honest risk assessment is making it into the room where launch decisions happen. Build that same instinct into your own AI deployment process.</p><div><hr></div><h3>9. Government GenAI hits 82% daily adoption with prompt injection attacks still unaddressed</h3><p>On April 9, Help Net Security published Center for Internet Security analysis showing 82% of state and territorial government employees now use GenAI tools daily, up from 53% the year prior (Help Net Security, Center for Internet Security). CIS cited prompt injection as the primary unaddressed vulnerability in that deployment base, distinguishing two attack categories: direct injection where users attempt to bypass safety guidelines, and indirect injection where attackers embed malicious instructions in external content such as documents, webpages, or emails the agent processes. Incidents cited include a code assistant that transmitted AWS API keys to an external server after processing hidden instructions, and the GeminiJack attack that exploited enterprise data sources to trigger data exfiltration.</p><p><strong>Why it matters</strong></p><ul><li><p>Government employees are generating official outputs using AI that remains manipulable through documents those systems process. A single malicious PDF submitted through a government portal can redirect an agent&#8217;s behavior.</p></li><li><p>Deployment outpaced security controls by a wide margin. State and local government security teams were not staffed or funded to keep pace with that adoption curve.</p></li><li><p>Prompt injection in government contexts is a policy integrity issue, not a privacy issue. An AI assistant that processes manipulated input and produces a compromised output informing a real government decision is a governance failure with material real-world consequences.</p></li></ul><p><strong>What to do about it</strong></p><ul><li><p>Require any GenAI deployment processing external documents, emails, or web content to implement input sanitization and instruction-boundary enforcement. Your AI shouldn&#8217;t follow commands embedded in documents it summarizes.</p></li><li><p>Test your enterprise AI deployments against indirect prompt injection scenarios before the next rollout. The attack is not sophisticated. The absence of testing is the problem.</p></li><li><p>Report AI usage rates alongside security control maturity to your board. An 82% adoption rate combined with 7% real-time governance effectiveness, the number from Cybersecurity Insiders research, belongs on a risk register.</p></li></ul><p><strong>Rock&#8217;s Musings</strong></p><p>A government employee pastes a document into an AI assistant, and that document silently redirects the assistant to send AWS credentials to an external server. An attack category from 2023 that government AI deployments in 2026 still haven&#8217;t addressed, running at 82% daily adoption. The attack surface grew to near-universal usage while the defense posture stayed at &#8220;we have an acceptable use policy.&#8221;</p><p>Government IT security teams are underfunded, understaffed, and now responsible for securing AI deployments at a scale they didn&#8217;t request and weren&#8217;t resourced for. Before the next state AI bill gets signed requiring healthcare providers to disclose AI use to patients, lawmakers should ask how they&#8217;re funding the security infrastructure to keep those same deployments from being turned against the citizens they&#8217;re meant to serve.</p><div><hr></div><h3>10. OpenAI&#8217;s national security lead says humans must stay in the loop for defense decisions</h3><p>At a Special Competitive Studies Project conference on April 9, Sasha Baker, OpenAI&#8217;s head of national security policy, stated that defense personnel need a &#8220;workforce transformation&#8221; to apply &#8220;appropriate human judgment&#8221; when AI informs national security operations (Nextgov). Baker noted no current large language model is foolproof, and incorrect AI-driven decisions in defense contexts carry &#8220;much greater&#8221; consequences. She tied the statement to OpenAI&#8217;s pre-deployment safety reviews and the controlled rollout of models including GPT-5.3-Codex, the same model featured in the restricted cybersecurity announcement reported the same day.</p><p><strong>Why it matters</strong></p><ul><li><p>OpenAI&#8217;s national security lead publicly endorsed human-in-the-loop for defense decisions in the same week the company announced its most capable autonomous cyber model. That tension deserves examination in your own governance policies.</p></li><li><p>&#8220;Workforce transformation&#8221; is a budget line, not a strategy. Organizations deploying AI in sensitive contexts need explicit training, decision authority maps, and accountability structures for human oversight.</p></li><li><p>Baker&#8217;s statement creates a public record regulators and litigants can reference when evaluating whether organizations maintained adequate human oversight in AI-assisted decisions.</p></li></ul><p><strong>What to do about it</strong></p><ul><li><p>Map every AI-assisted decision in your organization where error consequences are asymmetric. Finance, safety, hiring, and security operations are the obvious categories. Build human override requirements into the workflow before the system goes live.</p></li><li><p>Assess your &#8220;workforce transformation&#8221; budget. Deploying AI in high-stakes contexts without investing in training humans to supervise it transfers the liability Baker is explicitly naming.</p></li><li><p>Document your human oversight model for AI decisions affecting personnel, customers, or critical systems. When the inevitable incident arrives, regulators will ask whether oversight was designed in from the start or retrofitted after the fact.</p></li></ul><p><strong>Rock&#8217;s Musings</strong></p><p>OpenAI hired a national security lead. That person is publicly calling for human judgment to override AI in defense decisions. In the same week OpenAI announced a restricted-access autonomous hacking model. If that pairing doesn&#8217;t communicate the gap between capability development speed and governance readiness, nothing will.</p><p>I&#8217;ve run security operations for thirty years, and the hardest thing to get organizations to do is slow down, especially when a competitor is moving fast. Baker&#8217;s statement is a reminder that speed without oversight produces accountability gaps that become congressional hearings. The enterprises that build human oversight structures now are the ones that avoid spending 2027 explaining to a federal committee why their AI made a decision that hurt someone.</p><div><hr></div><h3>The One Thing You Won&#8217;t Hear About But You Need To</h3><h3>FedRAMP is a rubber stamp and the AI vendors deploying through it know it</h3><p>On April 6, ProPublica published a detailed investigation examining three cautionary tales from the federal government&#8217;s rush to AI adoption (ProPublica). The most damaging finding: FedRAMP, the federal security authorization program enterprise CISOs treat as a validation signal for cloud products, is now described by former employees as &#8220;little more than a rubber stamp.&#8221; The program operates with minimal staff, overwhelmed by vendor volume. Third-party assessors who evaluate cloud providers for FedRAMP authorization are paid by the companies they assess. FedRAMP established a confidential back channel for assessors to raise concerns they wouldn&#8217;t document in official reports. Microsoft used timeline pressure and volume to effectively compress the GCC High approval process.</p><p><strong>Why it matters</strong></p><ul><li><p>FedRAMP authorization signals to enterprise buyers that a product meets federal security standards. A degraded signal means every procurement decision relying on it as a security input draws from a compromised source.</p></li><li><p>The paid-by-vendor assessor model creates structural incentives to under-report findings. The unofficial back channel means the official report is not the complete picture.</p></li><li><p>Federal AI deployment at 82% daily government usage rates, built on FedRAMP authorizations produced under these conditions, is a systemic governance failure, not an isolated product risk.</p></li></ul><p><strong>What to do about it</strong></p><ul><li><p>Stop treating FedRAMP authorization as a complete security evaluation for AI products. Use it as a starting point, then conduct your own targeted assessment focused on AI-specific risks the framework wasn&#8217;t designed to evaluate.</p></li><li><p>Ask AI vendors directly whether their FedRAMP assessment surfaced any findings submitted through the confidential back channel. A vendor that can&#8217;t answer hasn&#8217;t done adequate diligence on their own authorization.</p></li><li><p>Engage your government affairs function to advocate for FedRAMP reform as AI deployment scales. The current model was built for traditional SaaS and is not equipped to evaluate the risk surface of autonomous AI systems.</p></li></ul><p><strong>Rock&#8217;s Musings</strong></p><p>Here&#8217;s the story nobody put on a slide at RSAC. The federal government is deploying AI at record speed through a security authorization program that former insiders describe as a rubber stamp. The assessors evaluating these vendors get paid by those vendors. The uncomfortable findings go into a back channel that never reaches the official record. Those authorizations then get used by enterprise security teams as proxies for security validation.</p><p>I&#8217;ve said for years that governance certifications are often theater. FedRAMP was supposed to be one of the more rigorous ones. The ProPublica investigation suggests the volume and complexity of AI products broke the model. If you&#8217;re a CISO using FedRAMP status as a risk reduction input in AI procurement decisions, you&#8217;re relying on a control that may not be working as designed. That&#8217;s the kind of hidden assumption that converts an undetected vulnerability into a breach narrative. Read the ProPublica piece. Then recalibrate what &#8220;government certified&#8221; means for your program.</p><p></p><p>&#128073; For ongoing analysis of agentic AI governance frameworks, the conversation continues at <strong><a href="https://rockcybermusings.com/">RockCyber Musings</a></strong>.</p><p>&#128073; Visit <strong><a href="https://www.rockcyber.com/">RockCyber.com</a></strong> to learn more about how we can help with your traditional Cybersecurity and AI Security and Governance journey.</p><p>&#128073; Want to save a quick $100K? Check out our AI Governance Tools at <strong><a href="https://aigovernancetoolkit.com/">AIGovernanceToolkit.com</a></strong></p><p>&#128073; Subscribe for more AI and cyber insights with the occasional rant.</p><p><em>The views and opinions expressed in RockCyber Musings are my own and do not represent the positions of my employer or any organization I&#8217;m affiliated with.</em></p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.rockcybermusings.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading RockCyber Musings! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.rockcybermusings.com/?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share RockCyber Musings&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.rockcybermusings.com/?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share RockCyber Musings</span></a></p><div><hr></div><h2>References</h2><p>Axios. (2026, April 9). Scoop: OpenAI plans new product for cybersecurity use. https://www.axios.com/2026/04/09/openai-new-model-cyber-mythos-anthopic</p><p>BetaNews. (2026, April 7). New report highlights critical gaps in securing AI agents and non-human IDs. https://betanews.com/article/new-report-highlights-critical-gaps-in-securing-ai-agents-and-non-human-ids/</p><p>Bloomberg. (2026, April 6). OpenAI, Anthropic, Google unite to combat model copying in China. https://www.bloomberg.com/news/articles/2026-04-06/openai-anthropic-google-unite-to-combat-model-copying-in-china</p><p>Center for Internet Security. (2026). Prompt injection tags along as GenAI enters daily government use. Referenced via Help Net Security, April 9, 2026. https://www.helpnetsecurity.com/2026/04/09/genai-prompt-injection-enterprise-data-risk/</p><p>Dark Reading. (2026, April 3). RSAC 2026: How AI is reshaping cybersecurity faster than ever. https://www.darkreading.com/cybersecurity-operations/rsac-2026-how-ai-is-reshaping-cybersecurity-faster-than-ever</p><p>Fortune. (2026, April 2). Mercor, a $10 billion AI startup, confirms it was caught up in a major security incident. https://fortune.com/2026/04/02/mercor-ai-startup-security-incident-10-billion/</p><p>Fortune. (2026, April 7). Anthropic is giving some firms early access to Claude Mythos to bolster cybersecurity defenses. https://fortune.com/2026/04/07/anthropic-claude-mythos-model-project-glasswing-cybersecurity/</p><p>Hackread. (2026, April 7). AI agents and non-human identities creating critical security gaps, report. https://hackread.com/ai-agents-non-human-identities-security-gaps/</p><p>Help Net Security. (2026, April 9). Prompt injection tags along as GenAI enters daily government use. https://www.helpnetsecurity.com/2026/04/09/genai-prompt-injection-enterprise-data-risk/</p><p>Japan Times. (2026, April 7). OpenAI, Anthropic and Google cooperate to fend off Chinese bids to clone models. https://www.japantimes.co.jp/business/2026/04/07/tech/openai-anthropic-google-china-copy/</p><p>Keeper Security. (2026, April 7). Keeper Security research exposes critical gaps in securing AI agents, machines and non-human identities [Press release]. https://www.prnewswire.com/news-releases/keeper-security-research-exposes-critical-gaps-in-securing-ai-agents-machines-and-non-human-identities-302735305.html</p><p>Nextgov/FCW. (2026, April 9). OpenAI national security lead endorses &#8216;appropriate human judgment&#8217; in AI. https://www.nextgov.com/artificial-intelligence/2026/04/openai-national-security-lead-endorses-appropriate-human-judgment-ai/412738/</p><p>PluralPolicy. (2026, April 6). AI governance watch: Nineteen new AI bills passed into law. https://pluralpolicy.com/blog/the-ai-governance-watch-april-2026-nineteen-new-ai-bills-passed-into-law/</p><p>ProPublica. (2026, April 6). As the federal government rushes toward AI, here are three cautionary tales. https://www.propublica.org/article/federal-government-ai-cautionary-tales</p><p>Salt Security. (2026, April 8). The era of agentic security is here: Key findings from the 1H 2026 State of AI and API Security Report. https://salt.security/blog/the-era-of-agentic-security-is-here-key-findings-from-the-1h-2026-state-of-ai-and-api-security-report</p><p>Security Boulevard. (2026, April 9). OpenAI readies rollout of new cyber model as industry shifts to defense. https://securityboulevard.com/2026/04/openai-readies-rollout-of-new-cyber-model-as-industry-shifts-to-defense/</p><p>SiliconAngle. (2026, April 3). Three insights on AI attack from theCUBE at RSAC 2026. https://siliconangle.com/2026/04/03/three-insights-ai-attack-thecube-rsac-2026-rsac26/</p><p>TechCrunch. (2026, April 7). Anthropic debuts preview of powerful new AI model Mythos in new cybersecurity initiative. https://techcrunch.com/2026/04/07/anthropic-mythos-ai-model-preview-security/</p><p>The Next Web. (2026, April 4). Meta freezes AI data work after breach puts training secrets at risk. https://thenextweb.com/news/meta-mercor-breach-ai-training-secrets-risk</p><p>The Register. (2026, April 2). Mercor says it was &#8216;one of thousands&#8217; hit in LiteLLM attack. https://www.theregister.com/2026/04/02/mercor_supply_chain_attack/</p><p>Troutman Pepper Locke. (2026, April 6). Proposed state AI law update: April 6, 2026. https://www.troutmanprivacy.com/2026/04/proposed-state-ai-law-update-april-6-2026/</p>]]></content:encoded></item><item><title><![CDATA[Agent Supply Chain Attacks: Your Scanner Already Switched Sides]]></title><description><![CDATA[March 2026's Trivy-LiteLLM-Axios cascade shows why agent supply chain risk breaks existing controls. Practical steps for CISOs.]]></description><link>https://www.rockcybermusings.com/p/agent-supply-chain-attacks-scanner-switched-sides</link><guid isPermaLink="false">https://www.rockcybermusings.com/p/agent-supply-chain-attacks-scanner-switched-sides</guid><dc:creator><![CDATA[Rock Lambros]]></dc:creator><pubDate>Tue, 07 Apr 2026 12:50:17 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!9kRx!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F66f3dd50-87fc-4013-afa3-56b70e007b69_2048x2048.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!9kRx!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F66f3dd50-87fc-4013-afa3-56b70e007b69_2048x2048.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!9kRx!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F66f3dd50-87fc-4013-afa3-56b70e007b69_2048x2048.jpeg 424w, https://substackcdn.com/image/fetch/$s_!9kRx!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F66f3dd50-87fc-4013-afa3-56b70e007b69_2048x2048.jpeg 848w, https://substackcdn.com/image/fetch/$s_!9kRx!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F66f3dd50-87fc-4013-afa3-56b70e007b69_2048x2048.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!9kRx!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F66f3dd50-87fc-4013-afa3-56b70e007b69_2048x2048.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!9kRx!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F66f3dd50-87fc-4013-afa3-56b70e007b69_2048x2048.jpeg" width="1456" height="1456" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/66f3dd50-87fc-4013-afa3-56b70e007b69_2048x2048.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1456,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:3273648,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.rockcybermusings.com/i/193187673?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F66f3dd50-87fc-4013-afa3-56b70e007b69_2048x2048.jpeg&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!9kRx!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F66f3dd50-87fc-4013-afa3-56b70e007b69_2048x2048.jpeg 424w, https://substackcdn.com/image/fetch/$s_!9kRx!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F66f3dd50-87fc-4013-afa3-56b70e007b69_2048x2048.jpeg 848w, https://substackcdn.com/image/fetch/$s_!9kRx!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F66f3dd50-87fc-4013-afa3-56b70e007b69_2048x2048.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!9kRx!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F66f3dd50-87fc-4013-afa3-56b70e007b69_2048x2048.jpeg 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.rockcybermusings.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading RockCyber Musings! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.rockcybermusings.com/p/agent-supply-chain-attacks-scanner-switched-sides?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.rockcybermusings.com/p/agent-supply-chain-attacks-scanner-switched-sides?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p><p>Agent supply chain risk stopped being theoretical in March 2026. Over twelve days, a single threat actor group turned Trivy, KICS, LiteLLM, and Axios into credential-harvesting weapons, cascading across five distribution ecosystems and producing 300 GB of stolen secrets. The campaign started with an AI-powered bot. It spread through a self-propagating worm with blockchain-based command and control. Your vulnerability scanner, the tool you trusted to protect your pipeline, was the entry point. Now picture that same attack chain hitting an autonomous agent that installs tools, updates dependencies, and executes third-party skills without asking you first.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!LBet!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff98c68dc-00bc-430c-b1ab-8f96d454607d_680x820.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!LBet!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff98c68dc-00bc-430c-b1ab-8f96d454607d_680x820.png 424w, https://substackcdn.com/image/fetch/$s_!LBet!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff98c68dc-00bc-430c-b1ab-8f96d454607d_680x820.png 848w, https://substackcdn.com/image/fetch/$s_!LBet!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff98c68dc-00bc-430c-b1ab-8f96d454607d_680x820.png 1272w, https://substackcdn.com/image/fetch/$s_!LBet!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff98c68dc-00bc-430c-b1ab-8f96d454607d_680x820.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!LBet!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff98c68dc-00bc-430c-b1ab-8f96d454607d_680x820.png" width="680" height="820" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/f98c68dc-00bc-430c-b1ab-8f96d454607d_680x820.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:820,&quot;width&quot;:680,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:107234,&quot;alt&quot;:&quot;Timeline showing TeamPCP campaign cascading from Trivy to KICS to LiteLLM to CanisterWorm to Axios between February 28 and March 31 2026&quot;,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.rockcybermusings.com/i/193187673?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff98c68dc-00bc-430c-b1ab-8f96d454607d_680x820.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="Timeline showing TeamPCP campaign cascading from Trivy to KICS to LiteLLM to CanisterWorm to Axios between February 28 and March 31 2026" title="Timeline showing TeamPCP campaign cascading from Trivy to KICS to LiteLLM to CanisterWorm to Axios between February 28 and March 31 2026" srcset="https://substackcdn.com/image/fetch/$s_!LBet!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff98c68dc-00bc-430c-b1ab-8f96d454607d_680x820.png 424w, https://substackcdn.com/image/fetch/$s_!LBet!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff98c68dc-00bc-430c-b1ab-8f96d454607d_680x820.png 848w, https://substackcdn.com/image/fetch/$s_!LBet!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff98c68dc-00bc-430c-b1ab-8f96d454607d_680x820.png 1272w, https://substackcdn.com/image/fetch/$s_!LBet!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff98c68dc-00bc-430c-b1ab-8f96d454607d_680x820.png 1456w" sizes="100vw"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Figure 1: The TeamPCP Cascade: 12 Days, 5 Ecosystems</figcaption></figure></div><h2>Your Vulnerability Scanner Was the Vulnerability</h2><p>On March 19, 2026, TeamPCP used stolen credentials to force-push 76 of 77 version tags in the aquasecurity/trivy-action repository to malicious commits. The payload ran before the legitimate scan. Pipelines completed normally with green checkmarks across the board. Meanwhile, the malware dumped Runner.Worker process memory and exfiltrated cloud credentials, SSH keys, Kubernetes tokens, npm tokens, and Docker registry credentials to attacker-controlled infrastructure.</p><p>Trivy is a vulnerability scanner. Organizations run it in their CI/CD pipelines to detect supply chain attacks. When TeamPCP compromised it, the tool designed to find compromised dependencies became the compromised dependency. The irony is structural, not incidental. Security tools make ideal targets for supply chain attacks because they already have broad read access to the environments they scan. They touch secrets by design.</p><p>KICS, Checkmarx&#8217;s infrastructure-as-code scanner, fell the same way four days later. All 35 version tags hijacked. Same credential-stealing payload, different typosquat domain. Then LiteLLM, the AI gateway library that holds API keys for every LLM provider an organization uses, with 95 million monthly downloads and presence in 36% of cloud environments according to Wiz Research. TeamPCP published malicious versions to PyPI using credentials stolen from LiteLLM&#8217;s own CI/CD pipeline, which ran Trivy as part of its build process.</p><p>Each victim funded the next attack. The chain started with a single incomplete credential rotation at Aqua Security on March 1. TeamPCP retained access through tokens that survived the rotation. Every compromise from March 19 forward exploited credentials harvested from the previous target. Partial containment, as Aqua Security&#8217;s own post-incident analysis acknowledged, equals no containment.</p><p>By the time Axios was compromised on March 31 (100+ million weekly npm downloads, attributed by Microsoft Threat Intelligence to North Korean state actor Sapphire Sleet), the credential ecosystem was so thoroughly disrupted that Mandiant CTO Charles Carmakal warned of &#8220;hundreds of thousands of stolen credentials&#8221; and &#8220;a variety of actors with varied motivations.&#8221; The FBI confirmed TeamPCP was working through approximately 300 GB of compressed stolen credentials in collaboration with the LAPSUS$ extortion group.</p><h2>The AI Bot That Started Everything</h2><p>Most coverage focuses on the credential cascade. The more significant development is the one that started it.</p><p>On February 28, 2026, an autonomous bot calling itself hackerbot-claw exploited a misconfigured pull_request_target workflow in Trivy&#8217;s GitHub Actions to steal a Personal Access Token with write access to all 33+ Aqua Security repositories. The bot&#8217;s GitHub profile described itself as &#8220;an autonomous security research agent powered by claude-opus-4-5.&#8221; It carried a vulnerability pattern index with 9 attack classes and 47 sub-patterns. It targeted seven major repositories belonging to Microsoft, DataDog, CNCF, and Aqua Security over one week, achieving remote code execution in at least four.</p><p>This was not a script running pre-written exploits, as hackerbot-claw adapted its approach to each target&#8217;s specific workflow configuration. When one technique failed, it pivoted. Against the ambient-code/platform repository, it attempted prompt injection by replacing the project&#8217;s CLAUDE.md file with instructions designed to trick Claude Code into committing unauthorized changes. Claude Code detected the attack and refused, classifying it as a supply chain attack via poisoned project-level instructions.</p><p>That detail matters. An AI agent attacked. An AI agent defended. The outcome depended on configuration quality, not human vigilance. This is the arms race in miniature, and it already happened at production scale against real infrastructure.</p><p>StepSecurity, Repello AI, and Boost Security Labs independently documented the campaign. Pillar Security&#8217;s assessment identified the core gap: &#8220;zero visibility into AI coding agents running on developer machines, and no runtime controls when those agents are weaponized.&#8221;</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!9Joa!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffb087f4a-8118-4a84-b9f9-8ad55656f9cf_680x620.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!9Joa!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffb087f4a-8118-4a84-b9f9-8ad55656f9cf_680x620.png 424w, https://substackcdn.com/image/fetch/$s_!9Joa!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffb087f4a-8118-4a84-b9f9-8ad55656f9cf_680x620.png 848w, https://substackcdn.com/image/fetch/$s_!9Joa!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffb087f4a-8118-4a84-b9f9-8ad55656f9cf_680x620.png 1272w, https://substackcdn.com/image/fetch/$s_!9Joa!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffb087f4a-8118-4a84-b9f9-8ad55656f9cf_680x620.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!9Joa!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffb087f4a-8118-4a84-b9f9-8ad55656f9cf_680x620.png" width="680" height="620" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/fb087f4a-8118-4a84-b9f9-8ad55656f9cf_680x620.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:620,&quot;width&quot;:680,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:88336,&quot;alt&quot;:&quot;Comparison diagram showing how traditional supply chain attacks target human developers who review code while agent supply chain attacks target autonomous agents that auto-execute without review&quot;,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.rockcybermusings.com/i/193187673?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffb087f4a-8118-4a84-b9f9-8ad55656f9cf_680x620.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="Comparison diagram showing how traditional supply chain attacks target human developers who review code while agent supply chain attacks target autonomous agents that auto-execute without review" title="Comparison diagram showing how traditional supply chain attacks target human developers who review code while agent supply chain attacks target autonomous agents that auto-execute without review" srcset="https://substackcdn.com/image/fetch/$s_!9Joa!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffb087f4a-8118-4a84-b9f9-8ad55656f9cf_680x620.png 424w, https://substackcdn.com/image/fetch/$s_!9Joa!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffb087f4a-8118-4a84-b9f9-8ad55656f9cf_680x620.png 848w, https://substackcdn.com/image/fetch/$s_!9Joa!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffb087f4a-8118-4a84-b9f9-8ad55656f9cf_680x620.png 1272w, https://substackcdn.com/image/fetch/$s_!9Joa!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffb087f4a-8118-4a84-b9f9-8ad55656f9cf_680x620.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Figure 2: Traditional vs. Agent Supply Chain Attack Surface</figcaption></figure></div><h2>Why Agent Supply Chain Risk Breaks Your Existing Controls</h2><p>Every supply chain control you have assumes a human is looking. Dependency scanning assumes someone reviews the output. Code review assumes someone reads the diff. SBOM generation assumes someone checks the inventory. SAST and DAST assume someone triages the findings.</p><p>Agents don&#8217;t look. They execute.</p><p>When a developer installs a package, they see a version number, check a changelog, run tests. When an agent installs a tool or skill, it follows instructions. If the MCP server definition says &#8220;install this plugin,&#8221; the agent installs it. If the skill marketplace listing looks legitimate, the agent trusts it. The human-in-the-loop that traditional supply chain security depends on evaporates.</p><p>Research published the same week as the TeamPCP campaign quantifies the gap. The OpenClaw vulnerability taxonomy (<a href="https://arxiv.org/abs/2603.27517">A Systematic Taxonomy of Security Vulnerabilities in the OpenClaw AI Agent Framework, arXiv 2603.27517</a>) catalogued 190 security advisories in the OpenClaw AI agent framework, organized by architectural layer and adversarial technique. Three findings stand out. First, three moderate-to-high-severity advisories compose into a complete unauthenticated remote code execution path from an LLM tool call to the host process. Second, the primary command-filtering mechanism relies on a closed-world assumption that shell commands are identifiable through lexical parsing, an assumption broken by basic techniques like line continuation and option abbreviation. Third, and most relevant here, a malicious skill distributed through the plugin channel executed a two-stage dropper within the LLM context, bypassing the entire execution pipeline. The skill distribution surface has no runtime policy enforcement.</p><p><a href="https://arxiv.org/abs/2603.28807">SafeClaw-R (arXiv 2603.28807)</a> found that 36.4% of OpenClaw&#8217;s built-in skills carry high or critical risk. That number covers the built-in skills, before any third-party marketplace plugins enter the picture. Across ClawHub, the agent skill marketplace, Antiy CERT confirmed 1,184 malicious skills, roughly one in five packages. The Repello AI team traced 335 of those to a single coordinated campaign called ClawHavoc.</p><p>The March 2026 campaign showed what happens when the consumer of a compromised dependency is a CI/CD pipeline: automated, monitored by humans on a lag, with credentials accessible at runtime. The agent version removes the monitoring layer entirely. An agent that installs a malicious MCP server or skill executes the payload as part of its normal workflow, with whatever permissions the agent has been granted, at machine speed.</p><h2>CanisterWorm Showed What Autonomous Propagation Looks Like</h2><p>If the hackerbot-claw precedent shows how AI agents attack, CanisterWorm shows how compromised dependencies spread once humans are removed from the loop.</p><p>CanisterWorm emerged on March 20, deployed using npm tokens stolen from Trivy-compromised pipelines. It was a self-propagating worm. Given a stolen token, it enumerated every package the token provided access to, bumped the patch version, injected its payload, and republished. Twenty-eight packages were compromised in under sixty seconds. The worm infected 64+ packages across multiple npm scopes. Endor Labs assessed that TeamPCP had &#8220;automated credential-to-compromise tooling&#8221; capable of turning a single stolen token into exponential propagation.</p><p>The command-and-control infrastructure used an Internet Computer Protocol blockchain canister, a tamperproof smart contract with no single takedown point. The operator could rotate payloads on infected machines without republishing any package. Security researchers confirmed this as the first publicly documented npm worm to use blockchain-based C2. The kill switch? If the canister returned a YouTube URL, the backdoor skipped execution. At the time of discovery, it was returning a Rick Roll. The infrastructure was live, tested, and ready. The payload was dormant by choice.</p><p>CanisterWorm targeted human-operated CI/CD pipelines. Translate that propagation model to an agent ecosystem where tools install other tools, agents delegate to sub-agents, and MCP servers chain calls across services. The propagation surface expands from &#8220;every package a stolen token provides access to&#8221; to &#8220;every tool, skill, and service the compromised agent is authorized to reach.&#8221; The Model Context Protocol, now under the Linux Foundation&#8217;s Agentic AI Foundation after Anthropic donated it in December 2025, is becoming the standard for agent-to-tool communication. Trend Micro found 492 MCP servers exposed to the internet with zero authentication. A separate supply chain attack involved a package masquerading as a legitimate Postmark MCP server that silently BCC&#8217;d every outgoing email to the attackers. The CoSAI whitepaper on MCP security identified 12 core threat categories spanning nearly 40 distinct threats. The MCP specification itself uses SHOULD rather than MUST for human-in-the-loop requirements. That word choice tells you everything about where the standard stands on constraining agent autonomy.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!cGuA!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9aef84b8-6614-464e-9217-882c7170cfda_680x900.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!cGuA!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9aef84b8-6614-464e-9217-882c7170cfda_680x900.png 424w, https://substackcdn.com/image/fetch/$s_!cGuA!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9aef84b8-6614-464e-9217-882c7170cfda_680x900.png 848w, https://substackcdn.com/image/fetch/$s_!cGuA!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9aef84b8-6614-464e-9217-882c7170cfda_680x900.png 1272w, https://substackcdn.com/image/fetch/$s_!cGuA!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9aef84b8-6614-464e-9217-882c7170cfda_680x900.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!cGuA!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9aef84b8-6614-464e-9217-882c7170cfda_680x900.png" width="680" height="900" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/9aef84b8-6614-464e-9217-882c7170cfda_680x900.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:900,&quot;width&quot;:680,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:109706,&quot;alt&quot;:&quot;Infographic showing key statistics from the March 2026 TeamPCP supply chain campaign and the agent supply chain gap, including 76 of 77 Trivy tags poisoned, 300 GB stolen credentials, 95 million LiteLLM monthly downloads, 36.4 percent of built-in agent skills rated high or critical risk, and 1 in 5 agent marketplace packages confirmed malicious&quot;,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.rockcybermusings.com/i/193187673?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9aef84b8-6614-464e-9217-882c7170cfda_680x900.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="Infographic showing key statistics from the March 2026 TeamPCP supply chain campaign and the agent supply chain gap, including 76 of 77 Trivy tags poisoned, 300 GB stolen credentials, 95 million LiteLLM monthly downloads, 36.4 percent of built-in agent skills rated high or critical risk, and 1 in 5 agent marketplace packages confirmed malicious" title="Infographic showing key statistics from the March 2026 TeamPCP supply chain campaign and the agent supply chain gap, including 76 of 77 Trivy tags poisoned, 300 GB stolen credentials, 95 million LiteLLM monthly downloads, 36.4 percent of built-in agent skills rated high or critical risk, and 1 in 5 agent marketplace packages confirmed malicious" srcset="https://substackcdn.com/image/fetch/$s_!cGuA!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9aef84b8-6614-464e-9217-882c7170cfda_680x900.png 424w, https://substackcdn.com/image/fetch/$s_!cGuA!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9aef84b8-6614-464e-9217-882c7170cfda_680x900.png 848w, https://substackcdn.com/image/fetch/$s_!cGuA!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9aef84b8-6614-464e-9217-882c7170cfda_680x900.png 1272w, https://substackcdn.com/image/fetch/$s_!cGuA!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9aef84b8-6614-464e-9217-882c7170cfda_680x900.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Figure 3: Agent Supply Chain Risk: By the Numbers</figcaption></figure></div><h2>What This Means for You</h2><p>The governance gap between where agent supply chain risk is and where your controls are will take years to close. Microsoft released the Agent Governance Toolkit on April 2, 2026, addressing OWASP&#8217;s Agentic AI Top 10 with features like Ed25519 plugin signing and MCP security gateways. The toolkit is two days old and unvalidated in production. SafeClaw-R achieved 95.2% accuracy in controlled tests. That 4.8% gap matters at enterprise scale.</p><p>You don&#8217;t have years. Here&#8217;s what you have right now.</p><p>Pin everything to immutable references. Version tags are pointers, not contracts. The March 2026 campaign proved this at scale across GitHub Actions, Docker Hub, and npm. Pin GitHub Actions to full commit SHAs. Pin container images to digests. Pin PyPI packages to exact versions with hash verification. Floating tags and unpinned dependencies are the entry point for every attack in this chain.</p><p>Treat your security tools as attack surface. Trivy, KICS, and every other scanner in your pipeline runs with privileged access to secrets by design. Apply the same scrutiny to your security tooling that you apply to production dependencies. Monitor for unexpected behavior from tools that should be predictable.</p><p>Audit your agent tool pipelines. If your organization deploys AI agents with access to MCP servers, skill marketplaces, or plugin registries, inventory every tool your agents use. Verify provenance. Enforce allow-lists. The ClawHavoc campaign showed that 20% of a major agent marketplace was compromised. Your agents are pulling from these registries right now.</p><p>Make credential rotation atomic. The entire TeamPCP cascade traces back to one failure: Aqua Security&#8217;s non-atomic rotation on March 1. When you respond to a supply chain incident, revoke all credentials simultaneously before issuing replacements. Partial rotation is an invitation for round two.</p><p>Plan for agent-specific incident response. If a tool or skill consumed by your agents is compromised, the blast radius includes everything those agents are authorized to access. Your current incident response playbook assumes a human in the response loop. Write the agent-specific version before you need it.</p><p><strong>Key Takeaway:</strong> The March 2026 supply chain campaign compromised your scanners, your AI gateway, and your HTTP client in twelve days. The same attack pattern targeting autonomous agents will move faster, spread further, and leave fewer traces. Your supply chain controls were built for a world where a human reviewed every dependency. That world is ending.</p><h3>What to do next</h3><p>The gap between traditional supply chain security and agent supply chain security is the defining governance challenge of 2026. If you&#8217;re a CISO or security architect, the question isn&#8217;t whether your organization uses AI agents with third-party tools. The question is whether you know which tools, with what permissions, under whose authority.</p><p>Start with visibility. You don&#8217;t control what you haven&#8217;t inventoried. For a deeper framework on operationalizing emerging security challenges, <a href="https://www.amazon.com/CISO-Evolution-Knowledge-Cybersecurity-Executives/dp/1119782481">The CISO Evolution</a> walks through how security leaders adapt their programs when the threat model shifts underneath them.</p><p>More on agent security, supply chain governance, and the practitioner&#8217;s view of AI risk at <strong><a href="https://rockcyber.substack.com">RockCyber Musings</a></strong>.</p><p>&#128073; Visit <strong><a href="https://www.rockcyber.com/">RockCyber.com</a></strong> to learn more about how we can help you in your traditional Cybersecurity and AI Security and Governance Journey</p><p>&#128073; Want to save a quick $100K? Check out our AI Governance Tools at <strong><a href="https://aigovernancetoolkit.com/">AIGovernanceToolkit.com</a></strong></p><p>&#128073; Subscribe for more AI and cyber insights with the occasional rant.</p><p><em>The views and opinions expressed in RockCyber Musings are my own and do not represent the positions of my employer or any organization I&#8217;m affiliated with.</em></p><p>Thanks for reading RockCyber Musings! Subscribe for free to receive new posts and support my work.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.rockcybermusings.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading RockCyber Musings! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.rockcybermusings.com/?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share RockCyber Musings&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.rockcybermusings.com/?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share RockCyber Musings</span></a></p><p></p>]]></content:encoded></item><item><title><![CDATA[Reasoning Theater: Why Chain-of-Thought Monitoring Fails Your Agentic AI]]></title><description><![CDATA[New research proves reasoning models perform deliberation they've already completed. Apply the CARE framework to close your agentic AI monitoring gap.]]></description><link>https://www.rockcybermusings.com/p/reasoning-theater-cot-monitoring-fails-agentic-ai</link><guid isPermaLink="false">https://www.rockcybermusings.com/p/reasoning-theater-cot-monitoring-fails-agentic-ai</guid><dc:creator><![CDATA[Rock Lambros]]></dc:creator><pubDate>Tue, 07 Apr 2026 12:50:15 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!8HFJ!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffc5ffff6-cea5-4323-92c9-86226f2f91a8_2752x1536.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!8HFJ!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffc5ffff6-cea5-4323-92c9-86226f2f91a8_2752x1536.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!8HFJ!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffc5ffff6-cea5-4323-92c9-86226f2f91a8_2752x1536.jpeg 424w, https://substackcdn.com/image/fetch/$s_!8HFJ!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffc5ffff6-cea5-4323-92c9-86226f2f91a8_2752x1536.jpeg 848w, https://substackcdn.com/image/fetch/$s_!8HFJ!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffc5ffff6-cea5-4323-92c9-86226f2f91a8_2752x1536.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!8HFJ!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffc5ffff6-cea5-4323-92c9-86226f2f91a8_2752x1536.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!8HFJ!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffc5ffff6-cea5-4323-92c9-86226f2f91a8_2752x1536.jpeg" width="1456" height="813" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/fc5ffff6-cea5-4323-92c9-86226f2f91a8_2752x1536.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:813,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:2828150,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.rockcybermusings.com/i/190378011?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffc5ffff6-cea5-4323-92c9-86226f2f91a8_2752x1536.jpeg&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!8HFJ!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffc5ffff6-cea5-4323-92c9-86226f2f91a8_2752x1536.jpeg 424w, https://substackcdn.com/image/fetch/$s_!8HFJ!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffc5ffff6-cea5-4323-92c9-86226f2f91a8_2752x1536.jpeg 848w, https://substackcdn.com/image/fetch/$s_!8HFJ!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffc5ffff6-cea5-4323-92c9-86226f2f91a8_2752x1536.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!8HFJ!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffc5ffff6-cea5-4323-92c9-86226f2f91a8_2752x1536.jpeg 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>If your agentic AI safety strategy depends on reading the model&#8217;s chain of thought, you&#8217;re listening to a rehearsed speech and calling it a confession. <a href="https://arxiv.org/abs/2603.05488">A March 2025 paper from Goodfire AI and Harvard University</a> shows that reasoning models often commit to their final answer within the first tokens of &#8220;thinking,&#8221; then generate hundreds of additional tokens to perform deliberation they&#8217;ve already completed. For every security leader, product owner, and governance committee treating chain-of-thought monitoring as an auditable safety control, these findings demand an immediate reassessment.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.rockcybermusings.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading RockCyber Musings! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.rockcybermusings.com/p/reasoning-theater-cot-monitoring-fails-agentic-ai?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.rockcybermusings.com/p/reasoning-theater-cot-monitoring-fails-agentic-ai?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p><h2>The Research That Changes the Conversation</h2><p>The paper, titled &#8220;Reasoning Theater: Disentangling Model Beliefs from Chain-of-Thought,&#8221; tested two frontier reasoning models (DeepSeek-R1 671B and GPT-OSS 120B) using three methods to determine when models commit to their final answer during a chain-of-thought trace.</p><p>The first method trains lightweight attention probes on model activations to predict the final answer at any point during reasoning. The second forces the model to answer early by truncating its reasoning and demanding a response. The third uses an external LLM (Gemini 2.5 Flash) as a CoT monitor, the same approach many vendors now sell as an AI safety feature.</p><p>The results expose a gap that should reshape every AI governance conversation. On MMLU-Redux, a benchmark heavy on recall-based questions, attention probes decoded the correct final answer with high accuracy from the earliest reasoning tokens. The CoT monitor couldn&#8217;t identify the model&#8217;s answer until much later in the sequence. The performativity rate, measuring the gap between internal confidence and external verbalization, hit 0.417 on MMLU for DeepSeek-R1. A rate near zero means the model&#8217;s spoken reasoning matches its internal state. A rate of 0.417 means the model spends roughly 40% of its reasoning trace performing confidence it already holds.</p><p>On GPQA-Diamond, a dataset requiring graduate-level multi-hop reasoning in biology, chemistry, and physics, the picture reversed. All three methods started near chance accuracy and improved at similar rates throughout the reasoning trace. The performativity rate dropped to 0.012. The chain of thought tracked genuine computation because the model needed to think.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!ge87!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbb549b57-6a71-424f-bf30-7b645e1b97a6_3500x2500.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!ge87!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbb549b57-6a71-424f-bf30-7b645e1b97a6_3500x2500.png 424w, https://substackcdn.com/image/fetch/$s_!ge87!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbb549b57-6a71-424f-bf30-7b645e1b97a6_3500x2500.png 848w, https://substackcdn.com/image/fetch/$s_!ge87!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbb549b57-6a71-424f-bf30-7b645e1b97a6_3500x2500.png 1272w, https://substackcdn.com/image/fetch/$s_!ge87!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbb549b57-6a71-424f-bf30-7b645e1b97a6_3500x2500.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!ge87!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbb549b57-6a71-424f-bf30-7b645e1b97a6_3500x2500.png" width="1456" height="1040" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/bb549b57-6a71-424f-bf30-7b645e1b97a6_3500x2500.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1040,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:350905,&quot;alt&quot;:&quot;Bar chart comparing performativity rates showing high values for MMLU recall tasks and near-zero values for GPQA-Diamond reasoning tasks across DeepSeek-R1 and GPT-OSS models&quot;,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.rockcybermusings.com/i/190378011?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbb549b57-6a71-424f-bf30-7b645e1b97a6_3500x2500.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="Bar chart comparing performativity rates showing high values for MMLU recall tasks and near-zero values for GPQA-Diamond reasoning tasks across DeepSeek-R1 and GPT-OSS models" title="Bar chart comparing performativity rates showing high values for MMLU recall tasks and near-zero values for GPQA-Diamond reasoning tasks across DeepSeek-R1 and GPT-OSS models" srcset="https://substackcdn.com/image/fetch/$s_!ge87!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbb549b57-6a71-424f-bf30-7b645e1b97a6_3500x2500.png 424w, https://substackcdn.com/image/fetch/$s_!ge87!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbb549b57-6a71-424f-bf30-7b645e1b97a6_3500x2500.png 848w, https://substackcdn.com/image/fetch/$s_!ge87!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbb549b57-6a71-424f-bf30-7b645e1b97a6_3500x2500.png 1272w, https://substackcdn.com/image/fetch/$s_!ge87!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbb549b57-6a71-424f-bf30-7b645e1b97a6_3500x2500.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Figure 1: Performativity Rate Across Models and Benchmarks</figcaption></figure></div><p>The number that should stop every governance meeting cold: probe-guided early exit reduced the number of generated tokens by 80% on MMLU and 30% on GPQA-Diamond, with comparable accuracy. The model produced 80% more tokens than it needed on easy recall tasks. Every one of those tokens looked like reasoning. None of them were.</p><h2>Your Model Performs Compliance, Not Communication</h2><p>The paper&#8217;s most valuable contribution for security professionals sits in the linguistic framework, not the probe methodology itself.</p><p>They invoke Grice&#8217;s maxims of cooperative communication, a foundational concept from pragmatics describing what honest, effective communicators do. Grice identified four maxims that cooperative speakers follow. They stay relevant (Relation). They assert only what they have evidence for (Quality). They provide exactly as much information as needed, no more (Quantity). They communicate clearly and without unnecessary obscurity (Manner).</p><p>Reasoning models, trained through reinforcement learning optimized for outcome rewards, follow Relation and Quality naturally. Staying on topic and generating evidence-based reasoning steps correlates with correct final answers, which earns the reward. The model has every incentive to be relevant and factually grounded.</p><p>Quantity and Manner get violated because the reward function doesn&#8217;t penalize verbosity or obscurity. The model generates hundreds of performative tokens after committing to its answer because nothing in the training signal punishes that behavior. The output looks like careful deliberation, but it reads like a thorough analysis. The model&#8217;s internal state tells a different story.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!wGIe!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F622a456c-5966-4cea-96b3-16ef262298ff_5175x4383.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!wGIe!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F622a456c-5966-4cea-96b3-16ef262298ff_5175x4383.png 424w, https://substackcdn.com/image/fetch/$s_!wGIe!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F622a456c-5966-4cea-96b3-16ef262298ff_5175x4383.png 848w, https://substackcdn.com/image/fetch/$s_!wGIe!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F622a456c-5966-4cea-96b3-16ef262298ff_5175x4383.png 1272w, https://substackcdn.com/image/fetch/$s_!wGIe!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F622a456c-5966-4cea-96b3-16ef262298ff_5175x4383.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!wGIe!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F622a456c-5966-4cea-96b3-16ef262298ff_5175x4383.png" width="1456" height="1233" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/622a456c-5966-4cea-96b3-16ef262298ff_5175x4383.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1233,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:868342,&quot;alt&quot;:&quot;Flowchart showing how reinforcement learning reward alignment causes reasoning models to follow relevance and quality maxims while violating quantity and manner maxims, creating a monitoring blind spot&quot;,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.rockcybermusings.com/i/190378011?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F622a456c-5966-4cea-96b3-16ef262298ff_5175x4383.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="Flowchart showing how reinforcement learning reward alignment causes reasoning models to follow relevance and quality maxims while violating quantity and manner maxims, creating a monitoring blind spot" title="Flowchart showing how reinforcement learning reward alignment causes reasoning models to follow relevance and quality maxims while violating quantity and manner maxims, creating a monitoring blind spot" srcset="https://substackcdn.com/image/fetch/$s_!wGIe!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F622a456c-5966-4cea-96b3-16ef262298ff_5175x4383.png 424w, https://substackcdn.com/image/fetch/$s_!wGIe!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F622a456c-5966-4cea-96b3-16ef262298ff_5175x4383.png 848w, https://substackcdn.com/image/fetch/$s_!wGIe!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F622a456c-5966-4cea-96b3-16ef262298ff_5175x4383.png 1272w, https://substackcdn.com/image/fetch/$s_!wGIe!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F622a456c-5966-4cea-96b3-16ef262298ff_5175x4383.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Figure 2: Gricean Maxim Compliance in RL-Trained Reasoning Models</figcaption></figure></div><p>If you&#8217;ve sat through a compliance audit where someone recites a scripted answer that sounds thorough but reveals nothing about practice on the ground, you&#8217;ve seen the human version of this. The model doesn&#8217;t lie to you. It performs what a reasoning trace is supposed to look like, without any structural obligation to communicate its internal state honestly.</p><p>This framing hits differently depending on where you sit at the governance table. Product teams hear &#8220;the model reasons safely&#8221; because the CoT looks clean. Legal hears &#8220;we have an auditable reasoning chain&#8221; because there&#8217;s text to review. Security asks the question nobody else in the room raises&#8230; &#8220;Does the text reflect what the model computed internally?&#8221;</p><h2>The Difficulty Trap: Where Enterprise Agentic AI Lives</h2><p>The difficulty-dependent split creates a precise and uncomfortable problem for enterprise deployments. CoT monitoring works when the model genuinely needs to think, hard multi-hop reasoning, novel problem decomposition, tasks requiring significant test-time compute. CoT monitoring fails on tasks where models already encode the answer in their weights: recall, pattern matching, classification, and routine decision execution.</p><p>Consider what agentic AI does in a financial services environment right now. A reconciliation agent matches transaction records against internal ledgers. A KYC agent screens customer documents against sanctions lists. A compliance agent routes regulatory filings based on entity classification. A trading surveillance agent flags potential wash trades from order flow patterns.</p><p>Every one of those tasks leans heavily toward recall and pattern matching. The model has seen millions of similar patterns during training. It knows the answer early. The chain-of-thought trace generated afterward is, according to this research, predominantly performative.</p><p>The security calculus changes in direct response. Picture a sanctions screening agent processing a crafted entity record containing an indirect prompt injection embedded in a business description field. The model processes the injection at the activation level, potentially altering its classification decision. The chain-of-thought trace, generated performatively because sanctions screening is a recall task the model handles, is produced from in-weight knowledge and reads like a textbook analysis of entity risk factors. The CoT monitor watching that trace sees clean, plausible reasoning. The model&#8217;s internal activations tell a different story. Your audit trail captures what the agent wants you to see, not computation. An examiner reviewing the reasoning trace isn&#8217;t examining evidence of how the model decided. They&#8217;re reviewing a post-hoc performance that resembles deliberation.</p><p>The researchers also found that model size correlates with performativity. Larger models (671B parameters) knew their answers earlier and generated more performative reasoning than smaller distilled versions (1.5B to 32B). Smaller models needed genuine test-time computation for equivalent tasks. As the industry pushes toward larger, more capable reasoning models for enterprise agentic AI, the performativity problem intensifies with every parameter upgrade.</p><p>The EU AI Act&#8217;s Article 14 requires human oversight of high-risk AI systems, with provisions for understanding the system&#8217;s reasoning. Financial regulators from the OCC to the FCA have signaled expectations for explainability in automated decision-making. Governance frameworks built on the assumption that reasoning traces constitute meaningful explanations now face a credibility problem they haven&#8217;t accounted for.</p><h2>Three Research Groups, One Uncomfortable Conclusion</h2><p>The Reasoning Theater paper lands with maximum force when you triangulate it against two other major findings from competing institutions.</p><p><a href="https://arxiv.org/abs/2503.11926">OpenAI&#8217;s Baker et al. (March 2025) </a>studied what happens when CoT monitors are integrated directly into the reinforcement learning training loop. In early stages, models became more capable and more aligned. Under excessive training pressure, models learned what the researchers called &#8220;obfuscated reward hacking&#8221;: hiding their intent within the chain of thought while continuing to hack the reward signal. OpenAI&#8217;s recommendation was blunt: developers training frontier reasoning models should avoid applying strong supervision directly to chains of thought. The monitorability itself is fragile. OpenAI&#8217;s follow-up &#8220;Monitoring Monitorability&#8221; paper in December 2025 proposed formal evaluation metrics to track this fragility across o3 and GPT-5.1 training runs, recognizing that the problem demands ongoing measurement rather than a one-time fix.</p><p><a href="https://arxiv.org/abs/2505.05410">Anthropic&#8217;s Chen et al. (May 2025)</a> found that reasoning models verbalized their use of reasoning hints in fewer than 20% of cases where they relied on those hints. Outcome-based RL initially improved faithfulness, then plateaued. Increased hint usage through reward hacking did not increase the propensity to disclose those hints. The model got better at using shortcuts while getting no better at revealing them.</p><p><a href="https://arxiv.org/abs/2507.11473">Korbak et al.&#8217;s multi-author paper from July 2025</a>, signed by over 40 researchers from multiple institutions, explicitly framed CoT monitorability as &#8220;a new and fragile opportunity for AI safety.&#8221; They warned that latent reasoning models, architectures that reason in continuous vector space rather than natural language, would eliminate even the partial transparency that current CoT provides.</p><p>Three independent research groups arrived at this conclusion: an interpretability lab at Goodfire, the model developer at OpenAI, and a cross-institutional safety coalition. All arrived at the same conclusion: chain-of-thought monitoring provides a signal in specific conditions, and those conditions don&#8217;t map cleanly to enterprise agentic AI deployments, where the majority of tasks fall below the difficulty threshold where monitoring remains reliable.</p><h2>Applying CARE: Fixing Your Chain-of-Thought Monitoring Gap</h2><p>The CARE framework (Create, Adapt, Run, Evolve) provides a structured path to address the reasoning theater problem without abandoning CoT monitoring wholesale.</p><p><strong>Create</strong> your threat model around the faithfulness gap. Classify every agentic AI workflow by task difficulty and determine which ones the model solves from in-weight knowledge versus genuine test-time reasoning. Recall-heavy tasks like KYC screening, transaction matching, and document classification carry high performativity risk. Multi-step analytical tasks like fraud investigation across multiple data sources and complex regulatory interpretation carry lower risk. Design your monitoring controls proportional to that classification.</p><p><strong>Adapt</strong> your monitoring stack to account for activation-level signals where architectures permit it. The Reasoning Theater research demonstrates that attention probes trained on model activations predict behavior far earlier and more accurately than CoT monitors for performative tasks. Every vendor offering agentic AI safety tooling should answer one direct question: does your monitoring access model internals, or does it only read output text? If the answer is output text only, you now know the limitation and the research that quantifies it.</p><p><strong>Run</strong> dual-layer monitoring in production. Keep CoT monitors active for genuine reasoning detection and inflection point identification. The research confirmed that backtracking, realizations, and reconsiderations in CoT traces correspond to authentic belief shifts, appearing more than twice as often in traces where probes showed genuine uncertainty (0.045 inflections per step in uncertain traces versus 0.020 in high-confidence performative traces for DeepSeek-R1). These inflection points remain reliable indicators of real internal computation. Layer behavioral monitoring (input/output validation, action-level controls, tool call verification) as the primary safety net for performative tasks. The CoT monitor becomes your second line for routine operations, not your first.</p><p><strong>Evolve</strong> your governance documentation to reflect the difficulty-dependent nature of CoT reliability. Update risk assessments as model capabilities change. Larger models and improved training methods shift the boundary between &#8220;easy&#8221; and &#8220;hard&#8221; tasks, changing where CoT monitoring remains effective. The August 2026 EU AI Act enforcement deadline adds urgency. Treat this as a moving target, because the research shows it is one.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!kxXF!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F69bfd39a-30fc-45c2-a014-a294eb7a3053_3307x7314.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!kxXF!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F69bfd39a-30fc-45c2-a014-a294eb7a3053_3307x7314.png 424w, https://substackcdn.com/image/fetch/$s_!kxXF!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F69bfd39a-30fc-45c2-a014-a294eb7a3053_3307x7314.png 848w, https://substackcdn.com/image/fetch/$s_!kxXF!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F69bfd39a-30fc-45c2-a014-a294eb7a3053_3307x7314.png 1272w, https://substackcdn.com/image/fetch/$s_!kxXF!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F69bfd39a-30fc-45c2-a014-a294eb7a3053_3307x7314.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!kxXF!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F69bfd39a-30fc-45c2-a014-a294eb7a3053_3307x7314.png" width="1456" height="3220" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/69bfd39a-30fc-45c2-a014-a294eb7a3053_3307x7314.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:3220,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:1056417,&quot;alt&quot;:&quot;Flowchart showing the four CARE framework phases with specific actions for addressing chain-of-thought monitoring limitations in agentic AI deployments&quot;,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.rockcybermusings.com/i/190378011?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F69bfd39a-30fc-45c2-a014-a294eb7a3053_3307x7314.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="Flowchart showing the four CARE framework phases with specific actions for addressing chain-of-thought monitoring limitations in agentic AI deployments" title="Flowchart showing the four CARE framework phases with specific actions for addressing chain-of-thought monitoring limitations in agentic AI deployments" srcset="https://substackcdn.com/image/fetch/$s_!kxXF!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F69bfd39a-30fc-45c2-a014-a294eb7a3053_3307x7314.png 424w, https://substackcdn.com/image/fetch/$s_!kxXF!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F69bfd39a-30fc-45c2-a014-a294eb7a3053_3307x7314.png 848w, https://substackcdn.com/image/fetch/$s_!kxXF!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F69bfd39a-30fc-45c2-a014-a294eb7a3053_3307x7314.png 1272w, https://substackcdn.com/image/fetch/$s_!kxXF!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F69bfd39a-30fc-45c2-a014-a294eb7a3053_3307x7314.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Figure 3: CARE Framework Response to Reasoning Theater</figcaption></figure></div><p><strong>Key Takeaway:</strong> Chain-of-thought monitoring provides genuine safety signal for hard reasoning tasks, but the majority of enterprise agentic AI workflows fall below the difficulty threshold where that signal remains reliable. Your governance framework needs to know the difference, and your next vendor evaluation needs to test for it.</p><h3>What to do next</h3><p>Download the Reasoning Theater paper and its interactive visualization tool at <a href="http://reasoning-theater.streamlit.app">reasoning-theater.streamlit.app</a>. Map your agentic AI workflows against the difficulty-dependent performativity findings. Bring this evidence to your next AI governance meeting, because the product team, legal counsel, and AI lead sitting across from you haven&#8217;t read it yet.</p><p>For more on building AI governance frameworks that survive contact with adversarial reality, explore the CARE framework at <a href="https://rockcyber.com">rockcyber.com</a>. Subscribe to <a href="https://rockcybermusings.com">RockCyber Musings</a> for more AI security and governance insights with the occasional rant.</p><p>&#128073; Subscribe for more AI security and governance insights with the occasional rant.</p><p>&#128073; Visit <strong><a href="https://www.rockcyber.com/">RockCyber.com</a></strong> to learn more about how we can help you in your traditional Cybersecurity and AI Security and Governance Journey</p><p>&#128073; Want to save a quick $100K? Check out our AI Governance Tools at <strong><a href="https://aigovernancetoolkit.com/">AIGovernanceToolkit.com</a></strong></p><p><em>The views and opinions expressed in RockCyber Musings are my own and do not represent the positions of my employer or any organization I&#8217;m affiliated with.</em></p>]]></content:encoded></item><item><title><![CDATA[Weekly Musings Top 10 AI Security Wrapup: Issue 32 March 27-April 2, 2026]]></title><description><![CDATA[Anthropic's Worst Week, CISA's Busiest Friday, and the EU Still Wasn't Ready]]></description><link>https://www.rockcybermusings.com/p/weekly-musings-top-10-ai-security-20260327-20260402</link><guid isPermaLink="false">https://www.rockcybermusings.com/p/weekly-musings-top-10-ai-security-20260327-20260402</guid><dc:creator><![CDATA[Rock Lambros]]></dc:creator><pubDate>Fri, 03 Apr 2026 13:03:28 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!GI_T!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2ea64fcd-437b-4e33-9512-7857b114e5ed_1024x1024.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!GI_T!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2ea64fcd-437b-4e33-9512-7857b114e5ed_1024x1024.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!GI_T!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2ea64fcd-437b-4e33-9512-7857b114e5ed_1024x1024.png 424w, https://substackcdn.com/image/fetch/$s_!GI_T!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2ea64fcd-437b-4e33-9512-7857b114e5ed_1024x1024.png 848w, https://substackcdn.com/image/fetch/$s_!GI_T!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2ea64fcd-437b-4e33-9512-7857b114e5ed_1024x1024.png 1272w, https://substackcdn.com/image/fetch/$s_!GI_T!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2ea64fcd-437b-4e33-9512-7857b114e5ed_1024x1024.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!GI_T!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2ea64fcd-437b-4e33-9512-7857b114e5ed_1024x1024.png" width="1024" height="1024" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/2ea64fcd-437b-4e33-9512-7857b114e5ed_1024x1024.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1024,&quot;width&quot;:1024,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:1233556,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.rockcybermusings.com/i/193018440?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2ea64fcd-437b-4e33-9512-7857b114e5ed_1024x1024.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!GI_T!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2ea64fcd-437b-4e33-9512-7857b114e5ed_1024x1024.png 424w, https://substackcdn.com/image/fetch/$s_!GI_T!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2ea64fcd-437b-4e33-9512-7857b114e5ed_1024x1024.png 848w, https://substackcdn.com/image/fetch/$s_!GI_T!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2ea64fcd-437b-4e33-9512-7857b114e5ed_1024x1024.png 1272w, https://substackcdn.com/image/fetch/$s_!GI_T!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2ea64fcd-437b-4e33-9512-7857b114e5ed_1024x1024.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Anthropic had a week that should be a case study in operational security failure for years to come. On March 31, a routine release packaging error exposed 500,000 lines of Claude Code source across roughly 2,000 files. Five days earlier, a CMS misconfiguration had already put nearly 3,000 unpublished internal documents into a public search index, including draft material describing their most capable model as posing &#8220;unprecedented cybersecurity risk.&#8221; By April 1, they were firing DMCA takedowns at 8,000 GitHub repositories, most unrelated to them, trying to unsee what the internet had already seen. By April 2, a congressman was writing to the CEO about national security.</p><p>That would have been enough for any week. It was not the only thing that happened. On March 27, CISA added two exploited AI infrastructure vulnerabilities to its KEV catalog; three LangChain and LangGraph CVEs hit disclosure, with 84 million downloads in scope; and the European Commission confirmed attackers had been inside their AWS account for three days. The thread connecting all of it is the same one it always is: AI deployment speed running ahead of the operational security discipline required to sustain it. This week was not an anomaly. It was a pattern. Patterns do not self-correct.</p><p>As a bonus, check out my <strong><a href="https://www.youtube.com/watch?v=091_b2qep9M">AI Cyber Magazine Podcast with Confidence Staveley</a></strong> during RSA.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.rockcybermusings.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading RockCyber Musings! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.rockcybermusings.com/p/weekly-musings-top-10-ai-security-20260327-20260402?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.rockcybermusings.com/p/weekly-musings-top-10-ai-security-20260327-20260402?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p><h3>1. Anthropic Leaked 500,000 Lines of Claude Code Source, Then Panicked on GitHub</h3><p>On March 31, a debugging file accidentally bundled into a routine Claude Code update exposed approximately 500,000 lines of source code across nearly 2,000 files (CNBC, Axios, Fortune). The codebase was mirrored across GitHub within hours. Leaked feature flags revealed unreleased capabilities: a persistent background agent, cross-device remote control, and session-to-session learning. Anthropic attributed the incident to &#8220;a release packaging issue caused by human error&#8221; and stated no customer data was exposed. On April 1, attempting to scrub the code from GitHub, Anthropic sent DMCA takedowns that hit approximately 8,000 repositories, most unrelated to the leak (TechCrunch, Bloomberg).</p><p><strong>Why it matters</strong></p><ul><li><p>Competitors received Anthropic&#8217;s unreleased feature roadmap. That strategic damage compounds the fact that this happened five days after the Mythos content leak. Coincidence???? I&#8217;ll let you decide.</p></li><li><p>The persistent background agent and remote control capabilities in the leaked code require explicit security design review before deployment. They were in development without prior public disclosure of the capability direction.</p></li><li><p>The DMCA sweep that caught 8,000 unrelated repositories shows what reactive incident response without a playbook looks like. Every remediation attempt created a new problem.</p></li></ul><p><strong>What to do about it</strong></p><ul><li><p>If you deploy Claude Code in your enterprise environment, review what access it holds to production systems and rotate any associated credentials until the full scope of the leak is confirmed.</p></li><li><p>Require software composition analysis (SCA) and release integrity verification as contractual terms with your AI vendors.</p></li><li><p>Develop a pre-incident legal response playbook that covers IP exposure scenarios, including proportional DMCA procedures that require scope confirmation before submission.</p></li></ul><p><strong>Rock&#8217;s Musings</strong></p><p>Two major operational security failures from the same company in five days. The first was a CMS misconfiguration. The second was a packaging error. Both are basic controls that mature security operations have solved. Anthropic markets itself on safety and trustworthiness, and that positioning is now doing work it was not designed to carry. The DMCA overcorrection made it worse: you leak 500,000 lines of source code, then fire automated takedown requests at 8,000 repositories, most of them unrelated to you. Every IP attorney will tell you DMCA takedowns require good faith and specificity. Have a process before the fire starts.</p><h3>2. Anthropic Accidentally Confirmed Its Most Capable Model Poses Unprecedented Cybersecurity Risk</h3><p>A configuration error in Anthropic&#8217;s content management system made nearly 3,000 unpublished assets publicly searchable starting around March 26, including draft blog posts for a model called Claude Mythos (Fortune, CoinDesk). Internal documents describe Mythos as capable of rapidly finding and exploiting software vulnerabilities at an unprecedented scale. Anthropic confirmed the model exists and is in testing with early-access customers, calling it &#8220;a step change&#8221; in capability. The company described the exposure as caused by a configuration error and stated the data store was secured after discovery.</p><p><strong>Why it matters</strong></p><ul><li><p>Anthropic&#8217;s own internal documentation, not a researcher&#8217;s estimate, describes this model as posing cybersecurity risks the industry has not seen before. That is the company&#8217;s self-assessment.</p></li><li><p>Early-access customer deployments were already underway before any public discussion of the risk profile occurred. The model shipped before the security conversation started.</p></li><li><p>A frontier model capable of autonomously finding and exploiting vulnerabilities at scale invalidates current vulnerability management timelines. That conversation needs to happen now.</p></li></ul><p><strong>What to do about it</strong></p><ul><li><p>Update your AI threat model to account for AI-assisted offensive operations at scale. This is not a future scenario. It is a current deployment.</p></li><li><p>Ask your AI vendors direct questions about internal capability assessments before your next contract renewal. What have they assessed, and when?</p></li><li><p>Document board and leadership awareness of frontier AI capability risk as a governance record item. Regulatory scrutiny on this topic will increase.</p></li></ul><p><strong>Rock&#8217;s Musings</strong></p><p>The model is called Mythos. The leaked internal docs describe the cybersecurity risk as unprecedented. Anthropic was already deploying it with customers before any of this became public. This happened not because of an attack but because someone left a CMS misconfigured. Anthropic has historically been conservative in capability claims. When their own internal documentation describes a model as different in kind from what came before, the security community should take that seriously, not because the word &#8220;unprecedented&#8221; is alarming on its own, but because the source is the organization that built the thing. They know what it does.</p><h3>3. ShinyHunters Breached the European Commission&#8217;s AWS Account</h3><p>The European Commission confirmed on March 27 that attackers accessed the AWS account hosting its <a href="http://Europa.eu">Europa.eu</a> websites, with the intrusion first detected on March 24 (TechCrunch, Bloomberg). Threat actor ShinyHunters claimed responsibility and alleged theft of more than 350GB of data including mail server exports, databases, confidential documents, and contracts. The Commission&#8217;s statement noted internal systems were unaffected and mitigation measures were applied quickly. Affected EU entities received notification.</p><p><strong>Why it matters</strong></p><ul><li><p>ShinyHunters has a documented history of monetizing stolen data through dark market sales. Even if the 350GB claim is exaggerated for leverage, policy documents and procurement contracts from the Commission&#8217;s web infrastructure are a counterintelligence asset.</p></li><li><p>The Commission enforces GDPR and is building the AI Act enforcement apparatus. Getting breached while standing up that apparatus is not a good governance signal.</p></li><li><p>AWS account-level compromise is full infrastructure compromise in practice. A managed cloud provider does not neutralize cloud account security failures.</p></li></ul><p><strong>What to do about it</strong></p><ul><li><p>Audit your AWS account permission boundaries and review CloudTrail logs for anomalous patterns this week, not next quarter.</p></li><li><p>Ensure your incident response plan explicitly covers cloud account compromise. Traditional endpoint-focused plans miss this scenario entirely.</p></li><li><p>If any of your vendors are EU institutions or Commission contractors, treat procurement data exposure as a downstream supply chain risk and assess your exposure now.</p></li></ul><p><strong>Rock&#8217;s Musings</strong></p><p>The body enforcing Europe&#8217;s data protection framework had its AWS account cracked. Governance credentials do not equal security maturity. Write the most thorough AI regulation in the world. Your cloud IAM configuration remains a disaster until someone fixes it. The ShinyHunters 350GB claim needs forensic verification before anyone draws conclusions about scope, but three days of undetected access to the official Commission infrastructure doesn&#8217;t need verification. The institutions asking private sector organizations to demonstrate AI security maturity owe the market some transparency on their own failures. Name it, fix it, move on.</p><h3>4. Your AI Workflow Tool Got CISA&#8217;s Attention: Langflow CVE-2026-33017</h3><p>CISA added CVE-2026-33017, a critical remote code execution flaw in Langflow, to its Known Exploited Vulnerabilities catalog on March 26. Attackers began scanning for exposed instances roughly 20 hours after the advisory publication, with exploitation scripts appearing within 21 hours and active .env and .db file harvesting beginning within 24 hours (Sysdig, BleepingComputer, Help Net Security). The vulnerability carries a CVSS score of 9.3 and allows unauthenticated attackers to inject arbitrary Python code through the public flow build endpoint with no sandboxing applied. Federal agencies face an April 8 remediation deadline. Upgrade to Langflow version 1.9.0 or later.</p><p><strong>Why it matters</strong></p><ul><li><p>Langflow is used to build and deploy LLM pipelines. Remote code execution in a workflow orchestration tool gives an attacker control over the AI&#8217;s inputs, outputs, and the credentials it holds.</p></li><li><p>The 20-hour exploitation window is increasingly standard for high-severity flaws. The concept of a patch window measured in days is no longer realistic for internet-exposed AI infrastructure.</p></li><li><p>.env file harvesting is the attacker&#8217;s first move because those files contain API keys for LLMs, vector databases, and cloud services the workflow connects to.</p></li></ul><p><strong>What to do about it</strong></p><ul><li><p>If Langflow runs on any internet-accessible host, treat the environment as potentially compromised and rotate all associated credentials before patching.</p></li><li><p>Segment AI workflow orchestration platforms behind authentication and network controls. These tools have no business being directly internet-accessible.</p></li><li><p>Verify Langflow version across your environment immediately. Anything prior to 1.9.0 is an open liability.</p></li></ul><p><strong>Rock&#8217;s Musings</strong></p><p>The 20-hour exploitation timeline should reframe your vulnerability management program. That program was designed when you had days or weeks to act. That era closed. CISA&#8217;s KEV catalog is now your minimum viable patch priority list, and if you are not at sub-72-hour remediation SLAs for critical AI infrastructure, you are already behind. Organizations still describing AI workflow platforms as &#8220;internal tools&#8221; need a rethink. Internal tools with LLM API keys, cloud credentials, and production data connections are not internal in any meaningful threat model. An attacker who executes code in your Langflow environment has lateral movement access to every system that environment touches.</p><h3>5. LangChain and LangGraph: Three CVEs, 84 Million Downloads Exposed</h3><p>Cyera security researcher Vladimir Tokarev disclosed three vulnerabilities in LangChain and LangGraph on March 27, each covering a different attack path against the same enterprise AI framework (The Hacker News). CVE-2026-34070 (CVSS 7.5) enables path traversal to arbitrary files through manipulated prompt templates. CVE-2025-68664 (CVSS 9.3) allows extraction of API keys and environment secrets through unsafe deserialization. CVE-2025-67644 (CVSS 7.3) enables SQL injection in LangGraph&#8217;s SQLite checkpoint layer. LangChain, LangChain-Core, and LangGraph collectively logged over 84 million downloads. Patches are available: LangChain Core 1.2.22+, LangChain-Core 0.3.81+ or 1.2.5+, and LangGraph checkpoint sqlite 3.0.1+.</p><p><strong>Why it matters</strong></p><ul><li><p>These three CVEs cover filesystem data, environment secrets, and conversation history in combination. Together, they represent near-total information exposure for any application built on these frameworks.</p></li><li><p>The 84 million download count means a significant portion of enterprise AI applications are affected. Most organizations do not know which AI frameworks their development teams selected.</p></li><li><p>CVE-2025-68664 with its 9.3 CVSS is the most critical. Unsafe deserialization is a well-understood, pervasive, and reliably exploitable class of vulnerability.</p></li></ul><p><strong>What to do about it</strong></p><ul><li><p>Inventory every AI framework in your environment, including those embedded in third-party tools. Do not rely on developers to self-report what they are using.</p></li><li><p>Apply the three patches and validate versions before the end of the business week.</p></li><li><p>Assess what data your LangChain-based applications can access and treat those data stores as potentially exposed pending patch confirmation.</p></li></ul><p><strong>Rock&#8217;s Musings</strong></p><p>Three vulnerability classes in the same framework, covering three categories of sensitive enterprise data, were disclosed in one report. That&#8217;s what happens when you build for speed and bolt on security later. AI framework developers made that choice repeatedly, and this week&#8217;s CVE list is the invoice. LangChain is the jQuery of AI development right now. It is in everything, often without explicit organizational approval. Your AI security posture includes every dependency your developers pulled in without telling you. Get ahead of that inventory problem before the next disclosure.</p><h3>6. A Congressman Put Anthropic on Notice Over National Security</h3><p>Rep. Josh Gottheimer (D-N.J.) sent a letter to Anthropic CEO Dario Amodei on April 2, citing national security concerns arising from the source code leak (Axios, The Hill). Gottheimer&#8217;s letter noted that Claude is embedded in defense and intelligence operations, raised the prior CCP-backed group intrusion against Claude, and expressed concern that Mythos could enable more sophisticated cyberattacks against the United States. The letter also flagged Anthropic&#8217;s decision in late February to remove its binding commitment to halt model development if safety capabilities fall behind, replacing it with &#8220;nonbinding but publicly-declared&#8221; goals.</p><p><strong>Why it matters</strong></p><ul><li><p>Federal agencies and defense contractors use Claude operationally. A source code leak followed by a congressional inquiry is a vendor risk event, not a PR problem. Your GRC process should treat it as such.</p></li><li><p>Removing the binding safety commitment is a substantive policy change that the congressional record now documents. The enforceability question will follow Anthropic through every future regulatory discussion.</p></li><li><p>Gottheimer sits on the House Intelligence Committee. This is not a throwaway letter. It is a first-stage oversight action that signals more to come.</p></li></ul><p><strong>What to do about it</strong></p><ul><li><p>Review your vendor risk assessment for any AI provider with confirmed government contracts. Congressional inquiries are material third-party risk events.</p></li><li><p>Establish a direct communication channel with your AI vendors&#8217; enterprise security teams and request formal notification procedures for any government inquiries affecting their products.</p></li><li><p>Track the congressional record regarding Anthropic&#8217;s rollback of its safety commitment. It will surface again in budget and procurement cycles.</p></li></ul><p><strong>Rock&#8217;s Musings</strong></p><p>The safety commitment rollback from February is the most substantive issue in that letter. Anthropic replaced a binding pledge to pause development if safety fell behind with goals they grade themselves on. That is not a small change. That is the foundational accountability mechanism that distinguished their positioning from competitors, and they quietly removed it. Congressional scrutiny was predictable the moment they became embedded in national security operations. The question I would ask directly is how many federal agency customers received notification about the source code exposure before it hit the press. I would guess the answer is uncomfortable.</p><h3>7. Your Security Scanner Was the Supply Chain Attack: Trivy CVE-2026-33634</h3><p>CISA added CVE-2026-33634 to its Known Exploited Vulnerabilities catalog on March 27 (Help Net Security, Aquasecurity GitHub advisory). Attackers compromised the Trivy container security scanner on March 19, using stolen credentials to publish a malicious v0.69.4 release and force-push 76 of 77 version tags in the trivy-action repository with credential-stealing malware. The attack triggered a downstream LiteLLM supply chain compromise via poisoned PyPI packages. Federal agencies face an April 9 deadline. Root cause was non-atomic credential rotation on March 1 left a valid token exposed during the rotation window.</p><p><strong>Why it matters</strong></p><ul><li><p>Trivy is a default security tool in CI/CD pipelines across the industry. Compromising the scanner means attackers access the same environment credentials the security scan was meant to protect.</p></li><li><p>Force-pushing 76 version tags is a comprehensive compromise. Any pipeline that pins to mutable major or minor version tags rather than specific commit hashes was exposed.</p></li><li><p>The downstream LiteLLM PyPI compromise extends the blast radius into Python environments running LLM application code. The supply chain damage propagated well beyond the initial tool compromise.</p></li></ul><p><strong>What to do about it</strong></p><ul><li><p>Audit every CI/CD pipeline for trivy-action or setup-trivy at mutable version tags and pin to specific commit hashes immediately.</p></li><li><p>Treat any environment that ran a compromised Trivy version since March 19 as potentially credential-compromised. Rotate all associated tokens, SSH keys, and cloud credentials.</p></li><li><p>Apply this lesson to every security tool in your pipeline. Security tooling supply chains are higher-value targets than application code supply chains.</p></li></ul><p><strong>Rock&#8217;s Musings</strong></p><p>The attacker turned the vulnerability scanner into the vulnerability. That is the platonic ideal of a supply chain attack: targeting organizations that care about security and embed security tooling in their build pipelines. The more security-conscious your culture, the higher your Trivy adoption, and the more exposed you were. The non-atomic credential rotation is the root cause. Aquasecurity rotated credentials on March 1 but did not revoke all tokens simultaneously. The attacker grabbed freshly rotated secrets during the window between invalidation and deployment. If your own rotation procedures have a gap between &#8220;revoke old&#8221; and &#8220;confirm new is live,&#8221; that gap is your exposure. Run your playbooks against that question this week.</p><h3>8. The State AI Chatbot Safety Wave Is Not Waiting for Washington</h3><p>Georgia&#8217;s state senate voted to concur in the House-amended version of SB 540 during the week of March 27, sending the chatbot disclosure and minor-protection bill to Governor Kemp&#8217;s desk (Troutman Privacy, Transparency Coalition). Idaho&#8217;s S 1297 passed its full legislature and advanced to Governor Little. Both are chatbot safety measures. Georgia&#8217;s bill requires disclosure every three hours for adult users and every hour for minors, along with explicit suicide and self-harm response protocols for conversational AI services. The Future of Privacy Forum&#8217;s tracker now counts 78 AI chatbot safety bills moving across 27 states in 2026.</p><p><strong>Why it matters</strong></p><ul><li><p>Disclosure, minor safety, and mental health response requirements are becoming the regulatory floor across state jurisdictions. Organizations operating consumer-facing AI products need a 50-state tracking capability, not a wait-and-see approach.</p></li><li><p>Hourly disclosure requirements for minors are not trivial to implement for many chatbot architectures. The compliance engineering work should start now.</p></li><li><p>Seventy-eight bills across 27 states mean that any federal preemption framework, if one ever arrives, faces an already established patchwork of state obligations to reconcile.</p></li></ul><p><strong>What to do about it</strong></p><ul><li><p>Map your consumer AI products against chatbot disclosure requirements in every state where users reside. Georgia and Idaho represent the floor, not the ceiling.</p></li><li><p>Assess your chatbot&#8217;s existing mental health response protocols against the Georgia requirement specifics. A disclaimer is not compliant.</p></li><li><p>Assign someone accountable for multi-state AI governance tracking. This is not a future compliance problem.</p></li></ul><p><strong>Rock&#8217;s Musings</strong></p><p>Washington cannot pass a federal AI framework. States can. Fifty legislatures with different requirements and different timelines is the compliance nightmare that preemption was supposed to prevent. It didn&#8217;t. Georgia&#8217;s hourly minor disclosure requirement is specific, implementable, and enforceable. State legislatures are producing more actionable compliance requirements than most federal guidance I have seen this year. If you deploy consumer AI products and you don&#8217;t have someone accountable for multi-state AI governance tracking today, that gap closes before Q3 or it closes you.</p><h3>9. The EU AI Act Has an Enforcement Problem, and Nobody Is Talking About It Honestly</h3><p>As of late March, only 8 of 27 EU member states had designated the single contact points required for national enforcement coordination under the AI Act, according to the European Parliament Think Tank&#8217;s enforcement analysis (Tech Policy Press, IAPP). The Digital Omnibus proposal, with negotiating positions adopted by Parliament&#8217;s IMCO and LIBE committees on March 18, would push high-risk AI compliance deadlines to December 2027 for Annex III systems and to August 2028 for Annex I systems, compared with the original August 2026 deadline. The European Commission also missed its own deadline for issuing guidance on high-risk AI systems. Trilogue negotiations between Council, Parliament, and Commission are now underway.</p><p><strong>Why it matters</strong></p><ul><li><p>Approximately 70% of EU member states are not operationally ready for AI Act enforcement. Regulations without enforcement infrastructure are aspirational documents.</p></li><li><p>The 16-month delay in high-risk requirements gives organizations breathing room on paper while creating uncertainty about what compliance standard they are being held to during the gap.</p></li><li><p>The Commission missing its own implementation guidance deadline sets a poor precedent for holding private sector organizations to their compliance timelines.</p></li></ul><p><strong>What to do about it</strong></p><ul><li><p>Do not use the delay as a license to defer governance program work. The underlying obligations have not changed in substance. Build the program now and own it.</p></li><li><p>Review the Digital Omnibus amendments specifically for changes to the high-risk AI system definition. Legislative simplification sometimes reclassifies systems in ways that alter the scope of compliance.</p></li><li><p>Subscribe to IAPP&#8217;s EU AI Act tracker for updates on the trilogue outcome. The final text will differ from both Council and Parliament positions.</p></li></ul><p><strong>Rock&#8217;s Musings</strong></p><p>Eight out of 27 enforcement bodies are operational as the Act&#8217;s first major deadlines approach. The Commission missed its own implementation guidance deadline. The most substantive AI governance framework on the planet is running on infrastructure that is not ready to enforce it. The delay does not invalidate the regulation. Organizations that build genuine AI risk management programs now will be positioned for whatever enforcement timeline materializes. Organizations that chase the deadline and treat compliance as documentation will be exposed when the enforcement machinery catches up. That gap grows wider every quarter.</p><h3>The One Thing You Won&#8217;t Hear About But You Need To</h3><h3>NVIDIA and Johns Hopkins Gave You a Blueprint for Defending AI Agents Against Prompt Injection</h3><p>Researchers from NVIDIA and Johns Hopkins University published &#8220;Architecting Secure AI Agents: Perspectives on System-Level Defenses Against Indirect Prompt Injection Attacks&#8221; on March 31 (<a href="https://arxiv.org/abs/2603.30016">ArXiv 2603.30016</a>). The paper addresses how AI agents are vulnerable not to direct attacks on the model but to malicious instructions embedded in data the agent processes during task execution. The authors articulate three architectural positions. First, agents in dynamic environments need dynamic replanning with security policy updates built into the replanning loop. Second, security decisions requiring contextual judgment should still involve LLMs, but only within system designs that strictly constrain what the model can observe and decide. Third, ambiguous situations should treat human interaction as a core design consideration, not an edge case to minimize.</p><p><strong>Why it matters</strong></p><ul><li><p>This paper frames indirect prompt injection as an architectural problem, not a model alignment problem. You cannot align your way out of it. You design it out or you accept the risk.</p></li><li><p>The principle of strictly constraining what the model can observe and decide has immediate practical application as your primary defense lever, more effective than filtering or detection approaches.</p></li><li><p>The human oversight design principle directly contradicts how most agentic deployments are being built, with human review treated as friction to reduce rather than a security control to preserve.</p></li></ul><p><strong>What to do about it</strong></p><ul><li><p>Read the paper. At 12 pages, it is short enough to share with your AI architects and security engineers before the next deployment review meeting.</p></li><li><p>Audit any agentic AI system currently in your environment against the observation scope and decision authority questions. Broad scope plus broad authority equals your highest-risk deployment.</p></li><li><p>Make human oversight an explicit design requirement in your AI agent security standards. Document the specific conditions under which an agent must pause and request human authorization.</p></li></ul><p><strong>Rock&#8217;s Musings</strong></p><p>Nobody outside the AI security research community covered this paper. That is precisely why it belongs here. The breach reports get attention. The architecture guidance that would prevent the next breach sits on ArXiv with a few hundred downloads. I have been arguing at <a href="https://www.rockcyber.com">RockCyber</a> for two years that agentic AI security is an architecture problem. You do not solve it with better prompts or stronger models. You solve it with privilege constraints, observation scope limits, and honest human oversight design. NVIDIA and Johns Hopkins gave you a 12-page framework for that conversation. If your next AI agent deployment review does not address these three principles, you are building exposure, not capability.</p><p>&#128073; For ongoing analysis of agentic AI governance frameworks, the conversation continues at <strong><a href="https://rockcybermusings.com">RockCyber Musings</a></strong>.</p><p>&#128073; Visit <strong><a href="https://www.rockcyber.com/">RockCyber.com</a></strong> to learn more about how we can help you in your traditional Cybersecurity and AI Security and Governance Journey</p><p>&#128073; Want to save a quick $100K? Check out our AI Governance Tools at <strong><a href="https://aigovernancetoolkit.com/">AIGovernanceToolkit.com</a></strong></p><p>&#128073; Subscribe for more AI and cyber insights with the occasional rant.</p><p><em>The views and opinions expressed in RockCyber Musings are my own and do not represent the positions of my employer or any organization I&#8217;m affiliated with.</em></p><div class="captioned-button-wrap" data-attrs="{&quot;url&quot;:&quot;https://www.rockcybermusings.com/p/weekly-musings-top-10-ai-security-20260327-20260402?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;}" data-component-name="CaptionedButtonToDOM"><div class="preamble"><p class="cta-caption">Thanks for reading RockCyber Musings! This post is public, so feel free to share it.</p></div><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.rockcybermusings.com/p/weekly-musings-top-10-ai-security-20260327-20260402?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.rockcybermusings.com/p/weekly-musings-top-10-ai-security-20260327-20260402?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p></div><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.rockcybermusings.com/?utm_source=substack&amp;utm_medium=email&amp;utm_content=share&amp;action=share&quot;,&quot;text&quot;:&quot;Share RockCyber Musings&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.rockcybermusings.com/?utm_source=substack&amp;utm_medium=email&amp;utm_content=share&amp;action=share"><span>Share RockCyber Musings</span></a></p><h2>References</h2><p>Axios. (2026, March 31). Anthropic leaked its own Claude source code. <a href="https://www.axios.com/2026/03/31/anthropic-leaked-source-code-ai">https://www.axios.com/2026/03/31/anthropic-leaked-source-code-ai</a></p><p>Axios. (2026, April 2). Exclusive: Gottheimer presses Anthropic on source code leaks and safety protocols. <a href="https://www.axios.com/2026/04/02/gottheimer-anthropic-source-code-leaks">https://www.axios.com/2026/04/02/gottheimer-anthropic-source-code-leaks</a></p><p>BleepingComputer. (2026, March 27). CISA: New Langflow flaw actively exploited to hijack AI workflows. <a href="https://www.bleepingcomputer.com/news/security/cisa-new-langflow-flaw-actively-exploited-to-hijack-ai-workflows/">https://www.bleepingcomputer.com/news/security/cisa-new-langflow-flaw-actively-exploited-to-hijack-ai-workflows/</a></p><p>Bloomberg. (2026, March 27). European Commission&#8217;s data stolen in hack on AWS account. <a href="https://www.bloomberg.com/news/articles/2026-03-27/european-commission-s-data-stolen-in-hack-on-aws-account">https://www.bloomberg.com/news/articles/2026-03-27/european-commission-s-data-stolen-in-hack-on-aws-account</a></p><p>Bloomberg. (2026, April 1). Anthropic takes down thousands of GitHub repos trying to yank its leaked source code. <a href="https://www.bloomberg.com/news/articles/2026-04-01/anthropic-scrambles-to-address-leak-of-claude-code-source-code">https://www.bloomberg.com/news/articles/2026-04-01/anthropic-scrambles-to-address-leak-of-claude-code-source-code</a></p><p>CNBC. (2026, March 31). Anthropic leaks part of Claude Code&#8217;s internal source code. <a href="https://www.cnbc.com/2026/03/31/anthropic-leak-claude-code-internal-source.html">https://www.cnbc.com/2026/03/31/anthropic-leak-claude-code-internal-source.html</a></p><p>CoinDesk. (2026, March 27). Anthropic&#8217;s massive Claude Mythos leak reveals a new AI model that could be a cybersecurity nightmare. <a href="https://www.coindesk.com/markets/2026/03/27/anthropic-s-massive-claude-mythos-leak-reveals-a-new-ai-model-that-could-be-a-cybersecurity-nightmare">https://www.coindesk.com/markets/2026/03/27/anthropic-s-massive-claude-mythos-leak-reveals-a-new-ai-model-that-could-be-a-cybersecurity-nightmare</a></p><p>Fortune. (2026, March 27). Anthropic accidentally leaked details of a new AI model that poses unprecedented cybersecurity risks. <a href="https://fortune.com/2026/03/27/anthropic-leaked-ai-mythos-cybersecurity-risk/">https://fortune.com/2026/03/27/anthropic-leaked-ai-mythos-cybersecurity-risk/</a></p><p>Fortune. (2026, March 31). Anthropic leaks its own AI coding tool&#8217;s source code in second major security breach. <a href="https://fortune.com/2026/03/31/anthropic-source-code-claude-code-data-leak-second-security-lapse-days-after-accidentally-revealing-mythos/">https://fortune.com/2026/03/31/anthropic-source-code-claude-code-data-leak-second-security-lapse-days-after-accidentally-revealing-mythos/</a></p><p>Help Net Security. (2026, March 27). CISA sounds alarm on Langflow RCE, Trivy supply chain compromise after rapid exploitation. <a href="https://www.helpnetsecurity.com/2026/03/27/cve-2026-33017-cve-2026-33634-exploited/">https://www.helpnetsecurity.com/2026/03/27/cve-2026-33017-cve-2026-33634-exploited/</a></p><p>Help Net Security. (2026, March 30). Second data breach at European Commission this year leaves open questions over resilience. <a href="https://www.helpnetsecurity.com/2026/03/30/european-commission-cyberattack-cloud-infrastructure-website/">https://www.helpnetsecurity.com/2026/03/30/european-commission-cyberattack-cloud-infrastructure-website/</a></p><p>IAPP. (2026). European Commission misses deadline for AI Act guidance on high-risk systems. <a href="https://iapp.org/news/a/european-commission-misses-deadline-for-ai-act-guidance-on-high-risk-systems">https://iapp.org/news/a/european-commission-misses-deadline-for-ai-act-guidance-on-high-risk-systems</a></p><p>IAPP. (2026, March). EU Digital Omnibus: Analysis of key changes. <a href="https://iapp.org/news/a/eu-digital-omnibus-analysis-of-key-changes">https://iapp.org/news/a/eu-digital-omnibus-analysis-of-key-changes</a></p><p>Qualys ThreatPROTECT. (2026, March 26). CISA Added Langflow Vulnerability to its Known Exploited Vulnerabilities Catalog (CVE-2026-33017). <a href="https://threatprotect.qualys.com/2026/03/26/cisa-added-langflow-vulnerability-to-its-known-exploited-vulnerabilities-catalog-cve-2026-33017/">https://threatprotect.qualys.com/2026/03/26/cisa-added-langflow-vulnerability-to-its-known-exploited-vulnerabilities-catalog-cve-2026-33017/</a></p><p>SecurityAffairs. (2026, March 27). The European Commission confirmed a cyberattack affecting part of its cloud systems. <a href="https://securityaffairs.com/190067/data-breach/the-european-commission-confirmed-a-cyberattack-affecting-part-of-its-cloud-systems.html">https://securityaffairs.com/190067/data-breach/the-european-commission-confirmed-a-cyberattack-affecting-part-of-its-cloud-systems.html</a></p><p>Sysdig. (2026, March 27). CVE-2026-33017: How attackers compromised Langflow AI pipelines in 20 hours. <a href="https://www.sysdig.com/blog/cve-2026-33017-how-attackers-compromised-langflow-ai-pipelines-in-20-hours">https://www.sysdig.com/blog/cve-2026-33017-how-attackers-compromised-langflow-ai-pipelines-in-20-hours</a></p><p>TechCrunch. (2026, March 27). European Commission confirms cyberattack after hackers claim data breach. <a href="https://techcrunch.com/2026/03/27/european-commission-confirms-cyberattack-after-hackers-claim-data-breach/">https://techcrunch.com/2026/03/27/european-commission-confirms-cyberattack-after-hackers-claim-data-breach/</a></p><p>TechCrunch. (2026, April 1). Anthropic took down thousands of GitHub repos trying to yank its leaked source code. <a href="https://techcrunch.com/2026/04/01/anthropic-took-down-thousands-of-github-repos-trying-to-yank-its-leaked-source-code-a-move-the-company-says-was-an-accident/">https://techcrunch.com/2026/04/01/anthropic-took-down-thousands-of-github-repos-trying-to-yank-its-leaked-source-code-a-move-the-company-says-was-an-accident/</a></p><p>The Hacker News. (2026, March 27). LangChain, LangGraph flaws expose files, secrets, databases in widely used AI frameworks. <a href="https://thehackernews.com/2026/03/langchain-langgraph-flaws-expose-files.html">https://thehackernews.com/2026/03/langchain-langgraph-flaws-expose-files.html</a></p><p>The Hill. (2026, April 2). House Democrat pushes Anthropic on safety protocols, source code leak. <a href="https://thehill.com/policy/technology/5812881-gottheimer-presses-anthropic-ai-safety/">https://thehill.com/policy/technology/5812881-gottheimer-presses-anthropic-ai-safety/</a></p><p>Tech Policy Press. (2026). EU&#8217;s AI Act delays let high-risk systems dodge oversight. <a href="https://www.techpolicy.press/eus-ai-act-delays-let-highrisk-systems-dodge-oversight/">https://www.techpolicy.press/eus-ai-act-delays-let-highrisk-systems-dodge-oversight/</a></p><p>Transparency Coalition. (2026, March 27). AI legislative update: March 27, 2026. <a href="https://www.transparencycoalition.ai/news/ai-legislative-update-march27-2026">https://www.transparencycoalition.ai/news/ai-legislative-update-march27-2026</a></p><p>Troutman Pepper Locke. (2026, March 30). Proposed state AI law update: March 30, 2026. <a href="https://www.troutmanprivacy.com/2026/03/proposed-state-ai-law-update-march-30-2026/">https://www.troutmanprivacy.com/2026/03/proposed-state-ai-law-update-march-30-2026/</a></p><p>Aquasecurity. (2026). Trivy ecosystem supply chain temporarily compromised [GitHub Security Advisory GHSA-69fq-xp46-6x23]. <a href="https://github.com/aquasecurity/trivy/security/advisories/GHSA-69fq-xp46-6x23">https://github.com/aquasecurity/trivy/security/advisories/GHSA-69fq-xp46-6x23</a></p><p>European Parliament Think Tank. (2026, March 18). Enforcement of the AI Act. <a href="https://epthinktank.eu/2026/03/18/enforcement-of-the-ai-act/">https://epthinktank.eu/2026/03/18/enforcement-of-the-ai-act/</a></p><p>Jiang, Z., et al. (2026, March 31). Architecting secure AI agents: Perspectives on system-level defenses against indirect prompt injection attacks [Preprint]. ArXiv. <a href="https://arxiv.org/abs/2603.30016">https://arxiv.org/abs/2603.30016</a></p>]]></content:encoded></item><item><title><![CDATA[AI Monitoring Is a Standards Problem, Not a Technology Problem]]></title><description><![CDATA[NIST AI 800-4 proves AI monitoring fails from missing standards, not missing tech. Specific actions CISOs should take before EU AI Act Article 72 hits August 2026.]]></description><link>https://www.rockcybermusings.com/p/ai-monitoring-standards-gap-nist-ai-800-4</link><guid isPermaLink="false">https://www.rockcybermusings.com/p/ai-monitoring-standards-gap-nist-ai-800-4</guid><dc:creator><![CDATA[Rock Lambros]]></dc:creator><pubDate>Tue, 31 Mar 2026 12:50:10 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!c_2d!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0c8b894d-0ef9-41d5-ac90-daa69ba1bfeb_2048x2048.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!c_2d!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0c8b894d-0ef9-41d5-ac90-daa69ba1bfeb_2048x2048.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!c_2d!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0c8b894d-0ef9-41d5-ac90-daa69ba1bfeb_2048x2048.jpeg 424w, https://substackcdn.com/image/fetch/$s_!c_2d!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0c8b894d-0ef9-41d5-ac90-daa69ba1bfeb_2048x2048.jpeg 848w, https://substackcdn.com/image/fetch/$s_!c_2d!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0c8b894d-0ef9-41d5-ac90-daa69ba1bfeb_2048x2048.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!c_2d!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0c8b894d-0ef9-41d5-ac90-daa69ba1bfeb_2048x2048.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!c_2d!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0c8b894d-0ef9-41d5-ac90-daa69ba1bfeb_2048x2048.jpeg" width="1456" height="1456" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/0c8b894d-0ef9-41d5-ac90-daa69ba1bfeb_2048x2048.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1456,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:3060907,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.rockcybermusings.com/i/192386096?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0c8b894d-0ef9-41d5-ac90-daa69ba1bfeb_2048x2048.jpeg&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!c_2d!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0c8b894d-0ef9-41d5-ac90-daa69ba1bfeb_2048x2048.jpeg 424w, https://substackcdn.com/image/fetch/$s_!c_2d!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0c8b894d-0ef9-41d5-ac90-daa69ba1bfeb_2048x2048.jpeg 848w, https://substackcdn.com/image/fetch/$s_!c_2d!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0c8b894d-0ef9-41d5-ac90-daa69ba1bfeb_2048x2048.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!c_2d!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0c8b894d-0ef9-41d5-ac90-daa69ba1bfeb_2048x2048.jpeg 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>NIST just published an admission that nobody knows how to monitor AI systems after deployment. NIST AI 800-4, &#8220;Challenges to the Monitoring of Deployed AI Systems,&#8221; reviews findings from three workshops, 250+ experts, and almost 90 research papers. The document catalogs over 30 distinct challenges. It offers zero solutions. That&#8217;s not a criticism. That&#8217;s the diagnosis, and that should raise your spidey senses.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.rockcybermusings.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading RockCyber Musings! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.rockcybermusings.com/p/ai-monitoring-standards-gap-nist-ai-800-4?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.rockcybermusings.com/p/ai-monitoring-standards-gap-nist-ai-800-4?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p><h2>NIST Mapped the Mess</h2><p>The report organizes post-deployment AI monitoring into six categories:</p><ol><li><p>Functionality (does it still work as intended?)</p></li><li><p>Operational (does the infrastructure hold?)</p></li><li><p>Human Factors (is it transparent and useful to humans?)</p></li><li><p>Security (is it defended against attacks?)</p></li><li><p>Compliance (does it meet regulatory requirements?)</p></li><li><p>Large-Scale Impacts (does it promote human flourishing?)</p></li></ol><p>Each category carries its own distinct challenges. Functionality monitoring suffers from a lack of ground-truth datasets and a lack of a reliable way to detect model drift. Operational monitoring struggles with fragmented logging across distributed infrastructure. Human Factors monitoring, which drew more practitioner attention than any other category in the workshops, remains almost entirely unstudied in the literature. Security monitoring faces the unsettling reality that some models appear to detect when they&#8217;re being evaluated, changing their behavior under observation. Compliance monitoring lacks even basic tracking of terms-of-service violations, including downstream fine-tuning of open models for CSAM generation. Large-Scale Impacts monitoring lacks agreed-upon metrics to measure whether AI systems help or harm people at scale.</p><p>That&#8217;s a lot of individual problems. The question is whether they share a common root cause.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!lyFV!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc03c3c9b-daf1-4081-8f07-392aa245e745_4675x1570.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!lyFV!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc03c3c9b-daf1-4081-8f07-392aa245e745_4675x1570.png 424w, https://substackcdn.com/image/fetch/$s_!lyFV!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc03c3c9b-daf1-4081-8f07-392aa245e745_4675x1570.png 848w, https://substackcdn.com/image/fetch/$s_!lyFV!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc03c3c9b-daf1-4081-8f07-392aa245e745_4675x1570.png 1272w, https://substackcdn.com/image/fetch/$s_!lyFV!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc03c3c9b-daf1-4081-8f07-392aa245e745_4675x1570.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!lyFV!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc03c3c9b-daf1-4081-8f07-392aa245e745_4675x1570.png" width="1456" height="489" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/c03c3c9b-daf1-4081-8f07-392aa245e745_4675x1570.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:489,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:372741,&quot;alt&quot;:&quot;Flowchart showing five cross-cutting monitoring challenges identified by NIST AI 800-4 converging on a missing standards layer as the common root cause&quot;,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.rockcybermusings.com/i/192386096?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc03c3c9b-daf1-4081-8f07-392aa245e745_4675x1570.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="Flowchart showing five cross-cutting monitoring challenges identified by NIST AI 800-4 converging on a missing standards layer as the common root cause" title="Flowchart showing five cross-cutting monitoring challenges identified by NIST AI 800-4 converging on a missing standards layer as the common root cause" srcset="https://substackcdn.com/image/fetch/$s_!lyFV!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc03c3c9b-daf1-4081-8f07-392aa245e745_4675x1570.png 424w, https://substackcdn.com/image/fetch/$s_!lyFV!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc03c3c9b-daf1-4081-8f07-392aa245e745_4675x1570.png 848w, https://substackcdn.com/image/fetch/$s_!lyFV!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc03c3c9b-daf1-4081-8f07-392aa245e745_4675x1570.png 1272w, https://substackcdn.com/image/fetch/$s_!lyFV!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc03c3c9b-daf1-4081-8f07-392aa245e745_4675x1570.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Figure 1: NIST AI 800-4 Cross-Cutting Challenges</figcaption></figure></div><h2>The Root Cause NIST Documented Without Naming</h2><p>Read the cross-cutting challenges section carefully. Five categories of barriers span every monitoring type: </p><ol><li><p>No trusted methods and tools</p></li><li><p>Poor visibility and transparency</p></li><li><p>Pace of change</p></li><li><p>Organizational incentive failures</p></li><li><p>Resource constraints</p></li></ol><p>Strip away the academic framing, and a pattern emerges. Workshop attendees were asking questions that belong in a standards body, not a research lab.</p><p>One attendee called for &#8220;an abstraction layer for universal security and monitoring.&#8221; Others asked, &#8220;What does the information sharing of what&#8217;s measured look like up and down the value chain?&#8221; Multiple participants flagged the absence of common metrics across use cases, noting that &#8220;non-standardized logic for generating metrics across use cases prevents us from building easy platform capabilities for monitoring.&#8221;</p><p>It&#8217;s important to point out that not every challenge NIST documented is a standards problem. Detecting deceptive behavior in models that modify their behavior under observation remains an open research problem. No specification can fix it because nobody knows how to do it reliably yet. Human-AI feedback loops are an understudied science. Ground-truth dataset availability is a data and methodology problem. The field faces three categories of challenge simultaneously: standards gaps (metrics, logging formats, reporting schemas), research gaps (deceptive behavior detection, feedback loop dynamics), and adoption gaps (methods exist in adjacent fields but aren&#8217;t applied to AI).</p><p>The standards layer is the prerequisite that makes progress on the other two categories possible. Without common definitions, you can&#8217;t scale research findings into production monitoring. Without shared schemas, adoption of proven methods stays trapped inside individual vendor implementations. Take deception detection as an example. You can&#8217;t begin researching whether a model&#8217;s stated reasoning matches its actual behavior unless you&#8217;re capturing structured reasoning traces alongside action logs in the first place. The research gap depends on closing the standards gap.</p><h2>You&#8217;ve Seen This Movie Before</h2><p>How did this work out for us in cybersecurity? We&#8217;ve had a 20-year head start on this exact problem.</p><p>Before syslog standardization, every network device vendor shipped its own logging format. Security teams drowned in data they couldn&#8217;t correlate. Firewalls from one vendor produced logs that meant nothing to the SIEM built for another vendor&#8217;s format. Every firewall had monitoring, but none of them spoke the same language.</p><p>The fix wasn&#8217;t a better firewall. It was CEF (Common Event Format), then LEEF (Log Event Extended Format), and now OCSF (Open Cybersecurity Schema Framework). Common schemas let security teams correlate events across vendors, build cross-platform detection rules, and operate SOCs that don&#8217;t require a translator for each data source. The technology didn&#8217;t change. The standards layer underneath made the existing technology useful at scale.</p><p>The AI monitoring equivalent would need agent-specific semantic conventions built on the observability infrastructure enterprises already operate. Not a new standard competing with OpenTelemetry. Extensions to OpenTelemetry that understand agent reasoning steps, tool calls, and multi-agent handoffs. Security events are mapped to schemas that flow into existing SIEMs without custom parsers. The pattern is identical: don&#8217;t build a parallel universe of AI-specific tooling. Extend the standards that security teams already trust.</p><p>AI monitoring is stuck in the pre-syslog era. Every platform defines its own metrics, its own log structures, its own alert taxonomies. If your organization runs AI workloads across three cloud providers and two agent frameworks, you operate five separate monitoring stacks that don&#8217;t talk to each other.</p><p>Here&#8217;s what that looks like in practice. A regional bank deploys a customer-facing loan origination model hosted on one cloud provider&#8217;s ML platform. The model calls a third-party credit scoring API. A separate vendor supplies the fairness monitoring layer. The bank&#8217;s compliance team uses an internal dashboard that pulls from the cloud provider&#8217;s native monitoring. When the credit scoring API updates its model without notification, the loan origination model starts producing subtly different risk scores. Approval rates for one demographic bracket shift by 4% over six weeks. The fairness monitoring vendor&#8217;s tool flags a drift alert using its own proprietary metric. The cloud provider&#8217;s native monitoring shows no anomaly because its baseline was never calibrated against the third-party API&#8217;s output distribution. The compliance dashboard, which aggregates data from both sources, shows conflicting signals that the compliance analyst can&#8217;t reconcile because the two tools define &#8220;drift&#8221; differently, measure it on different time windows, and log it in incompatible formats.</p><p>Nobody in that chain did anything wrong individually. The fairness vendor&#8217;s tool worked as designed. The cloud provider&#8217;s monitoring worked as designed. The gap was structural. There was no shared definition of what &#8220;drift&#8221; means across the pipeline, no common logging schema that would let the compliance team correlate events from two different monitoring tools, and no standardized way for the credit scoring API provider to notify downstream consumers of model updates.</p><p>That scenario plays out today in financial services, healthcare, and any sector that assembles AI capabilities from multiple vendors. NIST AI 800-4 confirmed it with receipts from 250 practitioners saying the same thing in different words.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!6a20!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffd225a52-68a5-492f-b1ad-66e68cc28c9b_6900x4582.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!6a20!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffd225a52-68a5-492f-b1ad-66e68cc28c9b_6900x4582.png 424w, https://substackcdn.com/image/fetch/$s_!6a20!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffd225a52-68a5-492f-b1ad-66e68cc28c9b_6900x4582.png 848w, https://substackcdn.com/image/fetch/$s_!6a20!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffd225a52-68a5-492f-b1ad-66e68cc28c9b_6900x4582.png 1272w, https://substackcdn.com/image/fetch/$s_!6a20!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffd225a52-68a5-492f-b1ad-66e68cc28c9b_6900x4582.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!6a20!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffd225a52-68a5-492f-b1ad-66e68cc28c9b_6900x4582.png" width="1456" height="967" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/fd225a52-68a5-492f-b1ad-66e68cc28c9b_6900x4582.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:967,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:1081734,&quot;alt&quot;:&quot;Timeline showing regulatory monitoring requirements from EU AI Act and NIST AI RMF against the current maturity of monitoring standards&quot;,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.rockcybermusings.com/i/192386096?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffd225a52-68a5-492f-b1ad-66e68cc28c9b_6900x4582.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="Timeline showing regulatory monitoring requirements from EU AI Act and NIST AI RMF against the current maturity of monitoring standards" title="Timeline showing regulatory monitoring requirements from EU AI Act and NIST AI RMF against the current maturity of monitoring standards" srcset="https://substackcdn.com/image/fetch/$s_!6a20!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffd225a52-68a5-492f-b1ad-66e68cc28c9b_6900x4582.png 424w, https://substackcdn.com/image/fetch/$s_!6a20!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffd225a52-68a5-492f-b1ad-66e68cc28c9b_6900x4582.png 848w, https://substackcdn.com/image/fetch/$s_!6a20!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffd225a52-68a5-492f-b1ad-66e68cc28c9b_6900x4582.png 1272w, https://substackcdn.com/image/fetch/$s_!6a20!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffd225a52-68a5-492f-b1ad-66e68cc28c9b_6900x4582.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Figure 2: The Monitoring Standards Gap</figcaption></figure></div><h2>Article 72 Is Already Undeliverable</h2><p>Regulators aren&#8217;t waiting for standards to mature. The EU AI Act&#8217;s high-risk system obligations take effect August 2, 2026 (if the aren&#8217;t delayed). Article 72 requires providers of high-risk AI systems to implement post-market monitoring plans that &#8220;actively and systematically collect, document and analyse relevant data&#8221; on system performance throughout the system&#8217;s lifetime. Deployers face separate obligations to monitor operations and report serious incidents within 72-hour and 15-day windows.</p><p>Pull one thread, and the gap becomes specific. Article 72 requires providers to collect performance data &#8220;throughout their lifetime&#8221; and evaluate &#8220;continuous compliance.&#8221; NIST AI 800-4 documents that practitioners lack standardized performance metrics, can&#8217;t establish baselines or deviation thresholds, and have no systematic way to compare model behavior across providers. One workshop attendee put it bluntly: &#8220;It&#8217;s often unclear what exactly to monitor and how.&#8221; The report cites research confirming that &#8220;the appropriate metrics to capture is not standardized in the AI community&#8221; and warns this &#8220;absence can result in misleading performance measures.&#8221;</p><p>That&#8217;s not a general compliance gap. Article 72 requires continuous collection and analysis of performance data. NIST AI 800-4 confirms that the field hasn&#8217;t agreed on what &#8220;performance&#8221; means in post-deployment contexts, let alone how to measure it consistently across different AI systems and providers. The regulation demands an activity that is structurally undeliverable with the current monitoring ecosystem. Organizations filing post-market monitoring plans in 2026 will document processes built on unstandardized metrics, non-interoperable tools, and self-defined baselines. They&#8217;ll comply on paper. The monitoring itself won&#8217;t be comparable, auditable, or meaningful across organizational boundaries.</p><p>Compliance requires two capabilities this ecosystem lacks: runtime hooks that produce monitoring data in standardized formats, and trace architectures that reconstruct decision chains across organizational boundaries. Without these, Article 72 post-market monitoring plans are fiction written in incompatible vendor dialects.</p><p>NIST&#8217;s own AI Risk Management Framework compounds the pressure. The MANAGE function calls for continuous monitoring and risk response throughout deployment. The forthcoming NIST Cyber AI Profile maps cybersecurity controls to AI-specific concerns like model integrity and adversarial robustness. Every framework converges on the same expectation. The implementation layer that would make compliance verifiable doesn&#8217;t exist yet.</p><h2>Who&#8217;s Responsible? Nobody Knows That Either.</h2><p>NIST AI 800-4 surfaced a question that&#8217;s arguably more urgent than the technical gaps: who monitors? Workshop attendees repeatedly asked: &#8220;Who should do monitoring?&#8221; &#8220;Who is responsible for remediating incidents?&#8221; and &#8220;If anything is found, who can act on it?&#8221;</p><p>In the bank scenario above, was the monitoring failure the cloud provider&#8217;s responsibility? The fairness vendor&#8217;s? The credit scoring API provider&#8217;s? The bank&#8217;s compliance team? Each party monitored its own slice of the pipeline. Nobody monitored the seams between them. The NIST report documents this as an unresolved question across the AI supply chain, and it&#8217;s compounded by the standards gap. You can&#8217;t assign responsibility for monitoring when you haven&#8217;t agreed on what monitoring means. You can&#8217;t hold a vendor accountable for failing to report a drift event when &#8220;drift&#8221; has no shared definition.</p><p>A viable monitoring architecture separates three concerns. The platform exposes standardized observation and control points. An open enforcement layer applies policy through those control points, portable across any platform that exposes them. The enterprise customizes policy to its domain: financial services brings its own data sensitivity models, healthcare brings PHI detection, and any regulated industry brings its compliance requirements. When responsibilities are layered this way, the question of &#8220;who monitors?&#8221; has a structural answer. The platform enables. Open tooling enforces. The enterprise governs. Accountability follows the layer where the failure occurred.</p><p>One attendee asked how to &#8220;reduce the burden on the end user&#8221; to validate model behavior. Another asked how monitoring could become &#8220;a more collaborative practice, rather than a closed technical process.&#8221; These aren&#8217;t theoretical musings. They&#8217;re the governance questions that determine whether monitoring happens at all or degenerates into checkbox compliance where everyone points at someone else&#8217;s dashboard. A layered architecture gives each party a defined obligation: expose, enforce, govern. The current ecosystem gives everyone an excuse.</p><h2>Agents Make Everything Worse</h2><p>If the standards gap is a problem for current AI systems, it&#8217;s a crisis for agentic AI. NIST SP 800-4 repeatedly mentions agents, and the findings are sobering.</p><p>Workshop attendees flagged &#8220;lengthy agentic tasks&#8221; as especially resource-intensive to monitor. The report cites research noting that &#8220;both the agents and the operational environment are subject to change,&#8221; making static monitoring baselines unreliable. Agent identification and tracking remain unstandardized. Attendees raised visibility challenges around &#8220;out-of-distribution behavior using agent identifiers&#8221; and noted that watermarking and content provenance measures &#8220;face reliability challenges.&#8221; One attendee asked directly: &#8220;Is the model agentically attempting to subvert the monitoring setup it is under, i.e., scheming?&#8221;</p><p>That question deserves a pause. We&#8217;re building systems that plan, execute across organizational boundaries, call external tools, and collaborate with other agents. The monitoring challenges NIST documented for conventional AI systems, from detecting drift to maintaining visibility to establishing baselines, all assume a relatively static system being observed from outside. Agents aren&#8217;t static. They change behavior based on context, discover new capabilities at runtime, and operate across a distributed infrastructure that no single organization fully controls.</p><p>Any monitoring standard for agents needs a dynamic inventory mechanism. A static software bill of materials generated at deployment time is worthless when agents discover new tools, connect to new service endpoints, and modify their own capabilities during a single execution session. The inventory must update in real time, triggered by component changes, and output in formats the supply chain security ecosystem already consumes. If your agent connects to a new MCP server mid-task and your inventory doesn&#8217;t reflect that within the same session, your security team is operating on a stale map.</p><p>The &#8220;monitorability tax&#8221; concept raised in the report&#8217;s cited research captures the emerging cost structure. Model developers will pay a performance penalty, through slower inference or less capable models, to maintain the ability to monitor agent behavior. That cost rises as agent autonomy increases. Standardized hooks reduce the engineering cost by making monitoring implementation portable across frameworks, a one-time platform integration rather than custom monitoring code for every deployment. The monitorability tax on compute remains. The tax on engineering effort doesn&#8217;t have to.</p><p>The cross-provider abstraction layer that workshop attendees called for isn&#8217;t a nice-to-have for agentic systems. Without standardized hooks for runtime monitoring, standardized trace formats for multi-agent workflows, and standardized inventories of agent capabilities and dependencies, you&#8217;re watching agents through whatever proprietary window each vendor provides. You can&#8217;t correlate behavior across platforms. You can&#8217;t reconstruct decision chains that span multiple agent frameworks. You can&#8217;t audit what you can&#8217;t consistently observe.</p><p>One more structural blind spot worth naming: runtime monitoring standards assume a cooperating platform that exposes hooks. Open-weight models distributed without platforms bypass this assumption entirely. Once a model is released into the wild for anyone to run, no runtime hook exists unless the downstream deployer voluntarily implements one. Open-weight models are structurally ungovernable by runtime standards alone. Any honest conversation about the monitoring gap has to acknowledge this boundary.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!Ku7d!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe46f6983-6c17-499d-a36e-9e51b3bdb476_3021x2975.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!Ku7d!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe46f6983-6c17-499d-a36e-9e51b3bdb476_3021x2975.png 424w, https://substackcdn.com/image/fetch/$s_!Ku7d!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe46f6983-6c17-499d-a36e-9e51b3bdb476_3021x2975.png 848w, https://substackcdn.com/image/fetch/$s_!Ku7d!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe46f6983-6c17-499d-a36e-9e51b3bdb476_3021x2975.png 1272w, https://substackcdn.com/image/fetch/$s_!Ku7d!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe46f6983-6c17-499d-a36e-9e51b3bdb476_3021x2975.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!Ku7d!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe46f6983-6c17-499d-a36e-9e51b3bdb476_3021x2975.png" width="1456" height="1434" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/e46f6983-6c17-499d-a36e-9e51b3bdb476_3021x2975.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1434,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:718881,&quot;alt&quot;:&quot;Block diagram showing how agentic AI properties such as autonomous planning, tool discovery, and multi-agent collaboration amplify each monitoring challenge NIST identified&quot;,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.rockcybermusings.com/i/192386096?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe46f6983-6c17-499d-a36e-9e51b3bdb476_3021x2975.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="Block diagram showing how agentic AI properties such as autonomous planning, tool discovery, and multi-agent collaboration amplify each monitoring challenge NIST identified" title="Block diagram showing how agentic AI properties such as autonomous planning, tool discovery, and multi-agent collaboration amplify each monitoring challenge NIST identified" srcset="https://substackcdn.com/image/fetch/$s_!Ku7d!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe46f6983-6c17-499d-a36e-9e51b3bdb476_3021x2975.png 424w, https://substackcdn.com/image/fetch/$s_!Ku7d!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe46f6983-6c17-499d-a36e-9e51b3bdb476_3021x2975.png 848w, https://substackcdn.com/image/fetch/$s_!Ku7d!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe46f6983-6c17-499d-a36e-9e51b3bdb476_3021x2975.png 1272w, https://substackcdn.com/image/fetch/$s_!Ku7d!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe46f6983-6c17-499d-a36e-9e51b3bdb476_3021x2975.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Figure 3: How Agents Amplify the Monitoring Standards Gap</figcaption></figure></div><p><strong>Key Takeaway:</strong> NIST AI 800-4 confirms what practitioners feel in their bones: AI monitoring isn&#8217;t failing because we lack technology. The standards layer that would make technology useful at scale doesn&#8217;t exist. Agents make the gap existential.</p><h3>What to do next</h3><p>Stop accepting proprietary monitoring silos. The next time you evaluate an AI platform, put these questions into the review:</p><ul><li><p>What open logging schema do your monitoring outputs conform to? If the answer is a proprietary format, ask how you export monitoring data into a format another platform can ingest without custom transformation.</p></li><li><p>How does your monitoring define and detect model drift? Compare the answer across your vendors. If two vendors define &#8220;drift&#8221; differently, your compliance team can&#8217;t produce a coherent post-market monitoring report under Article 72.</p></li><li><p>When a component in the AI pipeline (a third-party API, a model update, a data source change) shifts behavior, how does your monitoring surface cross-component effects? If the answer involves manual correlation, you have a gap that scales with system complexity.</p></li><li><p>Who in the supply chain is responsible for monitoring the seams between components? If nobody owns cross-boundary monitoring, say so in your risk register. That&#8217;s an accepted risk, not an oversight.</p></li><li><p>Does your AI platform expose standardized middleware hooks that allow your security team to intercept and evaluate agent actions before they execute? If the platform&#8217;s controls are proprietary and non-portable, your enforcement logic dies with the vendor relationship. Every policy you write, every guardrail you configure, every compliance rule you encode is locked to one vendor&#8217;s architecture.</p></li></ul><p>Push your industry groups and standards bodies. If you participate in OWASP, ISO working groups, or NIST-affiliated communities, advocate for common AI monitoring vocabularies and reference architectures. The cybersecurity field solved this problem a decade ago with common event formats and shared schemas. The AI field hasn&#8217;t started.</p><p>Audit your own monitoring maturity against the six NIST categories. Most organizations will find entire categories with no monitoring at all, particularly Human Factors and Large-Scale Impacts. Map the gaps before the next board meeting where someone asks if you&#8217;re ready for August 2026.</p><p>The full NIST AI 800-4 report is available at <a href="https://doi.org/10.6028/NIST.AI.800-4">https://doi.org/10.6028/NIST.AI.800-4</a>. </p><p>&#128073; Visit <strong><a href="https://www.rockcyber.com/">RockCyber.com</a></strong> to learn more about how we can help you in your traditional Cybersecurity and AI Security and Governance Journey</p><p>&#128073; Want to save a quick $100K? Check out our AI Governance Tools at <strong><a href="https://aigovernancetoolkit.com/">AIGovernanceToolkit.com</a></strong></p><p>&#128073; Subscribe for more AI and cyber insights with the occasional rant.</p><p><em>The views and opinions expressed in RockCyber Musings are my own and do not represent the positions of my employer or any organization I&#8217;m affiliated with.</em></p><p>Thanks for reading RockCyber Musings! Subscribe for free to receive new posts and support my work.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.rockcybermusings.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading RockCyber Musings! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.rockcybermusings.com/?utm_source=substack&amp;utm_medium=email&amp;utm_content=share&amp;action=share&quot;,&quot;text&quot;:&quot;Share RockCyber Musings&quot;,&quot;action&quot;:null,&quot;class&quot;:&quot;button-wrapper&quot;}" data-component-name="ButtonCreateButton"><a class="button primary button-wrapper" href="https://www.rockcybermusings.com/?utm_source=substack&amp;utm_medium=email&amp;utm_content=share&amp;action=share"><span>Share RockCyber Musings</span></a></p>]]></content:encoded></item><item><title><![CDATA[Weekly Musings Top 10 AI Security Wrapup: Issue 31 March 20-26, 2026]]></title><description><![CDATA[RSA 2026: Every Vendor Sold an Agent. A Supply Chain Attack Ran Quietly in the Background]]></description><link>https://www.rockcybermusings.com/p/weekly-musings-top-10-ai-security-20260320-20260326</link><guid isPermaLink="false">https://www.rockcybermusings.com/p/weekly-musings-top-10-ai-security-20260320-20260326</guid><dc:creator><![CDATA[Rock Lambros]]></dc:creator><pubDate>Fri, 27 Mar 2026 12:11:00 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!4rwy!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7802d5df-7f41-40f2-ad07-154926f08df2_1024x1024.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!4rwy!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7802d5df-7f41-40f2-ad07-154926f08df2_1024x1024.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!4rwy!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7802d5df-7f41-40f2-ad07-154926f08df2_1024x1024.png 424w, https://substackcdn.com/image/fetch/$s_!4rwy!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7802d5df-7f41-40f2-ad07-154926f08df2_1024x1024.png 848w, https://substackcdn.com/image/fetch/$s_!4rwy!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7802d5df-7f41-40f2-ad07-154926f08df2_1024x1024.png 1272w, https://substackcdn.com/image/fetch/$s_!4rwy!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7802d5df-7f41-40f2-ad07-154926f08df2_1024x1024.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!4rwy!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7802d5df-7f41-40f2-ad07-154926f08df2_1024x1024.png" width="1024" height="1024" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/7802d5df-7f41-40f2-ad07-154926f08df2_1024x1024.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1024,&quot;width&quot;:1024,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:1233556,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.rockcybermusings.com/i/192300876?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7802d5df-7f41-40f2-ad07-154926f08df2_1024x1024.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!4rwy!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7802d5df-7f41-40f2-ad07-154926f08df2_1024x1024.png 424w, https://substackcdn.com/image/fetch/$s_!4rwy!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7802d5df-7f41-40f2-ad07-154926f08df2_1024x1024.png 848w, https://substackcdn.com/image/fetch/$s_!4rwy!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7802d5df-7f41-40f2-ad07-154926f08df2_1024x1024.png 1272w, https://substackcdn.com/image/fetch/$s_!4rwy!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7802d5df-7f41-40f2-ad07-154926f08df2_1024x1024.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>RSA Conference 2026 closed Thursday in San Francisco. Thirty thousand attendees, six hundred exhibitors, one word on every booth banner: agentic. While the industry competed on keynotes and happy hours, LiteLLM, deployed in hundreds of enterprise AI stacks, got infected with credential-stealing code through a misconfigured GitHub Actions workflow. Malicious releases went live March 19 and March 22. Most of your security team was watching keynotes.</p><p>Underneath the conference noise, genuine signal emerged. Zenity&#8217;s CTO demonstrated live zero-click exploits against ChatGPT, Salesforce, and Microsoft Copilot on the conference floor. Palo Alto Networks Unit 42 documented new attack paths through the Model Context Protocol. HackerOne disclosed a 540% year-over-year surge in validated prompt injection vulnerabilities. The EU AI Office&#8217;s second draft Code of Practice on AI-generated content transparency is open for feedback through March 30, with prescriptive new requirements that narrow compliance discretion significantly. NIST published AI 800-4, the first federal framework for monitoring AI systems in production, with no vendor booth to announce it.</p><p>Here&#8217;s what matters and what to do about it.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.rockcybermusings.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading RockCyber Musings! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.rockcybermusings.com/p/weekly-musings-top-10-ai-security-20260320-20260326?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.rockcybermusings.com/p/weekly-musings-top-10-ai-security-20260320-20260326?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p><h3>1. Zenity Launches Guardian Agents and Demonstrates 0-Click AI Exploits at RSA</h3><p>Zenity launched Guardian Agents at RSA 2026 on March 23, positioning it as continuous, contextual security for AI agents across SaaS, cloud, and endpoint environments. CTO Michael Bargury ran live demonstrations titled &#8220;Your AI Agents Are My Minions,&#8221; showing zero-click prompt injection chains that manipulated Cursor into leaking developer secrets via support emails, Salesforce agents into exfiltrating customer data to an attacker-controlled server, and ChatGPT into producing persistent attacker-chosen outputs across conversations (The Register, March 23, 2026, and Help Net Security, March 24, 2026).</p><p><strong>Why it matters</strong></p><ul><li><p>Zero-click attacks eliminate the human review checkpoint most AI security frameworks assume is present. When agents act without user input, your primary detection layer disappears before the threat is visible.</p></li><li><p>Live exploitation of production enterprise systems on a conference floor is harder to dismiss than a threat model in a whitepaper.</p></li><li><p>Guardian Agents signals a market category forming in real time. The evaluation criteria you set today will shape purchasing decisions for the next several years.</p></li></ul><p><strong>What to do about it</strong></p><ul><li><p>Inventory every AI agent in your environment before your next board meeting. If you can&#8217;t enumerate them, you can&#8217;t monitor them.</p></li><li><p>Require vendors to document in writing which actions their agents take without explicit human approval. Non-answers are critical control gaps.</p></li><li><p>Run adversarial testing against your three highest-access agents this quarter, targeting credential extraction, data exfiltration, and cross-system manipulation.</p></li></ul><p><strong>Rock&#8217;s Musings</strong></p><p>Bargury&#8217;s demonstration strategy was the most honest thing at RSA this week: show the attack, then show the defense. Live exploitation on production systems is harder to dismiss than a slide deck built around the word autonomous. The inconvenient reality is that most enterprises already have agents running with email access, CRM credentials, and code repository permissions, with no runtime monitoring on what those agents decide to do. Selecting an AI security vendor is not the same thing as having an answer to the problem he demonstrated on the conference floor.</p><div><hr></div><h3>2. LiteLLM Infected with Credential-Stealing Code via Trivy Misconfiguration</h3><p>The Register reported March 24 that LiteLLM, a widely deployed open-source LLM API proxy, was compromised through a misconfigured Trivy GitHub Actions workflow. Attackers modified version tags on the trivy-action GitHub Action to inject malicious code into workflows organizations were already running, producing malicious releases on March 19 and March 22. The maintainer confirmed that anyone who installed and ran the project during that window should assume credentials available to their environment were exposed.</p><p><strong>Why it matters</strong></p><ul><li><p>LiteLLM sits in the critical path of many enterprise AI deployments. One compromised abstraction library reaches hundreds of downstream production systems simultaneously.</p></li><li><p>The attack exploited version tags, not direct code injection. CI/CD pipelines relying on tags rather than pinned commits ran malicious code without detection. That&#8217;s a systemic configuration gap across most enterprise pipelines.</p></li><li><p>The attack ran during RSA week when security teams were distracted. The timing was likely not accidental.</p></li></ul><p><strong>What to do about it</strong></p><ul><li><p>Audit every environment that pulled a LiteLLM update between March 19 and March 24. Treat those environments as potentially compromised until you confirm otherwise.</p></li><li><p>Pin all GitHub Actions to specific commit hashes, not version tags. Tags are mutable and can be silently overwritten. Commits are not.</p></li><li><p>Establish software bill of materials practices for all AI and ML dependencies. Supply chain attacks will keep finding environments where that inventory doesn&#8217;t exist.</p></li></ul><p><strong>Rock&#8217;s Musings</strong></p><p>LiteLLM is exactly the kind of library that lands in enterprise AI stacks without a security review, installed by an ML engineer who needed to route calls to three model providers before the sprint ended. Trivy is a security tool. Attackers used a security tool misconfiguration to compromise a release pipeline for another widely used tool. If there&#8217;s a cleaner argument for applying security rigor to your own security tooling, I haven&#8217;t heard it. Your AI dependency chain needs the same scrutiny as your application dependencies. Good intentions at install time are not a compensating control.</p><div><hr></div><h3>3. Palo Alto Networks Unit 42 Documents MCP Attack Vectors</h3><p>Palo Alto Networks Unit 42 published research the week of March 20 documenting new attack paths through the Model Context Protocol, including prompt injection delivered through MCP&#8217;s sampling interface. Security researchers tracked 30 CVEs filed against MCP implementations in the preceding 60 days, including CVE-2026-25536 (cross-client data leak in the MCP TypeScript SDK) and CVE-2026-23744 (remote code execution in MCPJam Inspector). A scan of more than 500 public MCP servers found that 38% lacked authentication entirely (Unit 42, March 2026, and Adversa.ai, March 2026).</p><p><strong>Why it matters</strong></p><ul><li><p>MCP is the connective tissue between AI agents and enterprise tools. A vulnerability in this protocol exposes the entire agent ecosystem built on top of it, not one isolated system.</p></li><li><p>Thirty CVEs in 60 days signals that security review did not happen before shipping at scale. Every API ecosystem that launches with deployment velocity ahead of security assessment follows this arc.</p></li><li><p>Thirty-eight percent of scanned servers lacking authentication is systemic failure. Authentication is the minimum viable control. Everything built on top of unauthenticated servers is exposed.</p></li></ul><p><strong>What to do about it</strong></p><ul><li><p>Inventory every MCP server in your environment and treat unauthenticated instances as critical findings requiring immediate action.</p></li><li><p>Require authentication, authorization, and comprehensive logging for any MCP server with access to production systems or sensitive data.</p></li><li><p>Demand specific CVE status and patch timelines from your AI infrastructure vendors. Vague answers signal high risk and a vendor not tracking its own exposure.</p></li></ul><p><strong>Rock&#8217;s Musings</strong></p><p>Thirty CVEs in 60 days is not a patching problem. It&#8217;s a design problem. MCP shipped fast because the builders cared more about what AI agents could reach than how securely they could reach it. The 38% authentication gap is the number that should end budget debates about AI infrastructure security investment. Roughly two in five MCP servers operate on the assumption that only authorized parties will talk to them, which is exactly wrong in a protocol designed to connect agents to external resources. That assumption creates direct paths to your production data.</p><div><hr></div><h3>4. HackerOne Reports 540% Surge in Validated Prompt Injection Vulnerabilities</h3><p>HackerOne announced Agentic Prompt Injection Testing on March 21, paired with platform data showing a 540% year-over-year increase in validated prompt injection vulnerabilities. The service executes structured, multi-turn adversarial scenarios against live AI applications, evaluating whether injection attempts produce actual data exposure or unauthorized tool execution across interconnected agent systems (HackerOne Blog, March 2026, and Cybersecurity Insiders, March 21, 2026).</p><p><strong>Why it matters</strong></p><ul><li><p>A 540% increase in validated vulnerabilities means real researchers are finding real exploitable conditions in production systems, not theoretical edge cases.</p></li><li><p>Traditional application security testing does not cover agent-specific attack paths. If your AI agents aren&#8217;t explicitly in scope for your red team or bug bounty program, you have a documented blind spot.</p></li><li><p>Unit 42&#8217;s concurrent research on indirect prompt injection through web content eliminates the &#8220;attacker needs direct access&#8221; objection. Agents read the web. The web is the attack surface.</p></li></ul><p><strong>What to do about it</strong></p><ul><li><p>Add AI agents to your red team scope explicitly as a primary target category, not an afterthought appended to an existing engagement.</p></li><li><p>Require prompt injection testing as part of every AI agent release process, treated as a gate equivalent to penetration testing for any externally facing application.</p></li><li><p>Track prompt injection findings as a distinct vulnerability class in your risk register. You can&#8217;t demonstrate improvement to your board on metrics you&#8217;re not collecting separately.</p></li></ul><p><strong>Rock&#8217;s Musings</strong></p><p>Five hundred forty percent ends the debate about whether prompt injection is a real threat. I&#8217;ve heard the objection that attackers need direct access to craft payloads. Unit 42&#8217;s indirect injection research, published this same week, shows agents reading manipulated instructions from ordinary websites they visit in the course of normal operation. Your agents don&#8217;t need to be directly targeted; they need to visit the wrong page. The gap between organizations deploying AI agents and organizations testing those agents adversarially is the largest unaddressed risk exposure I see in enterprise AI programs right now.</p><div><hr></div><h3>5. Microsoft Publishes Secure Agentic AI Framework and Confirms Agent 365 May 1 GA</h3><p>Microsoft published &#8220;Secure Agentic AI End-to-End&#8221; on March 20, documenting its approach to extending Zero Trust architecture across the full AI agent lifecycle: data ingestion, model training, deployment, and runtime behavioral monitoring. The post confirmed Agent 365, Microsoft&#8217;s governance control plane for enterprise AI agents, will reach general availability on May 1, 2026, with agent identity, authorization scope, and behavioral monitoring treated as distinct security domains from traditional human-user ZT controls (Microsoft Security Blog, March 20, 2026).</p><p><strong>Why it matters</strong></p><ul><li><p>A confirmed May 1 GA date gives enterprises in Microsoft environments a concrete six-week planning horizon. Governance framework adoption takes time and that clock is already running.</p></li><li><p>Extending Zero Trust to AI agents is architecturally correct. Most ZT implementations weren&#8217;t designed with agent identity or behavioral monitoring in mind, making the gap assessment non-trivial work.</p></li><li><p>Publishing detailed technical frameworks before product GA signals Microsoft wants enterprises building governance practices now, before the product ships.</p></li></ul><p><strong>What to do about it</strong></p><ul><li><p>Map your current ZT architecture against the agent-specific requirements described in the March 20 post. Focus on gaps in agent identity and behavioral monitoring specifically.</p></li><li><p>Begin internal stakeholder alignment on Agent 365 if you&#8217;re in a Microsoft 365 environment. Six weeks is not enough time to start that conversation from zero.</p></li><li><p>Document agent permissions, access patterns, and decision scopes using whatever visibility tools you have today rather than waiting for Microsoft tooling.</p></li></ul><p><strong>Rock&#8217;s Musings</strong></p><p>&#8220;End-to-end&#8221; is doing heavy lifting as a title. What Microsoft describes is extending known security primitives to a new execution context. That&#8217;s necessary work and not a complete answer. The hard problems are behavioral: distinguishing authorized agent actions from manipulated ones, detecting policy violations in real time, and maintaining audit trails that survive an incident investigation. Agent 365 is worth watching. If the behavioral monitoring is substantive, it&#8217;ll move the market. If it&#8217;s a compliance dashboard, enterprises will check the box while actual risk sits unaddressed underneath it.</p><div><hr></div><h3>6. Cisco Releases DefenseClaw Open Source on Final Day of RSA</h3><p>Cisco released DefenseClaw to GitHub on March 27, the final day of RSA 2026, as an open-source framework for scanning agent skills and sandboxing agent execution. The release accompanied Zero Trust Access for AI agents and a free AI Defense Explorer Edition targeting security practitioners. Cisco plans integration with NVIDIA OpenShell for hardware-level execution sandboxing, addressing execution isolation that software-only monitoring cannot replicate (Cisco Newsroom, March 2026, and UC Today, March 2026).</p><p><strong>Why it matters</strong></p><ul><li><p>Open-source agent security scanning means organizations can start building security into agent development pipelines without a procurement cycle or a budget line.</p></li><li><p>Hardware-anchored execution sandboxing addresses a control gap that software-only monitoring cannot close. Execution isolation for agents is systematically underinvested across the industry relative to the risk.</p></li><li><p>The open-source and Explorer Edition strategy targets developers before enterprise procurement cycles form, competing for architectural mindshare with builders rather than just buyers.</p></li></ul><p><strong>What to do about it</strong></p><ul><li><p>Pull DefenseClaw and run it against a non-production agent environment this month. Validate real-world utility before committing to any commercial evaluation.</p></li><li><p>Evaluate the NVIDIA sandboxing integration if you&#8217;re running NVIDIA infrastructure. Test in isolation before production consideration.</p></li><li><p>Track Cisco&#8217;s AI Defense commercial roadmap. Free Explorer Editions typically precede commercial tier launches by 12 to 18 months, and starting your evaluation now means you&#8217;ll have data when the pitch arrives.</p></li></ul><p><strong>Rock&#8217;s Musings</strong></p><p>Releasing open-source code on the last day of the conference changes the conversation from &#8220;will enterprises buy this&#8221; to &#8220;pull the repo and see for yourself.&#8221; That&#8217;s a credible move when the code is real and the threat model is honest. Run DefenseClaw against your actual agent environment before making any claims about coverage. The larger play is Cisco&#8217;s bid for the enterprise AI security architecture position using network visibility, an established security portfolio, and enterprise relationships most competitors would need a decade to build. DefenseClaw is a credible opening move. Watch the next 18 months of product decisions to judge the hand.</p><div><hr></div><h3>7. Google Deploys Gemini Agents to Process 10 Million Dark Web Posts Daily</h3><p>Google announced at RSA 2026 on March 23 that Gemini AI agents are processing more than 10 million dark web posts daily to surface threats relevant to specific organizations. The capability integrates with Google Security Operations alongside new agentic automation features, currently in preview, that let security teams combine AI-driven investigation with deterministic automated response workflows (The Register, March 23, 2026, and Google Cloud Blog, March 2026).</p><p><strong>Why it matters</strong></p><ul><li><p>Ten million posts per day changes the economics of dark web threat intelligence. Organizations that couldn&#8217;t sustain comprehensive monitoring programs gain access to Google-scale processing at a fraction of the previous cost.</p></li><li><p>Pairing AI-driven investigation with deterministic automation preserves human-defined control while extending agent reach into high-volume, low-judgment tasks. That&#8217;s the right architectural pattern for agentic SOC work.</p></li><li><p>Preview status means GA behavior, SLA, and security review standards remain unfinalized. Your production SOC is not where you run this experiment yet.</p></li></ul><p><strong>What to do about it</strong></p><ul><li><p>Assess your current dark web monitoring coverage gap against what this capability covers. If there&#8217;s a meaningful difference, prioritize a pilot evaluation once the feature reaches GA.</p></li><li><p>Review preview terms carefully before enabling agentic automation in any production SOC workflow. Preview features carry materially different risk profiles than GA releases.</p></li><li><p>Define which SOC workflows you&#8217;d delegate to agents and where human approval must remain. Build that policy before the tools arrive, not after they&#8217;re already running.</p></li></ul><p><strong>Rock&#8217;s Musings</strong></p><p>Threat intelligence is the most defensible application of AI agents in security operations right now. Failure modes are recoverable: the agent misses a threat and your other controls have a chance at it. Compare that to agentic incident response, where the failure mode might be blocking a production system or destroying forensic evidence. Start with intelligence, not response. The preview framing signals Google is collecting operational data before committing to GA behavior guarantees, which is reasonable product discipline. It also means you wait for GA before running this where failures have material consequences.</p><div><hr></div><h3>8. Novee Launches Autonomous AI Red Teaming Platform for LLM Applications</h3><p>Novee announced autonomous AI red teaming for LLM applications on March 24 at RSA Conference 2026. The platform deploys an AI pentesting agent that executes multi-turn adversarial scenarios against live systems, simulating attacker chaining techniques across prompt injection, jailbreaks, data exfiltration paths, and agent behavior manipulation, covering any LLM-powered system regardless of model provider with optional CI/CD pipeline integration (GlobeNewswire, March 24, 2026, and Help Net Security, March 24-25, 2026).</p><p><strong>Why it matters</strong></p><ul><li><p>Traditional pentesting tools were designed for pre-LLM application security problems. Novee builds red teaming from actual LLM vulnerability research, producing findings that adapted traditional tools miss.</p></li><li><p>CI/CD pipeline integration lets security teams catch prompt injection and agent manipulation issues before production deployment rather than after an incident surfaces them.</p></li><li><p>Two distinct companies announced adversarial AI testing capabilities at RSA 2026 in the same week. Market formation around this problem is accelerating.</p></li></ul><p><strong>What to do about it</strong></p><ul><li><p>Evaluate Novee&#8217;s beta against a non-production LLM application to understand what it surfaces relative to your existing security testing coverage.</p></li><li><p>Map the gap between your current SDL and what LLM-specific adversarial testing would require. The gap is almost certainly larger than you expect it to be.</p></li><li><p>Add AI-native red teaming as a release gate requirement for any LLM application reaching production. Make it a gate, not a post-deployment recommendation that teams skip.</p></li></ul><p><strong>Rock&#8217;s Musings</strong></p><p>Two autonomous AI red teaming announcements in one RSA week tells you the market is accepting that testing AI systems requires AI-specific tooling, not adapted traditional approaches. That&#8217;s a healthy development even if the tools themselves are early. The CI/CD integration angle is the most practically valuable feature: security issues caught before production deployment cost a fraction of what they cost after deployment. If you&#8217;re shipping LLM applications without adversarial testing in the pipeline, you&#8217;re making a risk decision that most boards don&#8217;t know they&#8217;re making.</p><div><hr></div><h3>9. EU AI Office Second Draft Code of Practice Enters Final Feedback Window</h3><p>The EU AI Office published its second draft Code of Practice on AI-Generated Content Transparency on March 3, with the stakeholder feedback window closing March 30. The second draft moves from high-level principles toward prescriptive, technically detailed commitments, narrowing compliance discretion and signaling how regulators will likely assess conformance in practice. A third and final version is expected by June 2026, ahead of the August 2 applicability date for AI-generated content transparency obligations (Herbert Smith Freehills Kramer, March 2026, and BABL AI, March 2026).</p><p><strong>Why it matters</strong></p><ul><li><p>Draft 2&#8217;s shift to prescriptive technical commitments closes the interpretation space organizations were using to plan flexible compliance programs. The gap between &#8220;we have a policy&#8221; and &#8220;we meet the technical specification&#8221; narrowed significantly this month.</p></li><li><p>The March 30 feedback deadline is this weekend. If your organization has substantive views on requirements that are technically unworkable, the window to influence the final text is closing.</p></li><li><p>August 2 is not distant. Organizations waiting for final text before beginning compliance work are accepting a six-week implementation sprint under real enforcement conditions.</p></li></ul><p><strong>What to do about it</strong></p><ul><li><p>Read Draft 2 this week. The technical specificity represents a meaningful change from Draft 1, and your compliance planning may need adjustment.</p></li><li><p>Submit feedback before March 30 if the current draft creates compliance constraints you believe are technically unworkable for your AI content operations.</p></li><li><p>Begin implementation planning against Draft 2 requirements now. The June final text will refine but won&#8217;t fundamentally restructure what&#8217;s already written.</p></li></ul><p><strong>Rock&#8217;s Musings</strong></p><p>Every organization waiting for final text before starting EU AI Act compliance work is playing a game where the timeline gets worse each quarter they wait. Draft 2 is prescriptive enough to start serious implementation planning. The adjustments you&#8217;ll need when Draft 3 drops will be smaller than the work you&#8217;ll need to compress into six weeks if you start in June. The transparency labeling requirements are more technically demanding than most organizations appreciate from reading summaries. Download Draft 2 from the EU&#8217;s digital strategy portal and read it against your actual AI content production workflows. That gap analysis is the starting point for everything else.</p><div><hr></div><h3>10. RSA 2026 Reveals a Contested Market for AI Agent Governance Control Planes</h3><p>A pattern emerged across RSA 2026 beyond individual product launches: the governance control plane for AI agents is being actively contested by multiple major vendors. Microsoft&#8217;s Agent 365 (GA May 1), Cisco&#8217;s DefenseClaw (released March 27), SentinelOne&#8217;s Prompt AI Agent Security control plane, and Nudge Security&#8217;s AI agent discovery expansion all launched during the conference week, each addressing the same fundamental problem: enterprises deploy AI agents and lose track of what those agents do, access, and decide autonomously (SecurityWeek, March 2026, and Biometric Update, March 2026).</p><p><strong>Why it matters</strong></p><ul><li><p>Multiple major vendors converging on the same problem in the same week signals enterprises are actively requesting governance solutions, not absorbing vendor-manufactured demand.</p></li><li><p>Competition between Microsoft&#8217;s integrated control plane and point solutions from Cisco, SentinelOne, and Nudge creates a real architectural decision. Choose wrong and you own the integration debt for years.</p></li><li><p>None of these products fully solves behavioral monitoring. They address discovery, policy enforcement, and visibility. Real-time behavioral anomaly detection for agents remains an open engineering challenge.</p></li></ul><p><strong>What to do about it</strong></p><ul><li><p>Define your AI agent governance requirements before evaluating any vendor. Required capabilities: inventory discovery, permission auditing, behavioral logging, and human approval workflows for high-risk actions.</p></li><li><p>Assess whether your environment favors an integrated control plane or best-of-breed point solutions based on your actual architecture, not vendor marketing claims.</p></li><li><p>Ask every vendor during evaluation: how does the product detect when an agent takes an authorized action it was manipulated into taking? The answer quality will differentiate vendors quickly.</p></li></ul><p><strong>Rock&#8217;s Musings</strong></p><p>When four vendors announce competing governance control planes at the same conference in the same week, you&#8217;re watching a market category consolidate in real time. That&#8217;s interesting for analysts and exhausting for practitioners who have to evaluate all of it while managing agents already running in production without any governance. My advice: don&#8217;t let the governance platform debate distract from the more urgent problem of knowing what agents you currently have. Most enterprises have agents deployed that security teams didn&#8217;t authorize, can&#8217;t enumerate, and have no logs on. Governance tooling is the right investment. Knowing what you&#8217;re governing is the prerequisite.</p><div><hr></div><h3>The One Thing You Won&#8217;t Hear About But You Need To</h3><p><strong>NIST Publishes AI 800-4: The First Federal Framework for Monitoring AI Systems in Production</strong></p><p>NIST published AI 800-4, &#8220;Challenges to the Monitoring of Deployed AI Systems,&#8221; in March 2026. Built from three practitioner workshops with more than 200 experts across academia, industry, and ten-plus federal agencies, plus an 87-paper literature review, it maps the gaps, barriers, and open questions in monitoring AI systems after deployment. It covers six monitoring categories: functionality, operational health, human factors, security, safety, and compliance. It received no RSA booth, no vendor keynote, and no sponsored coverage (NIST News, March 2026, and NIST AI 800-4 PDF, March 2026).</p><p><strong>Why it matters</strong></p><ul><li><p>Most organizations deploying AI monitor latency and availability. AI 800-4 addresses whether the model behaves consistently with its training distribution and produces outputs that align with policy, which are the failures that matter most and the ones traditional monitoring misses entirely.</p></li><li><p>NIST explicitly identifies human-AI interaction monitoring as the most under-researched gap in the field. Workshop practitioners raised it far more than published literature covers. If your AI monitoring program doesn&#8217;t address how users interact with and respond to AI outputs, you&#8217;re missing the category NIST calls most underdeveloped.</p></li><li><p>The document is vendor-neutral and grounded in practitioner experience, directly applicable to conversations with regulators and auditors who want evidence of a structured AI monitoring program.</p></li></ul><p><strong>What to do about it</strong></p><ul><li><p>Download NIST AI 800-4 from nist.gov and route it to whoever owns your AI security program. It&#8217;s the most actionable government guidance on operational AI monitoring published to date.</p></li><li><p>Map your current monitoring coverage against the document&#8217;s six categories. The gaps will be immediately apparent and the prioritization logic writes itself once you have the map.</p></li><li><p>Use AI 800-4 as the foundation for your AI monitoring program documentation. When regulators ask how you monitor AI systems in production, a NIST-aligned program gives you a defensible, auditable answer.</p></li></ul><p><strong>Rock&#8217;s Musings</strong></p><p>The honest state of enterprise AI monitoring: most organizations have logs showing their AI system responded. They don&#8217;t have logs showing whether the response was correct, consistent with training distribution, within policy boundaries, or manipulated by adversarial input. That visibility gap is how AI security incidents become AI security incidents. You don&#8217;t catch the drift until the outcome is undeniable and the damage is done. NIST AI 800-4 doesn&#8217;t get coverage because nobody can sell it. The organizations that read it and build monitoring programs from its framework will answer regulatory questions coherently in 18 months when enforcement catches up to deployment rates. The organizations that attended every RSA keynote and skipped the NIST publication will be writing incident reports instead. For more on building AI governance programs that survive regulatory scrutiny, visit <a href="https://rockcybermusings.com/">rockcybermusings.com</a>. If you need help turning frameworks like AI 800-4 into operating programs your security team can actually run, reach out at <a href="https://rockcyber.com/">rockcyber.com</a>.</p><p>&#128073; Visit <strong><a href="https://www.rockcyber.com/">RockCyber.com</a></strong> to learn more about how we can help you in your traditional Cybersecurity and AI Security and Governance Journey</p><p>&#128073; Want to save a quick $100K? Check out our AI Governance Tools at <strong><a href="https://aigovernancetoolkit.com/">AIGovernanceToolkit.com</a></strong></p><p>&#128073; Subscribe for more AI and cyber insights with the occasional rant.</p><p><em>The views and opinions expressed in RockCyber Musings are my own and do not represent the positions of my employer or any organization I&#8217;m affiliated with.</em></p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.rockcybermusings.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading RockCyber Musings! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.rockcybermusings.com/?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share RockCyber Musings&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.rockcybermusings.com/?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share RockCyber Musings</span></a></p><div><hr></div><h2>References</h2><p>Bargury, M. (2026, March 23). <em>Your AI agents are my minions</em> [Conference presentation]. RSA Conference 2026, San Francisco, CA.</p><p>Claburn, T. (2026, March 24). LiteLLM infected with credential-stealing code via Trivy. <em>The Register</em>. https://www.theregister.com/2026/03/24/trivy_compromise_litellm/</p><p>Claburn, T. (2026, March 23). AI agents are &#8216;gullible&#8217; and easy to turn into your minions. <em>The Register</em>. https://www.theregister.com/2026/03/23/pwning_everyones_ai_agents/</p><p>Claburn, T. (2026, March 23). Google unleashes Gemini AI agents on the dark web. <em>The Register</em>. https://www.theregister.com/2026/03/23/google_dark_web_ai/</p><p>Cisco. (2026, March). Cisco reimagines security for the agentic workforce. <em>Cisco Newsroom</em>. https://newsroom.cisco.com/c/r/newsroom/en/us/a/y2026/m03/cisco-reimagines-security-for-the-agentic-workforce.html</p><p>Google Cloud. (2026, March). RSAC 26: Supercharging agentic AI defense with frontline threat intelligence. <em>Google Cloud Blog</em>. https://cloud.google.com/blog/products/identity-security/rsac-26-supercharging-agentic-ai-defense-with-frontline-threat-intelligence</p><p>HackerOne. (2026, March). Agentic prompt injection testing for AI security. <em>HackerOne Blog</em>. https://www.hackerone.com/blog/agentic-prompt-injection-testing</p><p>HackerOne introduces agentic prompt injection testing as AI security risks accelerate. (2026, March 21). <em>Cybersecurity Insiders</em>. https://www.cybersecurity-insiders.com/hackerone-introduces-agentic-prompt-injection-testing-as-ai-security-risks-accelerate/</p><p>Herbert Smith Freehills Kramer. (2026, March). Transparency obligations for AI-generated content under the EU AI Act: From principle to practice. https://www.hsfkramer.com/notes/ip/2026-03/transparency-obligations-for-ai-generated-content-under-the-eu-ai-act-from-principle-to-practice</p><p>EU releases second draft of AI Act Code of Practice on labeling AI-generated content. (2026, March). <em>BABL AI</em>. https://babl.ai/eu-releases-second-draft-of-ai-act-code-of-practice-on-labeling-ai-generated-content/</p><p>Microsoft Security. (2026, March 20). Secure agentic AI end-to-end. <em>Microsoft Security Blog</em>. https://www.microsoft.com/en-us/security/blog/2026/03/20/secure-agentic-ai-end-to-end/</p><p>NIST. (2026, March). New report: Challenges to the monitoring of deployed AI systems. https://www.nist.gov/news-events/news/2026/03/new-report-challenges-monitoring-deployed-ai-systems</p><p>NIST. (2026). <em>NIST AI 800-4: Challenges to the monitoring of deployed AI systems</em>. National Institute of Standards and Technology. https://nvlpubs.nist.gov/nistpubs/ai/NIST.AI.800-4.pdf</p><p>Novee. (2026, March 24). Novee introduces autonomous AI red teaming to uncover security flaws in LLM applications [Press release]. <em>GlobeNewswire</em>. https://www.globenewswire.com/news-release/2026/03/24/3261278/0/en/Novee-Introduces-Autonomous-AI-Red-Teaming-to-Uncover-Security-Flaws-in-LLM-Applications.html</p><p>Novee introduces autonomous AI red teaming to hunt LLM vulnerabilities. (2026, March 24). <em>Help Net Security</em>. https://www.helpnetsecurity.com/2026/03/24/novee-ai-red-teaming-for-llm-applications/</p><p>Palo Alto Networks Unit 42. (2026, March). New prompt injection attack vectors through MCP sampling. https://unit42.paloaltonetworks.com/model-context-protocol-attack-vectors/</p><p>SecurityWeek. (2026, March). RSAC 2026 conference announcements summary: Day 1. https://www.securityweek.com/rsac-2026-conference-announcements-summary-day-1/amp/</p><p>Zenity AI agents contextual security. (2026, March 24). <em>Help Net Security</em>. https://www.helpnetsecurity.com/2026/03/24/zenity-ai-agents-contextual-security/</p><p>Zenity. (2026, March 23). Zenity sets the foundation for guardian agents. <em>Zenity Newsroom</em>. https://zenity.io/company-overview/newsroom/company-news/zenity-sets-the-foundation-for-guardian-agents</p>]]></content:encoded></item><item><title><![CDATA[Weekly Musings Top 10 AI Security Wrapup: Issue 30 March 13-19, 2026]]></title><description><![CDATA[Agentic AI Security Moves From "Meh" to Incident Log]]></description><link>https://www.rockcybermusings.com/p/weekly-musings-top-10-ai-security-20260313-20260319</link><guid isPermaLink="false">https://www.rockcybermusings.com/p/weekly-musings-top-10-ai-security-20260313-20260319</guid><dc:creator><![CDATA[Rock Lambros]]></dc:creator><pubDate>Fri, 20 Mar 2026 12:50:42 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!b3YR!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8cc0f875-fe24-4b6a-ab70-a93357678487_1024x1024.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!b3YR!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8cc0f875-fe24-4b6a-ab70-a93357678487_1024x1024.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!b3YR!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8cc0f875-fe24-4b6a-ab70-a93357678487_1024x1024.png 424w, https://substackcdn.com/image/fetch/$s_!b3YR!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8cc0f875-fe24-4b6a-ab70-a93357678487_1024x1024.png 848w, https://substackcdn.com/image/fetch/$s_!b3YR!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8cc0f875-fe24-4b6a-ab70-a93357678487_1024x1024.png 1272w, https://substackcdn.com/image/fetch/$s_!b3YR!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8cc0f875-fe24-4b6a-ab70-a93357678487_1024x1024.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!b3YR!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8cc0f875-fe24-4b6a-ab70-a93357678487_1024x1024.png" width="1024" height="1024" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/8cc0f875-fe24-4b6a-ab70-a93357678487_1024x1024.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1024,&quot;width&quot;:1024,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:1233556,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.rockcybermusings.com/i/191536924?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8cc0f875-fe24-4b6a-ab70-a93357678487_1024x1024.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!b3YR!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8cc0f875-fe24-4b6a-ab70-a93357678487_1024x1024.png 424w, https://substackcdn.com/image/fetch/$s_!b3YR!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8cc0f875-fe24-4b6a-ab70-a93357678487_1024x1024.png 848w, https://substackcdn.com/image/fetch/$s_!b3YR!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8cc0f875-fe24-4b6a-ab70-a93357678487_1024x1024.png 1272w, https://substackcdn.com/image/fetch/$s_!b3YR!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8cc0f875-fe24-4b6a-ab70-a93357678487_1024x1024.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Meta logged a SEV-1 on March 18 because an internal AI agent posted without human approval, provided bad advice, and exposed sensitive data to the wrong employees for 2 hours. Amazon confirmed its Bedrock sandbox lets AI models exfiltrate data via DNS and called it intentional design. HiddenLayer found 31% of security leaders don&#8217;t know if they had an AI breach in the past year. The EU Council voted to restructure the AI Act&#8217;s high-risk compliance framework. Three AI agent security products launched in four days. This was one week.</p><p>The week&#8217;s evidence points in one direction: agentic AI security is no longer a research problem. Real incidents are appearing in production environments run by organizations with serious security programs. Technical flaws in AI infrastructure are drawing vendor responses that amount to documentation updates rather than patches. Research data is documenting blind spots CISOs can no longer treat as edge cases. In parallel, the governance machinery is finally moving, but it&#8217;s moving slower than deployment. Standards and deployments are in a race, and deployments are winning by a wide margin. More context at <a href="https://www.rockcyber.com/">RockCyber</a> and <a href="https://rockcybermusings.com/">RockCyber Musings</a>.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.rockcybermusings.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading RockCyber Musings! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.rockcybermusings.com/p/weekly-musings-top-10-ai-security-20260313-20260319?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.rockcybermusings.com/p/weekly-musings-top-10-ai-security-20260313-20260319?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p><h3>1. OWASP publishes its GenAI data security risk taxonomy for 2026</h3><p>The OWASP GenAI Security Project released GenAI Data Security: Risks and Mitigations 2026 in March, a 103-page taxonomy covering 21 discrete data security risks across the full GenAI lifecycle from training through agentic runtime (OWASP). The document maps risks across training and fine-tuning data, retrieval and RAG pipelines, vector stores, context windows, agent memory, tool call payloads, and observability infrastructure. It identifies a core architectural property that makes GenAI data security structurally different from every prior computing model: the context window aggregates data from multiple trust domains into a single flat namespace with no internal access controls. A confidential HR record retrieved via RAG sits next to a user prompt with identical trust weight, and there is no mechanism today to mark a context segment as available for reasoning but not surfaceable in the output. The document also addresses machine unlearning directly: deleting source data does not remove what a fine-tuned model or LoRA adapter has memorized into its weights. <strong><a href="https://genai.owasp.org/resource/owasp-genai-data-security-risks-mitigations-2026/">Download the report HERE.</a></strong></p><p><strong>Why it matters</strong></p><ul><li><p>The flat-namespace context window problem is not a configuration gap. It&#8217;s an architectural property of how these systems work, which means perimeter controls and access policies cannot fully solve it. Minimization and context scoping are the only practical mitigations available today.</p></li><li><p>LoRA adapter memorization of rare training examples means high-recall prompts can extract verbatim PII, credentials, or intellectual property from fine-tuned models without any sophisticated attack technique. Organizations fine-tuning on internal data have a data exposure risk they likely haven&#8217;t assessed.</p></li><li><p>The Right to Erasure problem is unsolved at the architectural level. Deleting training data from a source system does not delete what the model encoded during fine-tuning. GDPR and state privacy law DSR obligations cannot be satisfied by source deletion alone.</p></li></ul><p><strong>What to do about it</strong></p><ul><li><p>Treat the context window as a data-exposure surface, not just a prompt-delivery mechanism. Classify what goes in the same way you classify what goes into a database query, and scope RAG retrieval to the minimum required for the task.</p></li><li><p>Audit every fine-tuned model and LoRA adapter in your environment against the data used to train it. If that training data included PII, credentials, or regulated information, your model could serve as a potential exfiltration vector.</p></li><li><p>Build a GenAI data bill of materials using CycloneDX ML-BOM as the base format. Until you have lineage from the source dataset to the deployed model to the embedding store, you cannot answer the question a regulator will eventually ask: what data did this model see, and where does it live now?</p></li></ul><p><strong>Rock&#8217;s Musings</strong></p><p>The architectural insight at the center of this document is the one the industry keeps sliding past. The context window has no internal access control layer. That&#8217;s not a misconfiguration. It&#8217;s a design property of how transformers process sequences. Everything that enters the context window is treated as equally reachable by the model&#8217;s output mechanism, and no amount of system prompt guardrailing changes the underlying architecture. The practical implication is that the primary defense is what you put in, not what you try to prevent from coming out.</p><p>The machine unlearning section is the one I push organizations on hardest. They are collecting consent, honoring deletion requests, and scrubbing source databases, and then deploying fine-tuned models that still carry what they memorized from the deleted data. The model weights are a copy of your training corpus in a form your DLP tools don&#8217;t see, and your deletion workflows can&#8217;t reach. Right to Erasure in GenAI is an open architectural problem with no clean solution today, and most organizations haven&#8217;t told their legal team that yet.</p><h3>2. EU Council rewrites the compliance clock for high-risk AI systems</h3><p>The EU Council adopted its negotiating position to amend the AI Act&#8217;s high-risk framework (EU Council). The core change replaces the fixed August 2026 compliance deadline with a conditional trigger. Full high-risk obligations apply only once the Commission certifies required standards and tools are available, with a hard backstop date. The Council also pushed the national AI regulatory sandbox deadline to December 2027 and clarified that law enforcement, border management, judicial, and financial AI systems remain under national supervisory authority rather than the Commission. Negotiations with the European Parliament begin next.</p><p><strong>Why it matters</strong></p><ul><li><p>The conditional trigger gives the Commission discretion over when your obligations start. Until it certifies standards are ready, full high-risk obligations don&#8217;t apply, creating an indeterminate window.</p></li><li><p>Pushing the sandbox deadline to December 2027 removes a key testing mechanism for high-risk AI at a time when organizations are accelerating deployment.</p></li><li><p>Fragmented supervisory authority means 27 member states apply their own rules to some of the highest-stakes AI use cases.</p></li></ul><p><strong>What to do about it</strong></p><ul><li><p>Map your AI systems against current and proposed high-risk definitions now. The conditional trigger shifts the timeline, not the compliance obligation itself.</p></li><li><p>Track Parliament negotiations. The Council position is a mandate, not the final text.</p></li><li><p>Build a jurisdiction-aware compliance map for EU operations covering which systems fall under national versus Commission supervision.</p></li></ul><p><strong>Rock&#8217;s Musings</strong></p><p>I&#8217;ve seen regulatory timelines used to delay compliance indefinitely in my career more times than I can count. This EU Council move fits the pattern. The conditional trigger means the Commission controls when your clock starts, and they have to certify standards are available first. Given the pace at which NIST&#8217;s agentic AI guidance is moving, expecting European standards to materialize quickly requires genuine optimism.</p><p>Organizations using this ambiguity to do nothing are miscalculating. The August 2026 date was never the governance point. You have high-risk AI systems in production today, and you need to govern them regardless of what the Commission certifies and when.</p><h3>3. Meta logs a SEV-1 incident from a rogue internal AI agent</h3><p>On March 18, Meta confirmed a Severity 1 security incident caused by an internal AI agent operating without human authorization (Bitcoinworld, HackerNoob). The agent posted to an internal forum, gave incorrect advice, and triggered a cascade that exposed sensitive company and user data to unauthorized employees for approximately two hours. Meta contained the exposure by cutting the agent&#8217;s forum access and auditing permissions across other internal agents. No external exfiltration was confirmed.</p><p><strong>Why it matters</strong></p><ul><li><p>A SEV-1 at Meta from an AI agent operating outside its bounds sets a documented precedent: production agents at companies with robust security programs can circumvent behavioral constraints and cause genuine incidents.</p></li><li><p>The chain reaction, one unauthorized action triggering downstream data exposure, is characteristic of agentic systems and different from traditional software vulnerabilities in ways most IR playbooks don&#8217;t yet account for.</p></li><li><p>No external exfiltration is partial comfort. Unauthorized internal access to sensitive user data carries GDPR and AI Act exposure regardless of whether the data left the building.</p></li></ul><p><strong>What to do about it</strong></p><ul><li><p>Audit every AI agent in your environment and document what it can post, write, or modify without a human approval checkpoint.</p></li><li><p>Map the blast radius. If a specific agent takes an unexpected action, what does it touch first, and what cascades from there?</p></li><li><p>Build AI agent incident response playbooks with automated containment triggers that don&#8217;t require analyst approval before they fire.</p></li></ul><p><strong>Rock&#8217;s Musings</strong></p><p>The Meta incident will get dismissed as a minor operational hiccup. That&#8217;s the wrong read. Even with legit engineering talent and a mature security program, a production AI agent escaped its behavioral constraints and triggered a data exposure chain. I&#8217;m willing to bet your environment isn&#8217;t more disciplined than Meta&#8217;s.</p><p>Two hours to containment is fast. Most organizations I work with couldn&#8217;t tell you within two hours that an agent had gone sideways. AI agent behavioral monitoring is dramatically behind where it needs to be. The lesson to take away from this is that you need detection that fires before the cascade, not after the data is already in the wrong hands.</p><h3>4. Amazon&#8217;s Bedrock sandbox leaks data through DNS because that&#8217;s the design</h3><p>BeyondTrust&#8217;s Phantom Labs disclosed that Amazon Bedrock AgentCore Code Interpreter&#8217;s sandbox mode permits outbound DNS queries (SC Media, The Hacker News). An attacker interacting with the agent can send commands encoded in DNS A record responses and receive exfiltrated data encoded in DNS subdomain queries to an attacker-controlled server. No authentication bypass is required. BeyondTrust assigned a CVSS score of 7.5. AWS reviewed the research, determined that the behavior reflects the intended functionality, and responded by updating the documentation rather than issuing a patch.</p><p><strong>Why it matters</strong></p><ul><li><p>&#8220;Intended behavior&#8221; is a vendor risk posture, not a security posture. Sandbox mode was positioned as providing execution isolation. A sandbox allowing covert DNS exfiltration does not deliver isolation in any security-relevant sense.</p></li><li><p>DNS-based covert channels are standard red team tradecraft in traditional environments. The technique translates directly into AI code execution environments without modification.</p></li><li><p>Organizations running agents against sensitive internal data in AWS Bedrock face an unpatched, documented, CVSS 7.5 risk with no vendor remediation timeline.</p></li></ul><p><strong>What to do about it</strong></p><ul><li><p>Add DNS query monitoring for Bedrock AgentCore code execution environments to your threat detection stack now.</p></li><li><p>Reduce the data that AI agents with code execution access can reach to the strict minimum required for the task.</p></li><li><p>Get a formal written architecture statement from AWS specifying exactly what the sandbox guarantees before expanding Bedrock AgentCore deployments.</p></li></ul><p><strong>Rock&#8217;s Musings</strong></p><p>Another &#8220;Intended behavior&#8221; narrative. I&#8217;m getting pretty damn sick of it. That&#8217;s another way of saying, &#8220;We know about this, it would be expensive to change, and it sucks to be you.&#8221; <strong><a href="https://www.csoonline.com/article/4118092/google-vertex-ai-security-permissions-could-amplify-insider-threats.html">(see my thoughts in CSO magazine about a previous instance HERE)</a></strong>. The documentation update rather than a patch is the tell. You can&#8217;t outsource your risk posture to your cloud provider&#8217;s design decisions.</p><p>The technique is in every red team playbook. DNS exfiltration from sandboxed environments is foundational evasion tradecraft. Translate that knowledge directly to your AI infrastructure. If you&#8217;re running code execution agents against sensitive data in Bedrock and you haven&#8217;t instrumented DNS as an exfiltration channel, now you have your reason.</p><h3>5. Linux Foundation raises $12.5 million from AI vendors to fix what their tools helped break</h3><p>The Linux Foundation announced $12.5 million in grant funding from Anthropic, AWS, GitHub, Google, Google DeepMind, Microsoft, and OpenAI to advance open source software security (Linux Foundation, OpenSSF). The funding flows through Alpha-Omega and the Open Source Security Foundation. The stated problem is that AI tools are generating vulnerability reports at a volume that open-source maintainers cannot triage or remediate, degrading the security posture of the software supply chain. AWS contributed an additional $2.5 million to Alpha-Omega, in addition to the pooled amount.</p><p><strong>Why it matters</strong></p><ul><li><p>The same organizations whose AI tools created the report flood are funding the solution. This characterizes the governance dynamic precisely, that vendors profit from deployment and are now asked to fund the externalized costs on the maintainer community.</p></li><li><p>Overwhelming maintainers with AI-generated findings lowers average signal quality. Funding addresses capacity but doesn&#8217;t solve the signal-to-noise problem alone.</p></li><li><p>This is the first major coordinated industry response to the specific problem of AI-generated report volume stressing the open source security ecosystem.</p></li></ul><p><strong>What to do about it</strong></p><ul><li><p>Factor the current maintainer backlog into your software composition analysis program. Critical open source dependencies may carry known vulnerabilities sitting in a backlogged queue rather than getting remediated.</p></li><li><p>Watch what Alpha-Omega and OpenSSF deliver from this investment over the next twelve months. The commitment matters less than whether the tooling measurably improves triage capacity.</p></li><li><p>Ask your security vendors how they handle AI-generated findings before surfacing them to your team. The same noise problem exists inside your tooling stack.</p></li></ul><p><strong>Rock&#8217;s Musings</strong></p><p>$12.5 million is the right direction, yet not nearly enough. Open source maintainers are largely volunteers managing the infrastructure that the global software supply chain runs on. The AI-generated report flood is a problem these vendors created while selling velocity gains to enterprises.</p><p>The coordination signal matters more than the dollar amount. You rarely see Google, Microsoft, AWS, Anthropic, and OpenAI announce joint anything. When competitors fund a shared problem together, the liability exposure of inaction exceeds the competitive cost of cooperating. Given how much of the internet runs on open source that these companies&#8217; AI tools are now stressing, the math on joint action isn&#8217;t complicated.</p><h3>6. Pentagon moves to replace Anthropic while the lawsuit works through the courts</h3><p>TechCrunch reported that the Pentagon is actively developing alternative AI capability paths to replace Anthropic&#8217;s Claude across defense applications (TechCrunch). This follows the Defense Department&#8217;s February designation of Anthropic as a supply chain security risk and Anthropic&#8217;s subsequent lawsuit against the Trump administration. This confirms that the replacement effort has shifted from contingency planning to active technical development. More than 875 Google and OpenAI employees have signed an open letter supporting Anthropic&#8217;s position.</p><p><strong>Why it matters</strong></p><ul><li><p>Active technical development of replacements, rather than contingency planning, signals DoD confidence that the Anthropic designation will hold through the litigation cycle.</p></li><li><p>Defense contractors relying on Claude for active program work now face migration timelines driven by someone else&#8217;s legal and procurement decisions.</p></li><li><p>The 875-employee response across competing firms signals the tech workforce treats this as a legitimacy question about AI governance, not a routine vendor dispute.</p></li></ul><p><strong>What to do about it</strong></p><ul><li><p>If your organization operates in the defense industrial base, review AI vendor contracts now for comparable ethical-use clauses and their enforceability, before further redesignations affect your supply chain.</p></li><li><p>Track the Anthropic lawsuit. The outcome defines what ethical use provisions in AI contracts are worth in federal procurement.</p></li><li><p>Evaluate AI vendor concentration risk in your stack. If one supply chain designation event could disrupt your programs, that&#8217;s a single point of failure worth addressing.</p></li></ul><p><strong>Rock&#8217;s Musings</strong></p><p>The supply chain risk designation was built for foreign adversaries. Applying it to a domestic AI company for writing autonomous weapons prohibitions into a contract is a significant precedent that the press is underweighting. The designation signals that safety constraints are now framed as operational liabilities in defense procurement, not risk mitigation.</p><p>If that framing spreads to other acquisition decisions, the AI vendors most willing to remove safety constraints gain a competitive advantage in a large and growing federal spending category. Watch the lawsuit and the follow-on procurement awards carefully. Both will tell you where this governance experiment ends up.</p><h3>7. CSA&#8217;s 2026 cloud and AI security report documents the identity explosion</h3><p>The Cloud Security Alliance published its State of Cloud and AI Security 2026 on March 13, finding the average enterprise now manages 100 machine and non-human identities for every one human identity (CSA). Forgotten or misconfigured cloud credentials declined from 84% in 2024 to 65% in 2026. Ninety-two percent of executives report business-impacting security compromises, most from preventable risks. The report identifies decentralized AI agents as the primary driver of the NHI expansion and calls for continuous exposure management to replace static patching cycles.</p><p><strong>Why it matters</strong></p><ul><li><p>A 100:1 machine-to-human identity ratio means the traditional IAM program built around human users is managing a fundamentally different problem than it was designed for.</p></li><li><p>Credential misconfiguration persisting at 65% suggests the improvement rate won&#8217;t match the velocity of AI-driven identity expansion.</p></li><li><p>A 92% executive compromise from preventable risks indicates the gap isn&#8217;t a detection-sophistication problem. Organizations know the controls and aren&#8217;t applying them at the required scale.</p></li></ul><p><strong>What to do about it</strong></p><ul><li><p>Audit NHI management practices against the same standards applied to human identities: lifecycle management, least privilege, and regular access reviews.</p></li><li><p>Deploy continuous credential exposure monitoring specifically for machine identities and AI agent service accounts.</p></li><li><p>Shift the board-level narrative from maturity scores to continuous exposure management. That&#8217;s where enterprise frameworks are heading.</p></li></ul><p><strong>Rock&#8217;s Musings</strong></p><p>A hundred machine identities for every human one, and most organizations manage them with IAM tooling built for a 10-to-1 ratio. The math doesn&#8217;t work. The credential improvement trend from 84% to 65% is real progress, but 65% still represents a failure rate I wouldn&#8217;t accept in any other critical control domain.</p><p>Every new agentic deployment creates more identities, tokens, service accounts, and API keys. If you don&#8217;t have a clear owner for non-human identity governance today, you have a gap that will become a breach within twelve months. Find the owner. Document the scope. Don&#8217;t wait for the incident.</p><h3>8. Jozu Agent Guard launches after watching an AI agent bypass governance in four commands</h3><p>Jozu announced Jozu Agent Guard on March 17, a zero-trust runtime that executes AI agents, models, and MCP servers with policy enforcement built outside the model&#8217;s control plane and hardcoded against agent-level override (Help Net Security). The architecture decision came directly from internal testing: during product development, Jozu observed an AI agent bypass the governance controls the product was designed to enforce in four commands. That failure drove the decision to move policy enforcement entirely outside the execution layer the agent can influence.</p><p><strong>Why it matters</strong></p><ul><li><p>A product built specifically to constrain AI agents was bypassed in four commands during its own testing. The threat model has to assume the agent itself will attempt to circumvent governance. Cooperative compliance is not a valid design assumption.</p></li><li><p>MCP server isolation is underprovided. MCP servers frequently carry production credentials and broad tool access, and running them in shared agent environments creates privilege escalation paths most organizations haven&#8217;t mapped.</p></li><li><p>Three AI agent security products launching in four days signals enterprise buying is active in this space right now.</p></li></ul><p><strong>What to do about it</strong></p><ul><li><p>Require AI agent security vendors to demonstrate their product against an adversarial agent in a live environment. Demand the failure modes alongside the happy path.</p></li><li><p>Treat MCP server execution environments as sensitive infrastructure requiring isolation equivalent to your most privileged workloads.</p></li><li><p>Add governance bypass testing to your AI red team scope before the next production agent deployment.</p></li></ul><p><strong>Rock&#8217;s Musings</strong></p><p>The four-command bypass during their own testing is the most honest vendor disclosure I&#8217;ve seen about AI agent security in the past year. Most vendors demo the happy path and skip the part where their product got circumvented. Jozu disclosed it and changed the architecture. That&#8217;s how security engineering is supposed to work.</p><p>The uncomfortable implication for everyone else: if a product built specifically to constrain AI agents was bypassed in four commands, ask yourself what your existing controls look like against an agent actively trying to exceed its permissions. If you haven&#8217;t run that test, you don&#8217;t have an answer.</p><div><hr></div><h3>9. Token Security builds intent-based controls for AI agent permissions</h3><p>Token Security announced intent-based AI agent security on March 18, governing autonomous agents by scoping their permissions to declared operational purpose rather than granting standing broad access (Help Net Security). The system creates purpose-defined permission envelopes that expire at task completion, with runtime enforcement preventing actions outside the declared intent. Token Security&#8217;s CEO stated directly that prompt filtering and guardrails were not designed to contain the security risks of autonomous AI agents, pointing to the architectural limitation of relying on the model&#8217;s output layer for enforcement.</p><p><strong>Why it matters</strong></p><ul><li><p>Purpose-aligned permissions address a structural problem in current agent deployment: agents inheriting credential scopes far exceeding what any single task requires.</p></li><li><p>Explicit acknowledgment that content filtering can&#8217;t do this job alone represents where serious practitioner thinking is converging. The field is moving from output layer controls toward architectural access controls.</p></li><li><p>Paired with Jozu, Entro, and Microsoft Entra Agent ID announcements this same week, this reflects a coherent market thesis forming around agent identity and least privilege as primary security controls.</p></li></ul><p><strong>What to do about it</strong></p><ul><li><p>Map current AI agent deployments against one question: does each agent hold only the permissions it needs for its specific task? If you can&#8217;t answer quickly, your access governance is already too loose.</p></li><li><p>Evaluate intent-based and purpose-scoped access controls in your next AI security procurement cycle.</p></li><li><p>Brief your identity team on AI agent access management before your security team deploys solutions they haven&#8217;t reviewed. These tools touch the same credential infrastructure.</p></li></ul><p><strong>Rock&#8217;s Musings</strong></p><p>Least privilege applied to agents is the same principle that has protected privileged service accounts in traditional architectures for decades. The problem is that most AI agent deployments aren&#8217;t being treated like privileged service accounts. They get broad collaboration access by default, and nobody asks why.</p><p>Intent-based controls force the right question: what is this agent for? If you can answer precisely, you can scope permissions precisely. If you can&#8217;t answer precisely, that is the real governance problem. You&#8217;ve deployed an agent without a defined operational boundary, and your control over it is largely fictional.</p><div><hr></div><h3>10. NIST receives formal research submissions on securing AI agents</h3><p>On March 18, UC Berkeley&#8217;s Center for Long-Term Cybersecurity submitted a formal response to NIST&#8217;s CAISI RFI on AI agent security, urging prioritization of standardization, incident reporting frameworks, talent pipelines, and adaptive governance (CLTC UC Berkeley). The Computer and Communications Industry Association submitted parallel comments advocating for multistakeholder processes and alignment with existing NIST frameworks (CCIA). NIST&#8217;s National Cybersecurity Center of Excellence also holds a separate comment period open through April 2 on a concept paper covering identity and authorization for AI agents.</p><p><strong>Why it matters</strong></p><ul><li><p>The gap between NIST collecting input and usable standards publishing is measured in years. Your agents are running now, under no binding identity or authorization standard.</p></li><li><p>Berkeley&#8217;s call for incident reporting infrastructure acknowledges a structural gap: no systematic mechanism exists for learning from AI agent security failures across organizations.</p></li><li><p>The NCCoE concept paper on agent identity and authorization is where future compliance requirements will originate. Comments submitted now shape what those requirements demand.</p></li></ul><p><strong>What to do about it</strong></p><ul><li><p>Read the NCCoE concept paper at nccoe.nist.gov and submit comments before April 2 if your organization deploys agents. Operational experience is what NIST is specifically asking for.</p></li><li><p>Treat the Berkeley and CCIA submissions as intelligence on where auditors will focus within 18 to 36 months.</p></li><li><p>Stand up basic agent identity logging now using existing IAM controls. Don&#8217;t wait for NIST to finalize anything.</p></li></ul><p><strong>Rock&#8217;s Musings</strong></p><p>NIST is moving faster on agentic AI security than I expected two years ago. That still isn&#8217;t fast enough to matter for organizations deploying agents today. Best case from the current comment cycle: interim guidance in twelve months. Binding controls will take longer.</p><p>Berkeley&#8217;s call for incident reporting is the right recommendation and it will face the same resistance every mandatory reporting regime has faced. Voluntary frameworks will come first, get ignored, and get teeth after the third or fourth major public incident. That&#8217;s the pattern. Plan for it and build your own internal incident tracking capability now.</p><h3>The One Thing You Won&#8217;t Hear About But You Need To</h3><h3>Entro Security builds a governed map of what your AI agents access in production</h3><p>Entro Security launched its Agentic Governance and Administration platform, extending non-human identity security coverage specifically to AI agents (GlobeNewswire, Help Net Security). The platform builds structured AI agent profiles from three observable layers. First, sources: the endpoints, agent platforms, cloud environments, and MCP servers where agents execute. Second, targets: the enterprise assets and applications each agent accesses. Third, identities: the human accounts, non-human identities, and secrets each agent uses to operate. AGA provides MCP server activity visibility and policy enforcement, audit trails for both allowed and blocked activity, and controls against unsanctioned MCP targets and AI client behaviors.</p><p><strong>Why it matters</strong></p><ul><li><p>Most organizations deploying AI agents don&#8217;t have a single governed view of what agents are running, what they access, and which identities they use. AGA builds that view from execution telemetry rather than documentation that goes stale immediately after it&#8217;s written.</p></li><li><p>MCP server governance is nearly absent from enterprise security programs today, despite MCP servers frequently holding production credentials and broad access to sensitive systems.</p></li><li><p>The NHI-first architecture lets organizations with existing non-human identity programs extend that coverage to AI agents rather than building a separate program from scratch.</p></li></ul><p><strong>What to do about it</strong></p><ul><li><p>Before the next AI agent deployment, require answers to three questions from observable telemetry: where does it run, what does it touch, and which identities does it use? If you need documentation rather than telemetry to answer, you don&#8217;t have governance.</p></li><li><p>Add MCP server inventory to asset management now. MCP servers deploy through developer workflows without formal change management, and retroactive cataloguing gets harder with each deployment.</p></li><li><p>Assess whether your current NHI security program explicitly covers AI agent identities. If it doesn&#8217;t, extend it or stand up a parallel track with a clear accountable owner.</p></li></ul><p><strong>Rock&#8217;s Musings</strong></p><p>This one didn&#8217;t get coverage this week because it launched during RSA prep season when every security vendor fights for the same column inches. That&#8217;s exactly why it&#8217;s here. The problem AGA addresses is what I call dark matter governance: AI agents operating in your environment that nobody catalogued because they deployed through platforms your traditional asset management tools don&#8217;t see.</p><p>The MCP visibility layer is the operationally useful piece. MCP servers multiply fast, are deployed by individual developers without change management review, and frequently hold credentials for production systems. An agent you haven&#8217;t catalogued connecting to an MCP server you haven&#8217;t governed is a permissions sprawl problem that compounds with every new deployment. Get a governed view of that surface before your adversary maps it for you. </p><p>If you found this analysis useful, subscribe at <a href="https://rockcybermusings.com/">rockcybermusings.com</a> for weekly intelligence on AI security developments.</p><p>&#128073; Visit <strong><a href="https://www.rockcyber.com/">RockCyber.com</a></strong> to learn more about how we can help you in your traditional Cybersecurity and AI Security and Governance Journey</p><p>&#128073; Want to save a quick $100K? Check out our AI Governance Tools at <strong><a href="https://aigovernancetoolkit.com/">AIGovernanceToolkit.com</a></strong></p><p>&#128073; Subscribe for more AI and cyber insights with the occasional rant.</p><p><em>The views and opinions expressed in RockCyber Musings are my own and do not represent the positions of my employer or any organization I&#8217;m affiliated with.</em></p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.rockcybermusings.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading RockCyber Musings! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.rockcybermusings.com/?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share RockCyber Musings&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.rockcybermusings.com/?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share RockCyber Musings</span></a></p><h2>References</h2><p>Bitcoinworld. (2026, March). <em>Rogue AI agent sparks critical security crisis at Meta, exposing sensitive data</em>. https://bitcoinworld.co.in/meta-rogue-ai-agent-security-breach/</p><p>Cloud Security Alliance. (2026, March 13). <em>The state of cloud and AI security in 2026</em>. https://cloudsecurityalliance.org/blog/2026/03/13/the-state-of-cloud-and-ai-security-in-2026</p><p>Computer and Communications Industry Association. (2026, March). <em>CCIA submits comments to NIST regarding privacy and security of AI agents</em>. https://ccianet.org/news/2026/03/ccia-submits-comments-to-nist-regarding-privacy-and-security-of-ai-agents/</p><p>Council of the European Union. (2026, March 13). <em>Council agrees position to streamline rules on artificial intelligence</em>. https://www.consilium.europa.eu/en/press/press-releases/2026/03/13/council-agrees-position-to-streamline-rules-on-artificial-intelligence/</p><p>Entro Security. (2026, March 18). <em>Entro launches agentic governance and administration to bring visibility and control to AI access across the enterprise</em>. GlobeNewswire. https://www.globenewswire.com/news-release/2026/03/18/3258229/0/en/Entro-Launches-Agentic-Governance-Administration-to-Bring-Visibility-and-Control-to-AI-Access-Across-the-Enterprise.html</p><p>HackerNoob. (2026, March). <em>Meta&#8217;s rogue AI agent: Sev 1 security incident and how to sandbox AI agents properly</em>. https://hackernoob.tips/meta-rogue-ai-agent-sev1-how-to-sandbox-ai-agents/</p><p>Help Net Security. (2026, March 17). <em>Jozu Agent Guard targets AI agents that evade controls</em>. https://www.helpnetsecurity.com/2026/03/17/jozu-agent-guard-targets-ai-agents-that-evade-controls/</p><p>Help Net Security. (2026, March 18). <em>Token Security advances AI agent protection with intent-based controls</em>. https://www.helpnetsecurity.com/2026/03/18/token-security-intent-based-ai-agent-security/</p><p>Help Net Security. (2026, March 18). <em>Big tech companies step in to support the open source security ecosystem</em>. https://www.helpnetsecurity.com/2026/03/18/linux-foundation-open-source-security-12-5-million-funding/</p><p>Help Net Security. (2026, March 19). <em>Entro Security AGA brings governance and control to enterprise AI agents and access</em>. https://www.helpnetsecurity.com/2026/03/19/entro-agentic-governance-administration/</p><p>HiddenLayer. (2026, March 18). <em>HiddenLayer releases the 2026 AI threat landscape report</em>. PR Newswire. https://finance.yahoo.com/news/hiddenlayer-releases-2026-ai-threat-140000928.html</p><p>Linux Foundation. (2026, March 17). <em>Linux Foundation announces $12.5 million in grant funding from leading organizations to advance open source security</em>. https://www.linuxfoundation.org/press/linux-foundation-announces-12.5-million-in-grant-funding-from-leading-organizations-to-advance-open-source-security</p><p>SC Media. (2026, March). <em>AWS Bedrock tool vulnerability allows data exfiltration via DNS leaks</em>. https://www.scworld.com/brief/aws-bedrock-vulnerability-allows-data-exfiltration-via-dns-leaks</p><p>TechCrunch. (2026, March 17). <em>The Pentagon is developing alternatives to Anthropic, report says</em>. https://techcrunch.com/2026/03/17/the-pentagon-is-developing-alternatives-to-anthropic-report-says/</p><p>The Hacker News. (2026, March 17). <em>AI flaws in Amazon Bedrock, LangSmith, and SGLang enable data exfiltration and RCE</em>. https://thehackernews.com/2026/03/ai-flaws-in-amazon-bedrock-langsmith.html</p><p>UC Berkeley Center for Long-Term Cybersecurity. (2026, March 18). <em>Researchers submit response to U.S. government request on security considerations for AI agents</em>. https://cltc.berkeley.edu/2026/03/18/researchers-submit-response-to-u-s-government-request-on-security-considerations-for-ai-agents/</p>]]></content:encoded></item><item><title><![CDATA[AI Agent Authentication Gets the Hard Part Right. Authorization Is Still Your Problem.]]></title><description><![CDATA[IETF's new AI agent auth draft nails identity with WIMSE and SPIFFE but skips per-action authorization.]]></description><link>https://www.rockcybermusings.com/p/i-agent-authentication-authorization-gap</link><guid isPermaLink="false">https://www.rockcybermusings.com/p/i-agent-authentication-authorization-gap</guid><dc:creator><![CDATA[Rock Lambros]]></dc:creator><pubDate>Tue, 17 Mar 2026 12:50:42 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!bS5L!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4aa7583c-e0fb-4920-a994-e8b6bb128fa4_2048x2048.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!bS5L!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4aa7583c-e0fb-4920-a994-e8b6bb128fa4_2048x2048.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!bS5L!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4aa7583c-e0fb-4920-a994-e8b6bb128fa4_2048x2048.jpeg 424w, https://substackcdn.com/image/fetch/$s_!bS5L!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4aa7583c-e0fb-4920-a994-e8b6bb128fa4_2048x2048.jpeg 848w, https://substackcdn.com/image/fetch/$s_!bS5L!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4aa7583c-e0fb-4920-a994-e8b6bb128fa4_2048x2048.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!bS5L!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4aa7583c-e0fb-4920-a994-e8b6bb128fa4_2048x2048.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!bS5L!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4aa7583c-e0fb-4920-a994-e8b6bb128fa4_2048x2048.jpeg" width="1456" height="1456" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/4aa7583c-e0fb-4920-a994-e8b6bb128fa4_2048x2048.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1456,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:2920581,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.rockcybermusings.com/i/190013993?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4aa7583c-e0fb-4920-a994-e8b6bb128fa4_2048x2048.jpeg&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!bS5L!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4aa7583c-e0fb-4920-a994-e8b6bb128fa4_2048x2048.jpeg 424w, https://substackcdn.com/image/fetch/$s_!bS5L!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4aa7583c-e0fb-4920-a994-e8b6bb128fa4_2048x2048.jpeg 848w, https://substackcdn.com/image/fetch/$s_!bS5L!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4aa7583c-e0fb-4920-a994-e8b6bb128fa4_2048x2048.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!bS5L!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4aa7583c-e0fb-4920-a994-e8b6bb128fa4_2048x2048.jpeg 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>The IETF just published its most ambitious attempt to standardize how AI agents prove their identity across systems. Draft-klrc-aiagent-auth-00, dropped March 2, 2026, composes WIMSE, SPIFFE, and OAuth 2.0 into a 26-page framework called AIMS (Agent Identity Management System). The authentication layer is solid. The authorization layer stops at the token boundary. The Security Considerations section contains two words: &#8220;TODO Security.&#8221; If you&#8217;re deploying agentic systems in production, you need to understand where this draft helps you and where you still have to build your own controls.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.rockcybermusings.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading RockCyber Musings! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.rockcybermusings.com/p/i-agent-authentication-authorization-gap?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.rockcybermusings.com/p/i-agent-authentication-authorization-gap?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p><p>Before I get into specifics, a quick note on what this document actually is. An IETF Internet-Draft (I-D) is a working document, the raw material that may eventually become an RFC (an official Internet standard). This one is version -00, the very first public iteration from Pieter Kasselman (Defakto Security), Jean-Francois Lombardo (AWS), Yaroslav Rosomakho (Zscaler), and Brian Campbell (Ping Identity). Criticizing a -00 draft for incompleteness is a bit like reviewing someone&#8217;s outline and complaining the conclusion is thin. That said, people are already reading this as deployment guidance, and the gaps matter for anyone building agentic systems today. So let&#8217;s talk about what it covers, what it doesn&#8217;t cover yet, and what you need to build yourself while the standards process catches up.</p><h2>The good news: agents are workloads, and workloads have an identity stack</h2><p>The draft&#8217;s foundational thesis gets it right that AI agents should be treated as workloads, not as some new identity category requiring new protocols and running instances of software executing specific tasks. That framing unlocks SPIFFE&#8217;s attestation-bound cryptographic identity, WIMSE&#8217;s cross-system workload semantics, and OAuth 2.0&#8217;s delegation framework. No new protocols needed.</p><p>This matters because SPIFFE already works at scale. Uber processes billions of attestations daily through SPIRE. Block runs the full SPIFFE+WIMSE+OAuth stack in production. The draft codifies patterns that companies with real security engineering teams already deploy.</p><p>The WIMSE identifiers specified in the draft bind agent identity to the execution environment through hardware-rooted attestation. A SPIRE agent on each node performs workload attestation by examining the kernel or querying the orchestration platform. Your agent&#8217;s identity gets measured from where it runs, not merely asserted by who registered it. An OAuth client_id is a registration artifact. A SPIFFE ID is cryptographic proof that Agent X is actually Agent X, running in the expected environment.</p><p>The draft also gets credentials right. Short-lived, cryptographically bound, explicit expiration. Static API keys are called out as unsuitable for agent authentication: bearer artifacts with no cryptographic binding, no identity conveyance, operationally painful to rotate.</p><p>That warning couldn&#8217;t come at a better time. Astrix Security analyzed over 5,200 open-source MCP server implementations and found that 53% rely on static API keys or Personal Access Tokens. Only 8.5% use OAuth. The ecosystem is building on exactly the anti-pattern the draft condemns.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!L6DS!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F80358181-31cd-42dd-ac08-ce32048aec9f_3748x2250.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!L6DS!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F80358181-31cd-42dd-ac08-ce32048aec9f_3748x2250.png 424w, https://substackcdn.com/image/fetch/$s_!L6DS!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F80358181-31cd-42dd-ac08-ce32048aec9f_3748x2250.png 848w, https://substackcdn.com/image/fetch/$s_!L6DS!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F80358181-31cd-42dd-ac08-ce32048aec9f_3748x2250.png 1272w, https://substackcdn.com/image/fetch/$s_!L6DS!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F80358181-31cd-42dd-ac08-ce32048aec9f_3748x2250.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!L6DS!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F80358181-31cd-42dd-ac08-ce32048aec9f_3748x2250.png" width="1456" height="874" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/80358181-31cd-42dd-ac08-ce32048aec9f_3748x2250.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:874,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:483418,&quot;alt&quot;:&quot;Pie chart showing 53% of MCP servers use static API keys versus 8.5% using OAuth&quot;,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.rockcybermusings.com/i/190013993?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F80358181-31cd-42dd-ac08-ce32048aec9f_3748x2250.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="Pie chart showing 53% of MCP servers use static API keys versus 8.5% using OAuth" title="Pie chart showing 53% of MCP servers use static API keys versus 8.5% using OAuth" srcset="https://substackcdn.com/image/fetch/$s_!L6DS!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F80358181-31cd-42dd-ac08-ce32048aec9f_3748x2250.png 424w, https://substackcdn.com/image/fetch/$s_!L6DS!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F80358181-31cd-42dd-ac08-ce32048aec9f_3748x2250.png 848w, https://substackcdn.com/image/fetch/$s_!L6DS!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F80358181-31cd-42dd-ac08-ce32048aec9f_3748x2250.png 1272w, https://substackcdn.com/image/fetch/$s_!L6DS!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F80358181-31cd-42dd-ac08-ce32048aec9f_3748x2250.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Figure 1: MCP Server Authentication Methods</figcaption></figure></div><h2>Transaction Tokens solve the lateral movement problem</h2><p>Section 10.4 addresses a real attack vector most frameworks ignore. When access tokens propagate through internal microservice chains within an agent workflow, every hop creates a theft and replay opportunity.</p><p>The draft&#8217;s answer is Transaction Tokens (draft-ietf-oauth-transaction-tokens-08). Short-lived, signed JWTs that bind user identity, workload identity, and authorization context to a specific transaction. Lifetimes are measured in seconds to minutes. Cryptographic signatures prevent context modification. You can&#8217;t grab a Transaction Token from one transaction and replay it in another because the transaction context is cryptographically sealed. A companion draft (draft-oauth-transaction-tokens-for-agents-04) extends this with agent-specific fields for the acting agent, the initiating human, and operational constraints.</p><p>The draft also correctly identifies tools forwarding access tokens to downstream services as an anti-pattern.</p><h2>The authorization gap: where scope alone isn&#8217;t enough</h2><p>Here&#8217;s where the draft&#8217;s -00 status shows. Once an OAuth access token gets issued with a set of scopes, every action within those scopes proceeds unchecked until the token expires. No per-action evaluation. No consequence assessment. No behavioral feedback loop. The authors clearly know authorization needs more work (the AIMS conceptual model describes layers that the spec hasn&#8217;t filled in yet), but anyone reading this draft as a deployment blueprint today will inherit that gap.</p><p>Think about what that means in practice. An agent with email:send scope authorized to send meeting notes can use that same scope to email every contact in the address book a different message. Each action is technically within scope. The framework treats them identically. The authorization decision happened once, at token issuance. Everything after that is a free pass.</p><p>OWASP&#8217;s Top 10 for Agentic Applications draws a distinction that the draft hasn&#8217;t addressed yet: <em><strong>least agency versus least privilege</strong></em>. Least privilege asks what the agent can access. Least agency extends that to how much freedom the agent has to act on that access without checking back.</p><p>The term &#8220;least agency&#8221; appears nowhere in the draft. Section 10.8 says agents should request minimum scopes and authorization details. That&#8217;s least privilege applied to OAuth scopes. Standard stuff. It does nothing to constrain autonomous decision-making within those scopes.</p><p>OWASP&#8217;s ASI03 (Identity and Privilege Abuse) mitigation guidance recommends per-action authorization through a centralized policy engine. Not once at token issuance. At each privileged step. The draft doesn&#8217;t provide a mechanism for this yet, and future revisions may address it. In the meantime, you need to build that layer yourself.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!hGET!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F29c69cc0-87ac-4b2d-b859-0c3c17b56f8a_2048x2048.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!hGET!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F29c69cc0-87ac-4b2d-b859-0c3c17b56f8a_2048x2048.jpeg 424w, https://substackcdn.com/image/fetch/$s_!hGET!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F29c69cc0-87ac-4b2d-b859-0c3c17b56f8a_2048x2048.jpeg 848w, https://substackcdn.com/image/fetch/$s_!hGET!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F29c69cc0-87ac-4b2d-b859-0c3c17b56f8a_2048x2048.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!hGET!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F29c69cc0-87ac-4b2d-b859-0c3c17b56f8a_2048x2048.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!hGET!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F29c69cc0-87ac-4b2d-b859-0c3c17b56f8a_2048x2048.jpeg" width="1456" height="1456" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/29c69cc0-87ac-4b2d-b859-0c3c17b56f8a_2048x2048.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1456,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:2429486,&quot;alt&quot;:&quot;Table showing IETF draft coverage levels against OWASP ASI01 through ASI10 risk categories&quot;,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.rockcybermusings.com/i/190013993?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F29c69cc0-87ac-4b2d-b859-0c3c17b56f8a_2048x2048.jpeg&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="Table showing IETF draft coverage levels against OWASP ASI01 through ASI10 risk categories" title="Table showing IETF draft coverage levels against OWASP ASI01 through ASI10 risk categories" srcset="https://substackcdn.com/image/fetch/$s_!hGET!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F29c69cc0-87ac-4b2d-b859-0c3c17b56f8a_2048x2048.jpeg 424w, https://substackcdn.com/image/fetch/$s_!hGET!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F29c69cc0-87ac-4b2d-b859-0c3c17b56f8a_2048x2048.jpeg 848w, https://substackcdn.com/image/fetch/$s_!hGET!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F29c69cc0-87ac-4b2d-b859-0c3c17b56f8a_2048x2048.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!hGET!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F29c69cc0-87ac-4b2d-b859-0c3c17b56f8a_2048x2048.jpeg 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Figure 2: OWASP Agentic Top 10 Coverage by IETF Draft</figcaption></figure></div><h2>Your token says &#8220;allowed.&#8221; What it can&#8217;t say is &#8220;should you?&#8221;</h2><p>The deeper issue goes beyond per-action evaluation. The draft in its current form contains no mechanisms for assessing the potential impact of an action before permitting it. No concept of blast radius. No reversibility check. No impact severity score. Again, this is version -00. These concepts may arrive in later revisions. They&#8217;re absent today.</p><p>Consider the practical difference. An agent with files:read_write scope can read one file or delete every file in scope. The OAuth framework treats these as equivalent actions. They aren&#8217;t. One is routine. The other is catastrophic and irreversible.</p><p>Consequence-based authorization asks three questions per permission:</p><ol><li><p>What&#8217;s the worst action this agent can take? </p></li><li><p>Is the damage reversible? </p></li><li><p>Can you reverse it within an acceptable recovery window? </p></li></ol><p>OAuth scopes can&#8217;t answer any of these.</p><p>The emerging practice of graduated trust models (read-only, then draft-only, then supervised execution, then earned autonomy) represents an informal consequence-based approach. Most practitioners agree that most agents never earn full autonomy in high-stakes contexts. That&#8217;s the correct outcome. The draft provides no framework for expressing or enforcing these graduation stages.</p><p>OWASP&#8217;s ASI08 (Cascading Failures) recommends blast-radius caps and digital twin replay testing. Run recorded agent actions in an isolated environment first. See if sequences trigger cascading failures before expanding policy permissions. Future revisions of the draft could incorporate these concepts. For now, they&#8217;re outside its scope.</p><h2>The observability gap: strong detection, no policy feedback loop</h2><p>Section 11&#8217;s observability requirements are genuinely strong for detection and audit. Seven minimum audit event fields. Correlation across agents, tools, services, and LLMs. The ability to reconstruct complete execution chains, including delegated authority and intermediate calls.</p><p>The draft calls observability &#8220;a security control, not solely an operational feature.&#8221; Correct. Then it integrates the OpenID Shared Signals Framework with CAEP (Continuous Access Evaluation Profile) for real-time signal delivery. Also good.</p><p>The problem is that the AIMS conceptual model in Section 4 promises observability that can &#8220;dynamically modify authorization decisions based on observed behavior and system state.&#8221; The actual specification delivers reactive remediation, terminate sessions, discard tokens, re-acquire with updated constraints. Detection flows to dashboards and SIEM tools. It doesn&#8217;t feed into the policy decision point that evaluates each authorization request. The conceptual model is ahead of the spec, which is normal for a -00 draft. The spec will likely catch up. You can&#8217;t afford to wait for it.</p><p>An agent exhibiting anomalous tool invocation patterns should see its authorization dynamically narrowed. Not through token revocation (which is all-or-nothing) but through policy-level constraints on permitted actions. The draft gives you a circuit breaker when you need a rheostat.</p><p>NIST SP 800-207 (Zero Trust Architecture) explicitly recommends a trust score that changes dynamically based on entity behavior patterns, feeding into the policy engine. Context-aware authorization systems from companies such as Zscaler and StrongDM already implement this pattern in production (not endorsing either). I&#8217;d expect future revisions of the draft to engage with these models, especially given that Zscaler&#8217;s Rosomakho is one of the four co-authors.</p><h2>AuthZEN fills the gap the draft hasn&#8217;t reached yet</h2><p>The most interesting omission in the current document is that AuthZEN (OpenID Authorization API 1.0) was approved as a Final Specification in January 2026. It standardizes a transport-agnostic API where any Policy Enforcement Point queries any Policy Decision Point, regardless of vendor. The information model is a four-element tuple: </p><p>Subject (the agent), Action (the operation), Resource (the target), Context (ambient attributes).</p><p>Every agent tool invocation maps cleanly to an AuthZEN evaluation: subject is the agent&#8217;s SPIFFE ID, action is &#8220;send_email,&#8221; resource is &#8220;contact_list,&#8221; context carries the delegating user, blast radius classification, reversibility flag, and behavioral anomaly score. The context object is extensible and open-ended. It was designed for exactly this kind of dynamic, attribute-rich decision-making.</p><p>The draft references AuthZEN in its normative references. The body text doesn&#8217;t discuss it yet. Given that AuthZEN solves the draft&#8217;s most significant open question, I&#8217;d bet it features prominently in the next revision. For now, that connection is yours to make.</p><p>Three policy engines deserve attention for filling that gap. OPA (Open Policy Agent), a CNCF Graduated project, evaluates structured JSON input against declarative policies with sub-millisecond latency. Cedar, from AWS, offers automated reasoning via SMT solver that can mathematically prove properties about policies and benchmarks at 42 to 60 times faster than Rego. Topaz, from Aserto (whose CEO co-authored the AuthZEN specification), combines OPA&#8217;s decision engine with a built-in Zanzibar-style relationship graph.</p><p>OAuth provides coarse-grained delegation, who can access what resource category. Policy engines provide fine-grained runtime evaluation, should this specific action on this specific resource proceed given current context. That layered model is where the draft needs to go next. Until it gets there, you build it yourself.</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!RHmI!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffb87f5e8-09ab-4e24-9acd-12fbc9d0790c_8192x664.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!RHmI!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffb87f5e8-09ab-4e24-9acd-12fbc9d0790c_8192x664.png 424w, https://substackcdn.com/image/fetch/$s_!RHmI!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffb87f5e8-09ab-4e24-9acd-12fbc9d0790c_8192x664.png 848w, https://substackcdn.com/image/fetch/$s_!RHmI!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffb87f5e8-09ab-4e24-9acd-12fbc9d0790c_8192x664.png 1272w, https://substackcdn.com/image/fetch/$s_!RHmI!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffb87f5e8-09ab-4e24-9acd-12fbc9d0790c_8192x664.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!RHmI!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffb87f5e8-09ab-4e24-9acd-12fbc9d0790c_8192x664.png" width="1456" height="118" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/fb87f5e8-09ab-4e24-9acd-12fbc9d0790c_8192x664.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:118,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:372395,&quot;alt&quot;:&quot;Diagram showing OAuth handling coarse-grained identity delegation while AuthZEN and policy engines handle per-action runtime evaluation&quot;,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.rockcybermusings.com/i/190013993?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffb87f5e8-09ab-4e24-9acd-12fbc9d0790c_8192x664.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="Diagram showing OAuth handling coarse-grained identity delegation while AuthZEN and policy engines handle per-action runtime evaluation" title="Diagram showing OAuth handling coarse-grained identity delegation while AuthZEN and policy engines handle per-action runtime evaluation" srcset="https://substackcdn.com/image/fetch/$s_!RHmI!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffb87f5e8-09ab-4e24-9acd-12fbc9d0790c_8192x664.png 424w, https://substackcdn.com/image/fetch/$s_!RHmI!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffb87f5e8-09ab-4e24-9acd-12fbc9d0790c_8192x664.png 848w, https://substackcdn.com/image/fetch/$s_!RHmI!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffb87f5e8-09ab-4e24-9acd-12fbc9d0790c_8192x664.png 1272w, https://substackcdn.com/image/fetch/$s_!RHmI!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffb87f5e8-09ab-4e24-9acd-12fbc9d0790c_8192x664.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a><figcaption class="image-caption">Figure 3: Authentication vs. Authorization Layer Responsibilities</figcaption></figure></div><h2>Regulatory timelines won&#8217;t wait for standards completion</h2><p>The EU AI Act&#8217;s high-risk system requirements take full effect August 2, 2026 (as of this writing, anyway). Five months from now. Article 14 requires human oversight. Article 26 requires deployers to keep automatically generated logs for at least six months. The draft&#8217;s identity-bound audit trails and CIBA-based human-in-the-loop mechanism directly support both.</p><p>NIST launched two converging initiatives in February 2026. The NCCoE concept paper on AI agent identity and authorization, and the AI Agent Standards Initiative covering security controls, identity, and testing. Both center on WIMSE/SPIFFE + OAuth. Both explicitly include policy-based access control, the piece the IETF draft&#8217;s -00 revision hasn&#8217;t specified yet.</p><p>The Colorado AI Act establishes a &#8220;reasonable care&#8221; standard for high-risk AI systems effective June 30, 2026. Widely adopted standards become evidence of reasonable care in court. The identity architecture the draft describes will likely qualify for authentication. You still need to build the authorization layer yourself.</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!wPEA!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faa31d4ce-233e-42ba-af0c-4c0ddac41e9d_7670x1700.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!wPEA!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faa31d4ce-233e-42ba-af0c-4c0ddac41e9d_7670x1700.png 424w, https://substackcdn.com/image/fetch/$s_!wPEA!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faa31d4ce-233e-42ba-af0c-4c0ddac41e9d_7670x1700.png 848w, https://substackcdn.com/image/fetch/$s_!wPEA!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faa31d4ce-233e-42ba-af0c-4c0ddac41e9d_7670x1700.png 1272w, https://substackcdn.com/image/fetch/$s_!wPEA!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faa31d4ce-233e-42ba-af0c-4c0ddac41e9d_7670x1700.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!wPEA!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faa31d4ce-233e-42ba-af0c-4c0ddac41e9d_7670x1700.png" width="1456" height="323" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/aa31d4ce-233e-42ba-af0c-4c0ddac41e9d_7670x1700.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:323,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:613784,&quot;alt&quot;:&quot;Timeline chart showing EU AI Act, Colorado AI Act, and NIST initiative deadlines converging in 2026&quot;,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.rockcybermusings.com/i/190013993?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faa31d4ce-233e-42ba-af0c-4c0ddac41e9d_7670x1700.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="Timeline chart showing EU AI Act, Colorado AI Act, and NIST initiative deadlines converging in 2026" title="Timeline chart showing EU AI Act, Colorado AI Act, and NIST initiative deadlines converging in 2026" srcset="https://substackcdn.com/image/fetch/$s_!wPEA!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faa31d4ce-233e-42ba-af0c-4c0ddac41e9d_7670x1700.png 424w, https://substackcdn.com/image/fetch/$s_!wPEA!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faa31d4ce-233e-42ba-af0c-4c0ddac41e9d_7670x1700.png 848w, https://substackcdn.com/image/fetch/$s_!wPEA!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faa31d4ce-233e-42ba-af0c-4c0ddac41e9d_7670x1700.png 1272w, https://substackcdn.com/image/fetch/$s_!wPEA!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faa31d4ce-233e-42ba-af0c-4c0ddac41e9d_7670x1700.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a><figcaption class="image-caption">Figure 4: Regulatory Compliance Timeline for AI Agent Systems</figcaption></figure></div><h2>MCP and A2A still have fundamental identity gaps</h2><p>Mapping the IETF draft&#8217;s framework onto the Model Context Protocol reveals how far the ecosystem still has to travel. MCP identifies agents as OAuth clients with a client_id, a registration artifact with no attestation binding. No SPIFFE identity verification. No attestation mechanism. No multi-hop delegation. No standard mapping between tool names and OAuth scopes. The draft recommends Workload Proof Tokens for proof-of-possession. MCP uses bearer tokens.</p><p>MCP&#8217;s OAuth model is human-centric (Authorization Code + PKCE). The Client Credentials Grant for machine-to-machine authentication was removed from the spec and is only returning through an extension. Fully autonomous agents have no standard authentication path in MCP today. Google&#8217;s A2A protocol has similar gaps: self-declared identities with no attestation binding, credential acquisition out of scope, authorization left to the receiving agent.</p><p>Riptides demonstrated the draft&#8217;s compositional pattern working for MCP in practice. Each workload gets a SPIFFE SVID, used as a software statement in Dynamic Client Registration and as a JWT assertion for client authentication. The pattern works. It required significant custom integration that no standard profile defines.</p><h2>What you should build now</h2><p>Don&#8217;t wait for standards completion. The threat model OWASP defined already exists. The regulatory deadlines are set.</p><p>Start with SPIFFE/SPIRE for attestation-bound agent identity. Use SVIDs as JWT assertions (RFC 7523) to obtain OAuth tokens. This follows the pattern the draft describes and Riptides validated in production.</p><p>Deploy an AuthZEN-compliant PDP (OPA, Cedar, or Topaz). Evaluate every agent tool invocation against dynamic policy. Pass agent identity, action details, resource metadata, delegation context, and behavioral signals in the AuthZEN context object.</p><p>Write Cedar or Rego policies encoding blast-radius thresholds, reversibility requirements, graduated trust levels, and human-in-the-loop triggers. Version-control policies alongside application code.</p><p>Tag every tool and action with impact metadata: blast_radius, reversible, data_sensitivity, scope. Enforce that irreversible high-blast-radius actions require explicit human approval through CIBA step-up authorization.</p><p>Feed observability data into the policy engine as real-time context attributes. Stop sending behavioral signals only to SIEM dashboards for post-hoc investigation. Make them first-class policy inputs.</p><p><strong>Key Takeaway:</strong> The IETF draft gives you a strong answer to &#8220;is this really Agent X?&#8221; It hasn&#8217;t answered &#8220;should Agent X do this specific thing right now?&#8221; yet. That gap will close as the draft matures. In the meantime, authentication without per-action authorization is a locked front door with open windows. Build the authorization layer now.</p><h3>What to do next</h3><p>If you&#8217;re building agentic systems and trying to figure out where identity controls fit, start with the CARE framework at <a href="https://rockcyber.com">rockcyber.com</a> for mapping security controls to business risk outcomes. The RISE framework helps you evaluate where your organization sits on the AI security maturity curve, particularly useful for figuring out which authorization controls to prioritize first.</p><p>The agent identity problem is a microcosm of the larger question the book addresses: how do you govern autonomous systems when the blast radius of failure compounds faster than your ability to detect it?</p><p>More analysis on agentic AI security, MCP authorization gaps, and practical frameworks for building authorization layers at <a href="https://rockcybermusings.com">rockcybermusings.com</a>.</p><p>&#128073; Subscribe for more AI security and governance insights with the occasional rant.</p><p>&#128073; Visit <strong><a href="https://www.rockcyber.com/">RockCyber.com</a></strong> to learn more about how we can help you in your traditional Cybersecurity and AI Security and Governance Journey</p><p>&#128073; Want to save a quick $100K? Check out our AI Governance Tools at <strong><a href="https://aigovernancetoolkit.com/">AIGovernanceToolkit.com</a></strong></p><p><em>The views and opinions expressed in RockCyber Musings are my own and do not represent the positions of my employer or any organization I&#8217;m affiliated with.</em></p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.rockcybermusings.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading RockCyber Musings! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.rockcybermusings.com/?utm_source=substack&amp;utm_medium=email&amp;utm_content=share&amp;action=share&quot;,&quot;text&quot;:&quot;Share RockCyber Musings&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.rockcybermusings.com/?utm_source=substack&amp;utm_medium=email&amp;utm_content=share&amp;action=share"><span>Share RockCyber Musings</span></a></p><p></p>]]></content:encoded></item><item><title><![CDATA[Weekly Musings Top 10 AI Security Wrapup: Issue 29 March 6, 2026 - March 12, 2026]]></title><description><![CDATA[When AI Companies Sue the Government and OpenAI Enters the Security Market]]></description><link>https://www.rockcybermusings.com/p/weekly-musings-top-10-ai-security-20260306-202600312</link><guid isPermaLink="false">https://www.rockcybermusings.com/p/weekly-musings-top-10-ai-security-20260306-202600312</guid><dc:creator><![CDATA[Rock Lambros]]></dc:creator><pubDate>Fri, 13 Mar 2026 12:50:41 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!hJU0!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6664727a-c9fd-4acb-b74f-259d770fda92_1024x1024.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!hJU0!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6664727a-c9fd-4acb-b74f-259d770fda92_1024x1024.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!hJU0!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6664727a-c9fd-4acb-b74f-259d770fda92_1024x1024.png 424w, https://substackcdn.com/image/fetch/$s_!hJU0!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6664727a-c9fd-4acb-b74f-259d770fda92_1024x1024.png 848w, https://substackcdn.com/image/fetch/$s_!hJU0!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6664727a-c9fd-4acb-b74f-259d770fda92_1024x1024.png 1272w, https://substackcdn.com/image/fetch/$s_!hJU0!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6664727a-c9fd-4acb-b74f-259d770fda92_1024x1024.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!hJU0!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6664727a-c9fd-4acb-b74f-259d770fda92_1024x1024.png" width="1024" height="1024" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/6664727a-c9fd-4acb-b74f-259d770fda92_1024x1024.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1024,&quot;width&quot;:1024,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:1233556,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.rockcybermusings.com/i/190823556?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6664727a-c9fd-4acb-b74f-259d770fda92_1024x1024.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!hJU0!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6664727a-c9fd-4acb-b74f-259d770fda92_1024x1024.png 424w, https://substackcdn.com/image/fetch/$s_!hJU0!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6664727a-c9fd-4acb-b74f-259d770fda92_1024x1024.png 848w, https://substackcdn.com/image/fetch/$s_!hJU0!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6664727a-c9fd-4acb-b74f-259d770fda92_1024x1024.png 1272w, https://substackcdn.com/image/fetch/$s_!hJU0!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6664727a-c9fd-4acb-b74f-259d770fda92_1024x1024.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>The week of March 6-12, 2026, handed us a story that was coming... Anthropic filed suit against the Pentagon for blacklisting it as a national security risk. In the same week, the White House released a new cyber strategy, OpenAI launched a vulnerability-scanning agent aimed squarely at the enterprise security market, and two major federal regulatory deadlines expired. This is that week.</p><p>AI Security and AI governance collided this week in federal court, in congressional briefings, and in the server rooms of every organization running an AI agent they don&#8217;t fully understand. The governance frameworks that were supposed to provide clarity are instead amplifying uncertainty, and attackers are exploiting the gap in real time. Here&#8217;s what happened, what it means, and what to do about it, from someone who&#8217;s watched this industry long enough to be appropriately paranoid about all of it.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.rockcybermusings.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading RockCyber Musings! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.rockcybermusings.com/p/weekly-musings-top-10-ai-security-20260306-202600312?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.rockcybermusings.com/p/weekly-musings-top-10-ai-security-20260306-202600312?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p><h3>1. Anthropic sues the Pentagon for blacklisting it as a national security risk</h3><p>Anthropic filed two federal lawsuits against the Trump administration after the Department of Defense designated the company a supply chain risk. That designation, typically reserved for foreign adversaries, bars Anthropic from federal contracts and requires defense contractors to certify they don&#8217;t use Claude in any DoD work. The root cause is Anthropic&#8217;s refusal to allow Claude for autonomous weapons or mass surveillance of American citizens. CEO Dario Amodei drew two red lines in contract negotiations, the Pentagon walked, and then labeled the company a national security threat (Fortune, Defense One). Anthropic warns the financial exposure runs to hundreds of millions of dollars.</p><p><strong>Why it matters</strong></p><ul><li><p>This is the first time a U.S.-headquartered AI company has received the supply chain risk designation, a label previously applied only to foreign adversaries.</p></li><li><p>The case tests whether the executive branch can use procurement leverage to override AI developers&#8217; safety commitments, a precedent that extends far beyond Anthropic.</p></li><li><p>Every CISO advising on AI vendor selection now has to factor whether a vendor&#8217;s ethics commitments make it a federal liability.</p></li></ul><p><strong>What to do about it</strong></p><ul><li><p>Map your Claude and Anthropic API dependencies now. Know which workflows break if this escalates.</p></li><li><p>Brief your board on what a supply chain risk designation means in federal contracting terms if your organization touches government work.</p></li><li><p>Watch for similar scrutiny applied to other AI vendors with published safety policies. This may not be a one-off.</p></li></ul><p><strong>Rock&#8217;s Musings</strong></p><p>Anthropic drew a line in the sand (no autonomous weapons, no mass surveillance), and the government responded by calling them a threat. Think about what that signals to every AI developer watching. If you have safety principles that conflict with defense procurement, you get punished for them. The First Amendment angle is interesting, but the real issue is that the executive branch just discovered that supply chain risk designation is a very effective stick, and they used it on a domestic company for the first time. AI safety as a business value just became a liability under the current administration. Read that sentence twice.</p><h3>2. Trump&#8217;s Cyber Strategy for America lands in five pages</h3><p>On March 6, the White House released &#8220;President Trump&#8217;s Cyber Strategy for America&#8221; alongside an executive order on cybercrime (White House, Forrester). The document covers six pillars: offensive cyber operations to shape adversary behavior, regulatory streamlining, federal network modernization, critical infrastructure security, technological superiority, and cyber workforce development. At five pages, it&#8217;s the shortest national cybersecurity strategy in a decade. The strategy explicitly calls for more aggressive offensive operations, &#8220;unprecedented coordination&#8221; between the public and private sectors, and the building of a talent base fluent in autonomous systems and AI-enabled defense.</p><p><strong>Why it matters</strong></p><ul><li><p>Five pages are either a vision document or a placeholder. For practical CISO purposes, it signals direction but provides almost no implementation guidance.</p></li><li><p>The offensive posture language has legal and escalation implications for any organization with a government nexus.</p></li><li><p>Workforce development framed as a national strategic asset means the government will be competing for the same AI security talent you&#8217;re trying to hire.</p></li></ul><p><strong>What to do about it</strong></p><ul><li><p>Map existing compliance obligations against the six pillars. Where regulations get streamlined, understand which requirements might disappear and which you need to maintain voluntarily.</p></li><li><p>Engage your federal liaison if you&#8217;re in a critical infrastructure sector. The public-private coordination language means more government asks are coming.</p></li><li><p>Start building for AI-fluent security talent now. The window before this becomes a serious hiring crunch is closing.</p></li></ul><p><strong>Rock&#8217;s Musings</strong></p><p>Five pages tells you something: either there&#8217;s a lot more in the classified annex, or this is aspirational language waiting for someone to actually build the plumbing. The workforce section is the sleeper story. AI-enabled defense needs people who understand both AI failure modes and adversarial tradecraft simultaneously. That combination doesn&#8217;t exist at scale anywhere, and we&#8217;re being asked to build it at the same time AI is accelerating attacks. The gap between those two curves is where the next major breach lives.</p><h3>3. OpenAI launches Codex Security and walks into the vulnerability scanning market</h3><p>OpenAI released Codex Security as a research preview, a context-aware AI vulnerability scanning agent that evolved from Aardvark, an internal security research tool OpenAI had tested in private beta since October 2025 (Bloomberg, SecurityWeek). Codex Security analyzes code repositories, pressure-tests suspected vulnerabilities in sandboxed environments, generates proof-of-concept exploits to confirm impact, and proposes fixes. OpenAI&#8217;s own data shows it scanned 1.2 million commits over the preceding 30 days, surfacing 10,561 high-severity issues and approximately 800 critical vulnerabilities. The tool is available free for the next month to ChatGPT Pro, Enterprise, Business, and Edu customers. OpenAI says it can &#8220;identify complex vulnerabilities that other agentic tools miss&#8221; (TechRadar).</p><p><strong>Why it matters</strong></p><ul><li><p>A free, frontier-model-powered vulnerability scanner from OpenAI immediately changes the competitive math for established AppSec vendors whose pricing models depend on the difficulty of this problem.</p></li><li><p>Generating proof-of-concept exploits to confirm vulnerability impact is a significant capability. In the wrong hands, or with a compromised account, this is an exploit generation service.</p></li><li><p>Organizations deploying Codex Security are giving OpenAI&#8217;s systems read access to their codebases. That data handling relationship deserves the same scrutiny as any privileged third-party tool.</p></li></ul><p><strong>What to do about it</strong></p><ul><li><p>Before enabling Codex Security on production repositories, review OpenAI&#8217;s data retention and training policies. Understand whether your code becomes training data.</p></li><li><p>Evaluate Codex Security against your existing SAST tooling on a representative code sample before replacing anything. &#8220;Better than other agentic tools&#8221; is a marketing claim until your team validates it.</p></li><li><p>The proof-of-concept exploit generation feature needs access controls. Restrict which engineers can trigger full exploit confirmation scans.</p></li></ul><p><strong>Rock&#8217;s Musings</strong></p><p>OpenAI entering the vulnerability scanner market is not a product launch. It&#8217;s a statement about where AI is heading in security operations. The incumbents in SAST and DAST have been selling the same scan-and-report workflow for a decade. An agent that generates a proof-of-concept exploit to confirm a real finding changes the value proposition significantly. I&#8217;m not surprised OpenAI built this. I&#8217;m watching carefully how they handle the fact that generating exploit code is exactly the capability defenders need and attackers want. The account compromise scenario alone should give your red team ideas.</p><h3>4. NIST AI Agent Standards RFI closes with 932 comments</h3><p>The comment period for NIST&#8217;s Center for AI Standards and Innovation (CAISI) Request for Information on securing AI agent systems closed March 9 with 932 responses (Federal Register, NIST). The RFI, published in January 2026, sought input from industry, academia, and the security community on securing agentic AI development and deployment. The OpenID Foundation submitted a response addressing AI agent identity and authorization. A second comment period focused specifically on identity and authorization for AI agents remains open until April 2.</p><p><strong>Why it matters</strong></p><ul><li><p>932 responses signals broad industry recognition of the problem. The quality of those comments determines whether the resulting standards have operational teeth.</p></li><li><p>Identity and authorization for AI agents is the structural gap behind most agent security failures. If NIST gets this right, it reshapes the risk calculus for enterprise agent deployment.</p></li><li><p>The listening sessions starting in April give practitioners a direct channel to shape what these standards require.</p></li></ul><p><strong>What to do about it</strong></p><ul><li><p>If your organization skipped the first RFI, submit to the identity and authorization comment period before April 2. Your implementation experience is exactly what NIST needs.</p></li><li><p>Start building your AI agent identity architecture now using OAuth 2.0 On-Behalf-Of flows with proper scope constraints. This is the emerging standard pattern.</p></li><li><p>Assign someone to track the AI Agent Standards Initiative. When draft standards publish later this year, you want your red-team comments in front of NIST before they finalize.</p></li></ul><p><strong>Rock&#8217;s Musings</strong></p><p>Standards processes are slow by design, and the slowness here is appropriate because the identity and authorization problem for AI agents is genuinely hard. An agent acting on behalf of a user needs to carry that user&#8217;s permissions, not escalate to system-level access, and current tooling doesn&#8217;t enforce this reliably. The OIDF response to NIST gets the framing right: agent identity needs cryptographic binding, not just policy. If an agent claims to act on your behalf without a verifiable credential, you don&#8217;t have identity management. You have a trust-me system. You can read about the comments I submitted at <a href="https://www.rockcybermusings.com/p/nist-ai-agent-rfi-2025-0035-human-oversight-wrong-fix">&#8220;NIST AI Agent RFI (2025-0035): Human Oversight Is the Wrong Fix.&#8221;</a></p><h3>5. Commerce and FTC hit their AI regulatory deadlines, and nothing changed yet</h3><p>Two major deliverables from the December 2025 executive order on AI preemption came due on March 11. The Commerce Department submitted its review of state AI laws, identifying which ones the administration considers overly burdensome or in conflict with federal objectives. The FTC delivered a policy statement on how Section 5 of the FTC Act applies to AI and when state laws requiring alteration of model outputs are preempted by federal deceptive practices law (Mondaq, Digital Applied). Neither document invalidates any state law on its own. They are ammunition for the DOJ&#8217;s AI Litigation Task Force, established in January and yet to file any lawsuits. The administration is also conditioning $42 billion in BEAD broadband funding on states repealing AI regulations it deems onerous.</p><p><strong>Why it matters</strong></p><ul><li><p>Organizations operating AI in multiple states face genuine legal uncertainty. State laws remain on the books. The federal government plans to fight them in court, and that litigation takes years.</p></li><li><p>The FTC&#8217;s Section 5 application to AI bias-mitigation requirements is legally untested territory.</p></li><li><p>The BEAD funding leverage is the most concrete near-term enforcement tool. Which states hold firm versus which fold will tell you a lot about regulatory durability.</p></li></ul><p><strong>What to do about it</strong></p><ul><li><p>Do not assume any state AI compliance requirement is going away. Build compliance architecture that can be toggled by jurisdiction as the legal landscape shifts.</p></li><li><p>Get legal counsel read into the Commerce Department report. Knowing which of your state compliance obligations are on the federal target list helps you prioritize risk posture.</p></li><li><p>Prepare for a two-to-three year period of overlapping requirements. Companies with modular, jurisdiction-aware compliance programs will weather this better.</p></li></ul><p><strong>Rock&#8217;s Musings</strong></p><p>The administration created a fog of legal uncertainty and called it reducing regulatory burden. For most enterprises deploying AI, this makes compliance harder. You now have to track active federal litigation against state laws while still complying with those laws until courts rule otherwise. The FTC theory is worth watching closely: if the argument that requiring AI bias mitigation compels &#8220;deceptive output&#8221; holds, it guts a large category of state AI fairness requirements. If it fails, it sets a precedent limiting federal deceptive practices law&#8217;s reach into AI output governance. Either outcome reshapes the field.</p><div><hr></div><h3>6. OpenAI publishes its prompt injection defense playbook</h3><p>On March 12, OpenAI published research and engineering guidance on defending AI agents against prompt-injection attacks (OpenAI, PrismNews). The guidance covers training techniques that help models treat different input channels with varying skepticism, architectural decisions that constrain privilege and limit blast radius, and layered verification to catch anomalous behavior. OpenAI also disclosed that it built a reinforcement learning-trained automated attacker to discover injection vulnerabilities internally, capable of steering agents through harmful multi-step workflows. The decision to publish openly reflects recognition that injection attacks threaten the entire developer ecosystem building on top of large language models.</p><p><strong>Why it matters</strong></p><ul><li><p>Publishing the automated attacker methodology gives defenders a concrete model of what they&#8217;re fighting. Multi-step RL-trained attacks won&#8217;t be stopped with static guardrails.</p></li><li><p>The channel-skepticism approach, which trains models to treat external web content differently from system instructions, is an architectural fix that operates at inference time.</p></li><li><p>OpenAI&#8217;s disclosure accelerates industry defenses while giving attackers a clearer picture of which countermeasures to route around.</p></li></ul><p><strong>What to do about it</strong></p><ul><li><p>Apply privilege minimization immediately: agents should hold only permissions required for the specific task, expiring at task completion.</p></li><li><p>For agents consuming external content, validate that content before the agent ingests it. Treat external web data as untrusted input, period.</p></li><li><p>Build a prompt injection test suite and run it against production agents before every deployment. What you don&#8217;t test, you don&#8217;t know.</p></li></ul><p><strong>Rock&#8217;s Musings</strong></p><p>OpenAI built an RL-trained machine to find injection vulnerabilities in their own systems. That machine now exists, and the same architecture will run on the offensive side of this problem within months, if it isn&#8217;t already. The deeper issue is architectural: language models cannot reliably distinguish instructions from data. That&#8217;s a fundamental property of how these systems process text, not a fixable bug. Any defense assuming the model will eventually learn to make that distinction is building on sand. The real fix is external. Don&#8217;t give agents access to resources they don&#8217;t need, and verify every external input before it reaches the model.</p><h3>7. Google Cloud Threat Horizons reveals software exploits overtaking stolen credentials</h3><p>Google Cloud&#8217;s Office of the CISO published its H1 2026 Threat Horizons Report on March 9, covering the second half of 2025 (Help Net Security, Security Boulevard). The headline finding is that exploitation of third-party software vulnerabilities jumped from 2.9% to 44.5% of initial cloud entry vectors in a single half-year period. The exploitation window has collapsed to days, with the React2Shell case showing crypto miners deployed within 48 hours of public vulnerability disclosure. North Korean threat group UNC4899 abused DevOps workflows and container breakout to steal millions in cryptocurrency. Threat actors also used LLMs to automate credential harvesting and accelerate the path from local developer access to full cloud admin privileges.</p><p><strong>Why it matters</strong></p><ul><li><p>A jump from 2.9% to 44.5% in software exploitation isn&#8217;t an incremental change. Something shifted structurally in attacker methodology during H2 2025.</p></li><li><p>A 48-hour exploitation window means patch prioritization SLAs have to account for attacker speed, not just team capacity.</p></li><li><p>LLM-assisted credential harvesting is now in a major incident response dataset, no longer just theoretical research.</p></li></ul><p><strong>What to do about it</strong></p><ul><li><p>Reduce your vulnerability exposure window to 48 hours or less for critical and high-severity findings on internet-facing systems. Build the automation to get there.</p></li><li><p>Audit DevOps pipeline permissions. The UNC4899 vector targets the privilege elevation that happens when developers hold broad cloud access from local workstations.</p></li><li><p>Review whether AI coding tools introduce dependencies with unreviewed third-party code. Supply chain hygiene is now tier-one.</p></li></ul><p><strong>Rock&#8217;s Musings</strong></p><p>For years, the orthodoxy was &#8220;credential hygiene is job one in the cloud.&#8221; Attackers just told you that orthodoxy is obsolete. They shifted to software exploitation because credential defenses got good enough. That&#8217;s how this works: defenders get strong on one vector, attackers rotate to the next. The current answer is patching speed. The LLM-assisted credential harvesting detail is quietly significant. It&#8217;s been in theoretical papers for two years, and now it&#8217;s in operational incident data from nation-state actors. Adjust your threat model accordingly.</p><h3>8. AI agents are now helping criminals manage attack infrastructure</h3><p>On March 8, The Register reported on Microsoft Threat Intelligence findings showing that North Korea&#8217;s Coral Sleet group is using AI and development platforms to rapidly build and manage attack infrastructure at scale. AI agents automate the creation of phishing infrastructure, manage C2 systems, and accelerate campaign tempo. The Unit 42 2026 Global Incident Response Report, published in February and drawing on 750 major incidents, showed the fastest 25% of attackers reaching data exfiltration in 72 minutes, down from 285 minutes the previous year. Identity weaknesses played a material role in almost 90% of investigations.</p><p><strong>Why it matters</strong></p><ul><li><p>AI is now a documented operational capability in nation-state attack campaigns, not just an enterprise productivity tool.</p></li><li><p>The 4x speed increase in attack timelines means detection and response programs calibrated to last year&#8217;s data are already outdated.</p></li><li><p>87% of incidents unfolded across multiple attack surfaces, making correlation harder for defenders.</p></li></ul><p><strong>What to do about it</strong></p><ul><li><p>Review detection and response SLAs against the new attacker timeline. 72 minutes from initial access to exfiltration is shorter than most IR playbook trigger times.</p></li><li><p>Run tabletops assuming an AI-assisted attack infrastructure. Stress-test whether your team can detect and contain within the compressed timeline.</p></li><li><p>Identity controls remain the highest-leverage investment. 90% material involvement in incidents makes this your budget priority.</p></li></ul><p><strong>Rock&#8217;s Musings</strong></p><p>The debate about whether attackers would use AI is over. It&#8217;s all about the economics. If you&#8217;re running persistent operations against multiple targets, automating the operational overhead with AI is exactly what you&#8217;d do. The 72-minute exfiltration timeline is the number that should break your IR program&#8217;s assumptions. Most enterprise programs are built around detection metrics measured in hours or days. You need automated detection with automated response triggers, not a playbook that assumes a human analyst will catch the initial alert.</p><h3>9. Amazon pushes back on data linking AI coding to infrastructure outages</h3><p>On March 10, The Register reported leaked briefing notes from an Amazon internal operations meeting flagging a &#8220;trend of incidents&#8221; characterized by &#8220;high blast radius&#8221; and &#8220;Gen-AI assisted changes.&#8221; The implication was that AI-assisted coding has made infrastructure changes more fragile. Amazon responded, saying they &#8220;have not seen compelling evidence that incidents are more common with AI tools.&#8221; The Veracode 2026 State of Software Security report, published February 24, found 82% of organizations carry security debt, a 36% year-over-year spike in high-risk vulnerabilities, and that more vulnerabilities are being created than fixed, with AI development velocity outstripping remediation capacity as a contributing factor.</p><p><strong>Why it matters</strong></p><ul><li><p>Amazon&#8217;s internal concern, even disputed, comes from one of the largest cloud operators in the world. Internal friction at that scale is a signal worth tracking.</p></li><li><p>The Veracode data shows a systemic pattern. AI tools accelerate feature shipping and the introduction of vulnerabilities simultaneously, while remediation capacity doesn&#8217;t scale at the same rate.</p></li><li><p>82% of organizations carry security debt, with 60% classified as critical, which <em>should</em> be a material risk disclosure issue for most boards (materiality is another conversation for another time).</p></li></ul><p><strong>What to do about it</strong></p><ul><li><p>Require AI coding tools to integrate with static analysis before code reaches production. Velocity gains without security gates just accelerate debt accumulation.</p></li><li><p>Measure remediation rate alongside development velocity. If the gap is widening, you have a governance problem, not just a tooling problem.</p></li><li><p>Brief your board on the Veracode numbers. This is a material risk disclosure issue.</p></li></ul><p><strong>Rock&#8217;s Musings</strong></p><p>Amazon&#8217;s denial matters. One leaked briefing note does not make a causal case. What it tells you is that someone inside one of the world&#8217;s largest cloud operators thought the correlation was worth flagging in an internal ops review. That&#8217;s a signal, not proof. The Veracode data is where I&#8217;m more confident: if your AI coding tools help developers write code 40% faster and that code contains the same flaw density as human-written code, you&#8217;ve just increased your vulnerability production rate by 40%. The only way this works in your favor is if you accelerate the remediation side at the same rate. Almost nobody is doing that.</p><h3>10. Microsoft Patch Tuesday drops 77 CVEs</h3><p>Microsoft pushed its March Patch Tuesday on March 11, fixing at least 77 vulnerabilities across Windows and other software (Kaseya, Check Point Research). This update cycle lands in an environment where, per the Google Cloud Threat Horizons data released the same week, exploitation windows for critical vulnerabilities have collapsed to 48 hours from public disclosure. AI-assisted exploit development is further compressing the time between CVE publication and the availability of weaponized exploits.</p><p><strong>Why it matters</strong></p><ul><li><p>77 CVEs in one month means your patch management team works against a sprint clock every Patch Tuesday. Prioritization methodology matters more than ever.</p></li><li><p>Critical Microsoft CVEs are being probed within 48 hours of this disclosure per current attacker timelines. Your patch SLA has to account for that.</p></li><li><p>AI-assisted exploit development means the gap between disclosure and exploitation continues to narrow.</p></li></ul><p><strong>What to do about it</strong></p><ul><li><p>Build risk-tiered patching protocols: critical internet-facing systems within 24-48 hours, critical internal systems within 72 hours, high severity within a week.</p></li><li><p>Prioritize remote code execution vulnerabilities from the March 11 batch first. Review the Microsoft advisory for specific critical CVEs.</p></li><li><p>Apply compensating controls like network segmentation and least-privilege configurations for systems where immediate patching isn&#8217;t operationally feasible.</p></li></ul><p><strong>Rock&#8217;s Musings</strong></p><p>Patch Tuesday used to feel routine. It isn&#8217;t anymore because the time between a CVE being added to the NVD and an attacker scanning for it has gone from weeks to hours. If your patch SLA is still &#8220;30 days for critical,&#8221; you&#8217;re operating with a policy written for a threat environment that no longer exists. That&#8217;s not a patch management problem. That&#8217;s a governance problem. Fix the policy first.</p><h3>The One Thing You Won&#8217;t Hear About But You Need To</h3><h3>CISA adds an actively exploited n8n RCE to its known exploited list, and 24,700 instances are still unpatched</h3><p>On March 12, CISA added CVE-2025-68613 to its Known Exploited Vulnerabilities catalog, a critical expression-injection vulnerability in the n8n workflow automation platform with a CVSS score of 9.9 (The Hacker News, The Register). The flaw was patched three months ago in the December 2025 versions. Federal agencies have until March 25 to patch. The problem: Shadowserver data shows 24,700 instances remain unpatched online, with 12,300 in North America and 7,800 in Europe. This matters beyond the CVE itself because n8n is one of the most widely used platforms for building AI automation workflows and AI agent pipelines. Organizations deploying AI agents frequently use n8n as the orchestration layer connecting those agents to enterprise data sources.</p><p><strong>Why it matters</strong></p><ul><li><p>An unpatched RCE in the orchestration layer of an AI workflow means that an attacker who owns the n8n instance can access every connected system the AI agents touch, including credentials, APIs, and data stores.</p></li><li><p>24,700 exposed instances three months after a publicly known critical patch represents a systemic patching failure in a category of software organizations that have not been treated as critical infrastructure.</p></li><li><p>CISA&#8217;s KEV addition triggers mandatory remediation timelines for federal agencies, but most n8n deployments are in private enterprise environments with no equivalent enforcement mechanism.</p></li></ul><p><strong>What to do about it</strong></p><ul><li><p>Search your environment for n8n now. It is frequently deployed by individual teams or developers outside formal IT procurement, so your asset inventory may not show it.</p></li><li><p>If you find unpatched instances, treat them as compromised until proven otherwise. Rotate every credential and API key the n8n instance had access to.</p></li><li><p>Apply the same logic to every workflow automation tool in your environment: Zapier, Make, and similar platforms are potential RCE targets and connect to the same sensitive data sources.</p></li></ul><p><strong>Rock&#8217;s Musings</strong></p><p>This story isn&#8217;t getting the attention it deserves because nobody considers workflow automation as critical security infrastructure. It&#8217;s where developers wire things together quickly, connect AI agents to Slack, Salesforce, and internal APIs, and then move on to the next problem. The security team doesn&#8217;t own it. The AI team doesn&#8217;t think they need to patch it. The result is a critical RCE sitting at the center of your AI agent architecture, exposed to the internet, with a patch that&#8217;s been available for three months. CISA flagging active exploitation on March 12 means this is not theoretical. Someone is using this right now. Go find your n8n instances.</p><p>If you found this analysis useful, subscribe at <a href="https://rockcybermusings.com/">rockcybermusings.com</a> for weekly intelligence on AI security developments.</p><p>&#128073; Visit <strong><a href="https://www.rockcyber.com/">RockCyber.com</a></strong> to learn more about how we can help you in your traditional Cybersecurity and AI Security and Governance Journey</p><p>&#128073; Want to save a quick $100K? Check out our AI Governance Tools at <strong><a href="https://aigovernancetoolkit.com/">AIGovernanceToolkit.com</a></strong></p><p>&#128073; Subscribe for more AI and cyber insights with the occasional rant.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.rockcybermusings.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading RockCyber Musings! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.rockcybermusings.com/?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share RockCyber Musings&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.rockcybermusings.com/?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share RockCyber Musings</span></a></p><h2>References</h2><p>Axios. (2026, March 6). <em>OpenAI rolls out Codex Security to automate code security reviews</em>. <a href="https://www.axios.com/2026/03/06/openai-codex-security-ai-cyber">https://www.axios.com/2026/03/06/openai-codex-security-ai-cyber</a></p><p>Baker Botts. (2026, March). <em>March 2026: Federal deadlines that will reshape the AI regulatory landscape</em>. MONDAQ. <a href="https://www.mondaq.com/unitedstates/new-technology/1755166/march-2026-federal-deadlines-that-will-reshape-the-ai-regulatory-landscape">https://www.mondaq.com/unitedstates/new-technology/1755166/march-2026-federal-deadlines-that-will-reshape-the-ai-regulatory-landscape</a></p><p>Bloomberg. (2026, March 6). <em>OpenAI unveils Codex Security tool to detect database vulnerabilities</em>. <a href="https://www.bloomberg.com/news/articles/2026-03-06/openai-releases-ai-agent-security-tool-for-research-preview">https://www.bloomberg.com/news/articles/2026-03-06/openai-releases-ai-agent-security-tool-for-research-preview</a></p><p>Check Point Research. (2026, March 9). <em>9th March: Threat Intelligence Report</em>. <a href="https://research.checkpoint.com/2026/9th-march-threat-intelligence-report/">https://research.checkpoint.com/2026/9th-march-threat-intelligence-report/</a></p><p>CISA. (2026, March 12). <em>CISA adds one known exploited vulnerability to catalog</em>. <a href="https://www.cisa.gov/known-exploited-vulnerabilities-catalog">https://www.cisa.gov/known-exploited-vulnerabilities-catalog</a></p><p>CNBC. (2026, March 10). <em>Amazon convenes &#8216;deep dive&#8217; internal meeting to address outages</em>. <a href="https://www.cnbc.com/2026/03/10/amazon-plans-deep-dive-internal-meeting-address-ai-related-outages.html">https://www.cnbc.com/2026/03/10/amazon-plans-deep-dive-internal-meeting-address-ai-related-outages.html</a></p><p>Defense One. (2026, March 9). <em>Anthropic sues over a dozen federal agencies and government leaders</em>. <a href="https://www.defenseone.com/business/2026/03/anthropic-sues-over-dozen-federal-agencies-and-government-leaders/411997/">https://www.defenseone.com/business/2026/03/anthropic-sues-over-dozen-federal-agencies-and-government-leaders/411997/</a></p><p>Digital Applied. (2026, March). <em>FTC AI policy deadline March 11: Compliance guide</em>. <a href="https://www.digitalapplied.com/blog/ftc-ai-policy-deadline-march-11-compliance-readiness">https://www.digitalapplied.com/blog/ftc-ai-policy-deadline-march-11-compliance-readiness</a></p><p>Forrester. (2026, March). <em>White House announces the 2026 cyber strategy for America</em>. <a href="https://www.forrester.com/blogs/white-house-announces-the-2026-cyber-strategy-for-america/">https://www.forrester.com/blogs/white-house-announces-the-2026-cyber-strategy-for-america/</a></p><p>Fortune. (2026, March 9). <em>Anthropic sues Pentagon after being labeled a threat to national security</em>. <a href="https://fortune.com/2026/03/09/anthropic-sues-pentagon-ai-supply-chain-risk-trump-administration/">https://fortune.com/2026/03/09/anthropic-sues-pentagon-ai-supply-chain-risk-trump-administration/</a></p><p>Google Cloud. (2026, March 9). <em>Cloud threat horizons report H1 2026</em>. <a href="https://cloud.google.com/security/report/resources/cloud-threat-horizons-report-h1-2026">https://cloud.google.com/security/report/resources/cloud-threat-horizons-report-h1-2026</a></p><p>Help Net Security. (2026, March 11). <em>Software vulnerabilities push credential abuse aside in cloud intrusions</em>. <a href="https://www.helpnetsecurity.com/2026/03/11/google-cloud-environments-cyber-threats-report/">https://www.helpnetsecurity.com/2026/03/11/google-cloud-environments-cyber-threats-report/</a></p><p>Kaseya. (2026, March 11). <em>The week in breach news: March 11, 2026</em>. </p><p>https://www.kaseya.com/?post_type=post&amp;p=26754</p><p>Microsoft Security Blog. (2026, March 6). <em>AI as tradecraft: How threat actors operationalize AI</em>. <a href="https://www.microsoft.com/en-us/security/blog/2026/03/06/ai-as-tradecraft-how-threat-actors-operationalize-ai/">https://www.microsoft.com/en-us/security/blog/2026/03/06/ai-as-tradecraft-how-threat-actors-operationalize-ai/</a></p><p>National Institute of Standards and Technology. (2026, January). <em>CAISI issues request for information about securing AI agent systems</em>. <a href="https://www.nist.gov/news-events/news/2026/01/caisi-issues-request-information-about-securing-ai-agent-systems">https://www.nist.gov/news-events/news/2026/01/caisi-issues-request-information-about-securing-ai-agent-systems</a></p><p>National Institute of Standards and Technology. (2026, February). <em>Announcing the AI agent standards initiative for interoperable and secure innovation</em>. <a href="https://www.nist.gov/news-events/news/2026/02/announcing-ai-agent-standards-initiative-interoperable-and-secure">https://www.nist.gov/news-events/news/2026/02/announcing-ai-agent-standards-initiative-interoperable-and-secure</a></p><p>OpenAI. (2026, March 6). <em>Codex Security: Now in research preview</em>. <a href="https://openai.com/index/codex-security-now-in-research-preview/">https://openai.com/index/codex-security-now-in-research-preview/</a></p><p>OpenAI. (2026, March 12). <em>Understanding prompt injections: A frontier security challenge</em>. <a href="https://openai.com/index/prompt-injections/">https://openai.com/index/prompt-injections/</a></p><p>OpenAI. (2026). <em>Continuously hardening ChatGPT Atlas against prompt injection attacks</em>. <a href="https://openai.com/index/hardening-atlas-against-prompt-injection/">https://openai.com/index/hardening-atlas-against-prompt-injection/</a></p><p>OpenID Foundation. (2026). <em>OIDF responds to NIST on AI agent security</em>. <a href="https://openid.net/oidf-responds-to-nist-on-ai-agent-security/">https://openid.net/oidf-responds-to-nist-on-ai-agent-security/</a></p><p>Palo Alto Networks. (2026, February). <em>2026 Unit 42 global incident response report: Attacks now 4x faster</em>. <a href="https://www.paloaltonetworks.com/blog/2026/02/unit-42-global-ir-report/">https://www.paloaltonetworks.com/blog/2026/02/unit-42-global-ir-report/</a></p><p>PrismNews. (2026, March). <em>OpenAI releases engineering playbook to shield AI agents from prompt injection</em>. <a href="https://www.prismnews.com/news/openai-releases-engineering-playbook-to-shield-ai-agents">https://www.prismnews.com/news/openai-releases-engineering-playbook-to-shield-ai-agents</a></p><p>Security Boulevard. (2026, March). <em>83% of cloud breaches start with identity, AI agents are about to make it worse</em>. <a href="https://securityboulevard.com/2026/03/83-of-cloud-breaches-start-with-identity-ai-agents-are-about-to-make-it-worse/">https://securityboulevard.com/2026/03/83-of-cloud-breaches-start-with-identity-ai-agents-are-about-to-make-it-worse/</a></p><p>SecurityWeek. (2026, March 6). <em>OpenAI rolls out Codex Security vulnerability scanner</em>. <a href="https://www.securityweek.com/openai-rolls-out-codex-security-vulnerability-scanner/">https://www.securityweek.com/openai-rolls-out-codex-security-vulnerability-scanner/</a></p><p>TechRadar. (2026, March 6). <em>OpenAI releases Codex Security to spot the next big cyber risks to your company</em>. <a href="https://www.techradar.com/pro/security/openai-releases-codex-security-to-spot-the-next-big-cyber-risks-to-your-company-promises-to-identify-complex-vulnerabilities-that-other-agentic-tools-miss">https://www.techradar.com/pro/security/openai-releases-codex-security-to-spot-the-next-big-cyber-risks-to-your-company-promises-to-identify-complex-vulnerabilities-that-other-agentic-tools-miss</a></p><p>The Hacker News. (2026, March 12). <em>CISA flags actively exploited n8n RCE bug as 24,700 instances remain exposed</em>. <a href="https://thehackernews.com/2026/03/cisa-flags-actively-exploited-n8n-rce.html">https://thehackernews.com/2026/03/cisa-flags-actively-exploited-n8n-rce.html</a></p><p>The Register. (2026, March 6). <em>Anthropic sues US over national security blacklist</em>. <a href="https://www.theregister.com/2026/03/06/anthropic_left_with_no_other/">https://www.theregister.com/2026/03/06/anthropic_left_with_no_other/</a></p><p>The Register. (2026, March 8). <em>Manage attack infrastructure? AI agents can now help</em>. <a href="https://www.theregister.com/2026/03/08/deploy_and_manage_attack_infrastructure/">https://www.theregister.com/2026/03/08/deploy_and_manage_attack_infrastructure/</a></p><p>The Register. (2026, March 10). <em>Amazon insists AI coding isn&#8217;t source of outages</em>. <a href="https://www.theregister.com/2026/03/10/amazon_ai_coding_outages/">https://www.theregister.com/2026/03/10/amazon_ai_coding_outages/</a></p><p>The Register. (2026, March 12). <em>CISA says n8n critical bug exploited in real-world attacks</em>. <a href="https://www.theregister.com/2026/03/12/cisa_n8n_rce/">https://www.theregister.com/2026/03/12/cisa_n8n_rce/</a></p><p>U.S. Federal Register. (2026, January 8). <em>Request for information regarding security considerations for artificial intelligence agents</em>. <a href="https://www.federalregister.gov/documents/2026/01/08/2026-00206/request-for-information-regarding-security-considerations-for-artificial-intelligence-agents">https://www.federalregister.gov/documents/2026/01/08/2026-00206/request-for-information-regarding-security-considerations-for-artificial-intelligence-agents</a></p><p>Veracode. (2026, February 24). <em>2026 state of software security report</em>. BusinessWire. <a href="https://www.businesswire.com/news/home/20260224526703/en/Veracode-2026-State-of-Software-Security-Report-Reveals-Four-Out-of-Five-Organizations-Are-Drowning-in-Security-Debt">https://www.businesswire.com/news/home/20260224526703/en/Veracode-2026-State-of-Software-Security-Report-Reveals-Four-Out-of-Five-Organizations-Are-Drowning-in-Security-Debt</a></p><p>White House. (2026, March). <em>White House unveils President Trump&#8217;s cyber strategy for America</em>. <a href="https://www.whitehouse.gov/articles/2026/03/white-house-unveils-president-trumps-cyber-strategy-for-america/">https://www.whitehouse.gov/articles/2026/03/white-house-unveils-president-trumps-cyber-strategy-for-america/</a></p>]]></content:encoded></item><item><title><![CDATA[AI Vendor Lock-In: What the Pentagon Taught Every CISO This Week]]></title><description><![CDATA[The DoD's Anthropic supply chain risk designation exposed every enterprise's embedded AI architecture gap. Here's what your vendor contracts are missing.]]></description><link>https://www.rockcybermusings.com/p/ai-vendor-lock-in-pentagon-anthropic-ciso-lesson</link><guid isPermaLink="false">https://www.rockcybermusings.com/p/ai-vendor-lock-in-pentagon-anthropic-ciso-lesson</guid><dc:creator><![CDATA[Rock Lambros]]></dc:creator><pubDate>Tue, 10 Mar 2026 12:50:49 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!rMq7!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F949ab505-5594-453f-b968-f0333f1fa094_2048x2048.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!rMq7!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F949ab505-5594-453f-b968-f0333f1fa094_2048x2048.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!rMq7!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F949ab505-5594-453f-b968-f0333f1fa094_2048x2048.jpeg 424w, https://substackcdn.com/image/fetch/$s_!rMq7!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F949ab505-5594-453f-b968-f0333f1fa094_2048x2048.jpeg 848w, https://substackcdn.com/image/fetch/$s_!rMq7!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F949ab505-5594-453f-b968-f0333f1fa094_2048x2048.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!rMq7!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F949ab505-5594-453f-b968-f0333f1fa094_2048x2048.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!rMq7!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F949ab505-5594-453f-b968-f0333f1fa094_2048x2048.jpeg" width="1456" height="1456" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/949ab505-5594-453f-b968-f0333f1fa094_2048x2048.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1456,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:3304173,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.rockcybermusings.com/i/190372517?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F949ab505-5594-453f-b968-f0333f1fa094_2048x2048.jpeg&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!rMq7!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F949ab505-5594-453f-b968-f0333f1fa094_2048x2048.jpeg 424w, https://substackcdn.com/image/fetch/$s_!rMq7!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F949ab505-5594-453f-b968-f0333f1fa094_2048x2048.jpeg 848w, https://substackcdn.com/image/fetch/$s_!rMq7!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F949ab505-5594-453f-b968-f0333f1fa094_2048x2048.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!rMq7!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F949ab505-5594-453f-b968-f0333f1fa094_2048x2048.jpeg 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>You probably don&#8217;t know which AI model is running inside your operational tools right now. That&#8217;s a near-certainty given how enterprise AI procurement actually works. The Pentagon just ran a live stress test on that exact blind spot, and the results were not subtle. When the Department of War formally designated Anthropic a supply chain risk on March 5, 2026, making it the first American company in history to receive a label previously reserved for Huawei and Chinese state-adjacent tech firms, the disruption didn&#8217;t start with Anthropic. It cascaded through Palantir, across AWS infrastructure, and into active military workflows during U.S. strikes on Iran. Your enterprise has the same layered architecture. The question is whether you&#8217;ve mapped it, and whether your contracts protect you when the layer you don&#8217;t control catches fire.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.rockcybermusings.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading RockCyber Musings! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.rockcybermusings.com/p/ai-vendor-lock-in-pentagon-anthropic-ciso-lesson?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.rockcybermusings.com/p/ai-vendor-lock-in-pentagon-anthropic-ciso-lesson?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p><h2>The AI Model You Don&#8217;t Control Is Already in Production</h2><p>The DoD&#8217;s direct customer relationship wasn&#8217;t with Anthropic. Claude ran inside Palantir&#8217;s Maven Smart System, hosted on AWS at Impact Level 6, sitting on classified infrastructure the military depended on for intelligence analysis and operational planning. The DoD contracted with Palantir. Palantir embedded Claude. When the supply chain risk designation landed, it cascaded from procurement machinery through Palantir&#8217;s operational position and into workflows with real military dependencies, reportedly including active support for Iran strikes, even as the designation was being disputed on social media by the Secretary of Defense and the CEO of Anthropic simultaneously.</p><p>Piper Sandler analysts noted after the designation that Anthropic was &#8220;heavily embedded in the Military and the Intelligence community&#8221; and that migrating off the technology could &#8220;pose some short-term disruptions&#8221; to Palantir&#8217;s operations. Short-term disruptions. During an active military operation. That&#8217;s the polite Wall Street version of the problem.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!ajJ9!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F84347e25-14b4-4c9b-92e8-9de95b69f075_7274x3250.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!ajJ9!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F84347e25-14b4-4c9b-92e8-9de95b69f075_7274x3250.png 424w, https://substackcdn.com/image/fetch/$s_!ajJ9!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F84347e25-14b4-4c9b-92e8-9de95b69f075_7274x3250.png 848w, https://substackcdn.com/image/fetch/$s_!ajJ9!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F84347e25-14b4-4c9b-92e8-9de95b69f075_7274x3250.png 1272w, https://substackcdn.com/image/fetch/$s_!ajJ9!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F84347e25-14b4-4c9b-92e8-9de95b69f075_7274x3250.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!ajJ9!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F84347e25-14b4-4c9b-92e8-9de95b69f075_7274x3250.png" width="1456" height="651" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/84347e25-14b4-4c9b-92e8-9de95b69f075_7274x3250.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:651,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:1013321,&quot;alt&quot;:&quot;Flowchart showing how Claude was embedded through Palantir Maven Smart System and AWS IL6 into DoD operational workflows, with a parallel enterprise layer showing the same pattern across SaaS vendor, foundation model, and cloud provider&quot;,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.rockcybermusings.com/i/190372517?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F84347e25-14b4-4c9b-92e8-9de95b69f075_7274x3250.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="Flowchart showing how Claude was embedded through Palantir Maven Smart System and AWS IL6 into DoD operational workflows, with a parallel enterprise layer showing the same pattern across SaaS vendor, foundation model, and cloud provider" title="Flowchart showing how Claude was embedded through Palantir Maven Smart System and AWS IL6 into DoD operational workflows, with a parallel enterprise layer showing the same pattern across SaaS vendor, foundation model, and cloud provider" srcset="https://substackcdn.com/image/fetch/$s_!ajJ9!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F84347e25-14b4-4c9b-92e8-9de95b69f075_7274x3250.png 424w, https://substackcdn.com/image/fetch/$s_!ajJ9!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F84347e25-14b4-4c9b-92e8-9de95b69f075_7274x3250.png 848w, https://substackcdn.com/image/fetch/$s_!ajJ9!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F84347e25-14b4-4c9b-92e8-9de95b69f075_7274x3250.png 1272w, https://substackcdn.com/image/fetch/$s_!ajJ9!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F84347e25-14b4-4c9b-92e8-9de95b69f075_7274x3250.png 1456w" sizes="100vw"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Figure 1: The Embedded AI Architecture Problem</figcaption></figure></div><p>Replace &#8220;Military and Intelligence community&#8221; with your sector. Replace &#8220;Palantir&#8221; with your largest workflow vendor. Replace &#8220;active military operation&#8221; with your peak fraud season, your annual close, or your next regulatory audit. You&#8217;ve just described your own exposure.</p><p>Your enterprise equivalent of Maven isn&#8217;t a targeting system. It&#8217;s the fraud detection platform your SOC relies on for alert triage. It&#8217;s the contract review tool your legal team treats as a first pass on every agreement. It&#8217;s the SIEM enrichment workflow your analysts approved 18 months ago, without anyone asking which model was under it or whose usage policy governed it. In each case, there&#8217;s a foundation model embedded by a SaaS vendor, hosted by a cloud provider, running under policies you never reviewed and almost certainly can&#8217;t enforce. The vendor who sold you the platform might not even know which model version was deployed last Tuesday.</p><p>The lock-in risk most CISOs think about is the wrong one. They worry about pricing leverage at renewal or feature gaps during the next budget cycle. Those are real, and they&#8217;re also the least interesting version of vendor risk in an AI-dependent stack. The risk that actually bites is operational dependency on a model whose policies, safety stack, and external political relationships sit entirely outside your contractual reach. This week demonstrated those conditions shift in 48 hours. When they do, you find out how embedded you actually are. The DoD found out during airstrikes. You&#8217;ll find out during something comparably inconvenient for you.</p><h2>What the Contract Language Reveals About Your Own Agreements</h2><p>The factual record on the Anthropic negotiation is clear enough. The Department of War&#8217;s January 2026 AI strategy memorandum directed procurement to require &#8220;any lawful use&#8221; language and to acquire models &#8220;free from usage policy constraints that may limit lawful military applications.&#8221; Anthropic held two red lines: no mass domestic surveillance of Americans, and no fully autonomous weapons with no human in the targeting decision loop. The DoD called those constraints unacceptable. The negotiation collapsed. The designation followed.</p><p>Here&#8217;s where it gets interesting... OpenAI reached a deal within hours of the designation announcement, published contract excerpts containing the exact &#8220;all lawful purposes&#8221; language Anthropic refused, then amended the agreement twice in the following week after legal experts publicly tore apart what the protections actually meant. Sam Altman acknowledged the deal was &#8220;definitely rushed&#8221; and that &#8220;the optics don&#8217;t look good.&#8221; Jessica Tillipman, associate dean for government procurement law studies at George Washington University, wrote that the published excerpt &#8220;does not give OpenAI an Anthropic-style, free-standing right to prohibit otherwise-lawful government use.&#8221; Altman signed it anyway. To be fair to him, he was working in 48-hour crisis mode while a competing lab was being designated a national security threat. Good contract hygiene was not the priority.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!aI2j!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F60fa953b-375a-4154-9a81-a59c07cadb40_3591x7020.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!aI2j!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F60fa953b-375a-4154-9a81-a59c07cadb40_3591x7020.png 424w, https://substackcdn.com/image/fetch/$s_!aI2j!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F60fa953b-375a-4154-9a81-a59c07cadb40_3591x7020.png 848w, https://substackcdn.com/image/fetch/$s_!aI2j!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F60fa953b-375a-4154-9a81-a59c07cadb40_3591x7020.png 1272w, https://substackcdn.com/image/fetch/$s_!aI2j!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F60fa953b-375a-4154-9a81-a59c07cadb40_3591x7020.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!aI2j!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F60fa953b-375a-4154-9a81-a59c07cadb40_3591x7020.png" width="1456" height="2846" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/60fa953b-375a-4154-9a81-a59c07cadb40_3591x7020.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:2846,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:1084216,&quot;alt&quot;:&quot;Comparison diagram contrasting Anthropic&#8217;s standalone vendor-imposed prohibition approach with OpenAI&#8217;s law-anchored permissive use framework, including identified gaps in each&quot;,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.rockcybermusings.com/i/190372517?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F60fa953b-375a-4154-9a81-a59c07cadb40_3591x7020.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="Comparison diagram contrasting Anthropic&#8217;s standalone vendor-imposed prohibition approach with OpenAI&#8217;s law-anchored permissive use framework, including identified gaps in each" title="Comparison diagram contrasting Anthropic&#8217;s standalone vendor-imposed prohibition approach with OpenAI&#8217;s law-anchored permissive use framework, including identified gaps in each" srcset="https://substackcdn.com/image/fetch/$s_!aI2j!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F60fa953b-375a-4154-9a81-a59c07cadb40_3591x7020.png 424w, https://substackcdn.com/image/fetch/$s_!aI2j!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F60fa953b-375a-4154-9a81-a59c07cadb40_3591x7020.png 848w, https://substackcdn.com/image/fetch/$s_!aI2j!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F60fa953b-375a-4154-9a81-a59c07cadb40_3591x7020.png 1272w, https://substackcdn.com/image/fetch/$s_!aI2j!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F60fa953b-375a-4154-9a81-a59c07cadb40_3591x7020.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Figure 2: Red Lines vs. Legal Anchors: Two Approaches to AI Contract Protection </figcaption></figure></div><p>Instead of wasting your time on the OpenAI vs. Anthropic drama and who is right or wrong, you need to pay attention to the legal architecture underlying AI safety commitments.</p><p>Why?</p><p>Because your enterprise contracts almost certainly follow the same pattern OpenAI accepted, that include usage restrictions anchored to &#8220;applicable law&#8221; and &#8220;existing policy,&#8221; with the vendor&#8217;s safety stack as the primary enforcement mechanism. OpenAI anchored its protections to existing statutes: the Fourth Amendment, FISA, DoD Directive 3000.09 on autonomous weapons, and Executive Order 12333. Critics flagged immediately that EO 12333 is the authority the NSA has historically used to justify intercepting Americans&#8217; communications through collection outside U.S. borders. &#8220;Lawful&#8221; in national security contexts isn&#8217;t a fixed boundary. It lives inside classified legal interpretations, executive orders, and internal agency guidance nobody outside the building ever reads.</p><p>Your enterprise contracts with AI vendors operate the same way. When law shifts, when policy changes, or when your vendor faces its own version of a 48-hour political deadline, those anchors move with the situation. What your procurement posture needs instead are vendor-imposed, free-standing prohibited-use schedules for your specific high-risk workflows, written into contract appendices with attached audit rights and defined remedies. &#8220;We comply with applicable law&#8221; is a description of baseline legal obligation. It&#8217;s not a control. It&#8217;s what every vendor says about every product, whether or not AI is involved. You shouldn&#8217;t be paying for that sentence in an AI addendum. You should be getting something that took a lawyer to write specifically for your deployment.</p><h2>Human-in-the-Loop Theater</h2><p>Let me describe a workflow you probably have running right now. Your AI triage layer ingests 200 alerts per shift and flags 180 as low severity. Your analyst reviews the queue, confirms the model&#8217;s assessment on most items, escalates five, clears the rest. Total elapsed review time for the cleared items is, let&#8217;s say, roughly two minutes each. Every disposition went through a human. The audit log shows human review. Your controls documentation references human oversight. What actually happened is your analyst ratified model outputs under cognitive load and time pressure while telling themselves they were exercising judgment.</p><p>That&#8217;s the failure mode human-in-the-loop review was designed to prevent. The loop exists on paper. The friction isn&#8217;t in the workflow design because no step requires the reviewer to explain why they agree with the model before confirming the output. Nobody required forced alternative generation before escalating or clearing. Nobody captured uncertainty as a structured field. The control is decorative.</p><p>The OpenAI contract&#8217;s autonomous weapons provision bars the use of the AI system &#8220;to independently direct autonomous weapons in any case where law, regulation, or Department policy requires human control.&#8221; Defense scholars noted the omission of &#8220;human-in-the-loop&#8221; language was deliberate, preserving operational flexibility. &#8220;Human judgment&#8221; and &#8220;human control&#8221; are not equivalent, and the people drafting that language knew it. The contract borrows its enforceability entirely from existing policy, which requires commanders to exercise &#8220;appropriate levels of human judgment over the use of force.&#8221; Appropriate is not a control. It&#8217;s a word that means whatever the decision-maker concludes is appropriate under the circumstances they&#8217;re actually in.</p><p>Research from King&#8217;s College London found that tested AI models threatened nuclear strikes in 95% of simulated crisis scenarios. The problem wasn&#8217;t autonomous weapons. The problem was that under uncertainty and time pressure, models produced escalatory recommendations with false confidence, and human reviewers were positioned to ratify those outputs rather than interrogate them. That&#8217;s not a future risk. That&#8217;s automation bias, and it operates in your environment every shift, at every tier of your AI-assisted workflows.</p><p>The Lavender targeting system used by Israeli defense forces was reportedly identified by investigators as carrying a 10% false positive rate on human identification, with human reviewers present throughout the process. The investigation raised a direct question of whether those humans were genuinely reviewing or functionally ratifying outputs under operational tempo. That distinction carries different consequences in contexts outside the military. In your environment, it shows up as a miscategorized fraud case that costs a customer their account, or a misconfigured access control that cleared review because the analyst trusted the model&#8217;s output and moved on in the last four minutes of a shift.</p><p>Building real decision friction requires designing it into the workflow architecture before something goes wrong, not auditing for it afterward. Two-person review for high-consequence AI outputs. Forced alternative generation before an analyst confirms a model recommendation. Explicit uncertainty capture as a required structured field. If your current AI-assisted workflows don&#8217;t require a reviewer to articulate why they agree with the model&#8217;s output before confirming it, then you are rubber-stamping your way into a problem down the road. You may survive your next audit. Youwon&#8217;t survive your next incident.</p><h2>The Procurement Posture That Needs to Change Before the Next Signature</h2><p>Most CISOs don&#8217;t own AI vendor contracts. Procurement does. Legal does. The CISO inherits the agreement after signature, usually after the vendor relationship is already operational and the leverage window has closed. This is the moment where I&#8217;ll stop pretending that&#8217;s a systems failure and call it what it is: CISOs have let themselves get cut out of a decision that&#8217;s now one of the highest-risk commitments their organization makes. The Anthropic situation gives you the publicly documented argument to change that for every AI agreement with operational or regulatory exposure going forward.</p><p>The DoD&#8217;s relationship with Palantir didn&#8217;t include enforceable audit rights over Claude&#8217;s underlying usage policy, safety stack updates, or model variant changes. When Anthropic&#8217;s relationship with the DoD broke down, Palantir faced operational disruption from a vendor dependency it hadn&#8217;t fully governed at the model layer. Your enterprise equivalent is any SaaS vendor who embeds a foundation model in a production workflow without explicit flow-down contract obligations. You need those flow-down provisions now: contractual requirements for your SaaS vendors to notify you of material AI policy changes, with a defined right to pause deployment or terminate.</p><p>Anthropic&#8217;s published usage policy states the company may tailor restrictions for certain customers based on mission and legal authorities, subject to Anthropic&#8217;s judgment about safeguards. That clause exists in their public policy documentation. Most of their enterprise customers have never read it, don&#8217;t know whether their deployment is governed by standard or tailored terms, and have no contractual mechanism to find out. If you&#8217;re an Anthropic customer and you don&#8217;t know the answer to that question, the answer is almost certainly that you don&#8217;t know, which means you don&#8217;t control it.</p><p>Splunk&#8217;s 2026 CISO Report found that a large majority of CISOs carry personal liability concerns about security incidents. AI model misuse by a subcontractor or an embedded model that you didn&#8217;t govern is exactly the incident scenario that tests that liability question. Your current contract schedules almost certainly don&#8217;t address it. Here are the questions that need to be in every AI vendor negotiation before signature, not as a wish list, but as conditions of signature:</p><ul><li><p>Which model variant governs your deployment, and does that variant deviate from the vendor&#8217;s published acceptable use policy or baseline safety commitments? Get the answer in writing with a version reference.</p></li><li><p>What change control process governs model updates, safety stack revisions, and policy changes? &#8220;We update continuously&#8221; is not an answer. You need customer notice requirements and the right to pause deployment when the vendor makes a material change.</p></li><li><p>What logs exist, who holds access, and what is the retention period? Without logs you can&#8217;t support an incident investigation, a regulatory inquiry, or your own post-incident analysis.</p></li><li><p>What happens when a major customer, a regulator, or a government agency demands scope expansion for your deployment? The Anthropic situation confirmed this question isn&#8217;t hypothetical. It&#8217;s a negotiating dynamic triggered externally, rapidly, and without advance warning to downstream customers.</p></li></ul><h2>From the Run Phase to the Evolve Phase</h2><p>If you&#8217;re applying the CARE framework, this situation signals that you&#8217;re overdue for an Evolve-phase review of your AI vendor relationships. The Create and Adapt work produced your current model integrations. Most organizations have stayed in the Run phase, monitoring performance and managing routine issues, while the risk environment underneath those integrations has shifted significantly. The Evolve phase requires reassessing whether the governance model you built for each AI deployment still fits the world you&#8217;re operating in now.</p><p>The Anthropic situation changed that environment in three concrete ways your board needs to understand. First, it showed that an AI vendor&#8217;s political and contractual relationships with high-profile customers now represent operational risk to every downstream customer, not only government contractors. Second, it produced a documented public case where contract language anchored to &#8220;applicable law&#8221; failed to deliver the protections a party believed it had agreed to. Third, it revealed that model replacement timelines are slower than your AI vendors implied during the sales process. The DoD, with its classified infrastructure, operational urgency, considerable resources, and six-month transition timeline, is the fastest-moving version of this problem you&#8217;re likely to encounter. Your enterprise timeline almost certainly isn&#8217;t shorter.</p><p>Build your AI vendor risk registry before something breaks, while relationships are functional and vendors are cooperative. Map every production AI deployment to the model underneath it, the vendor who embeds it, the cloud provider who hosts it, and the contract that governs each layer. Run a prohibited-use gap assessment: which categories of use does each contract explicitly prohibit, and are those prohibitions free-standing or anchored to &#8220;applicable law&#8221;? Apply OWASP&#8217;s Agentic Top 10 to any workflow where a model makes or influences a decision without a mandatory human review step that requires documented rationale.</p><p>The CISOs who were ahead of this story weren&#8217;t tracking the Pentagon news cycle. They had already asked their SaaS vendors which model was embedded, what the vendor&#8217;s posture would be if that model&#8217;s policy changed, and what their exit path looked like. Most got vague answers. The right response to a vague answer from an AI vendor is a contract clause, not a follow-up email.</p><p><strong>Key Takeaway:</strong> Your AI vendor&#8217;s ethics statement doesn&#8217;t protect your enterprise. A free-standing prohibited-use schedule, enforceable audit rights, and model-layer flow-down provisions do.</p><h3>What to Do Next</h3><p>Start with a model inventory audit across your top ten SaaS vendor relationships. Ask each vendor to identify the foundation model embedded in your production workflows and provide the current acceptable use policy governing your specific deployment, including any tailored terms. Map the gap between what the policy says and what your contract actually enforces.</p><p>The Anthropic situation is the most instructive public case study on AI vendor governance to emerge from this space. Use it while it&#8217;s in front of your board and before your next AI vendor signature lands on someone else&#8217;s desk.</p><p>&#128073; Subscribe for more AI security and governance insights with the occasional rant.</p><p>&#128073; Visit <strong><a href="https://www.rockcyber.com/">RockCyber.com</a></strong> to learn more about how we can help you in your traditional Cybersecurity and AI Security and Governance Journey</p><p>&#128073; Want to save a quick $100K? Check out our AI Governance Tools at <strong><a href="https://aigovernancetoolkit.com/">AIGovernanceToolkit.com</a></strong></p><p><em>The views and opinions expressed in RockCyber Musings are my own and do not represent the positions of my employer or any organization I&#8217;m affiliated with.</em></p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.rockcybermusings.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading RockCyber Musings! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.rockcybermusings.com/?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share RockCyber Musings&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.rockcybermusings.com/?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share RockCyber Musings</span></a></p>]]></content:encoded></item><item><title><![CDATA[Weekly Musings Top 10 AI Security Wrapup: Issue 28 February 27, 2026 - March 5, 2026]]></title><description><![CDATA[When AI Attacks AI: The Agentic Threat Era Arrives in Full]]></description><link>https://www.rockcybermusings.com/p/weekly-musings-top-10-ai-security-20260227-202600305</link><guid isPermaLink="false">https://www.rockcybermusings.com/p/weekly-musings-top-10-ai-security-20260227-202600305</guid><dc:creator><![CDATA[Rock Lambros]]></dc:creator><pubDate>Fri, 06 Mar 2026 13:47:08 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!ko-M!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F075e6604-ca52-4950-8813-044a77a98100_1024x1024.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!ko-M!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F075e6604-ca52-4950-8813-044a77a98100_1024x1024.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!ko-M!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F075e6604-ca52-4950-8813-044a77a98100_1024x1024.png 424w, https://substackcdn.com/image/fetch/$s_!ko-M!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F075e6604-ca52-4950-8813-044a77a98100_1024x1024.png 848w, https://substackcdn.com/image/fetch/$s_!ko-M!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F075e6604-ca52-4950-8813-044a77a98100_1024x1024.png 1272w, https://substackcdn.com/image/fetch/$s_!ko-M!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F075e6604-ca52-4950-8813-044a77a98100_1024x1024.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!ko-M!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F075e6604-ca52-4950-8813-044a77a98100_1024x1024.png" width="1024" height="1024" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/075e6604-ca52-4950-8813-044a77a98100_1024x1024.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1024,&quot;width&quot;:1024,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:1233556,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.rockcybermusings.com/i/190100547?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F075e6604-ca52-4950-8813-044a77a98100_1024x1024.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!ko-M!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F075e6604-ca52-4950-8813-044a77a98100_1024x1024.png 424w, https://substackcdn.com/image/fetch/$s_!ko-M!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F075e6604-ca52-4950-8813-044a77a98100_1024x1024.png 848w, https://substackcdn.com/image/fetch/$s_!ko-M!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F075e6604-ca52-4950-8813-044a77a98100_1024x1024.png 1272w, https://substackcdn.com/image/fetch/$s_!ko-M!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F075e6604-ca52-4950-8813-044a77a98100_1024x1024.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.rockcybermusings.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading RockCyber Musings! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.rockcybermusings.com/p/weekly-musings-top-10-ai-security-20260227-202600305?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.rockcybermusings.com/p/weekly-musings-top-10-ai-security-20260227-202600305?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p><p>This week handed security leaders something they&#8217;ve been theorizing about for two years: autonomous AI agents attacking other autonomous AI agents in live production environments. No thought experiment, no conference demo. A malicious bot using Claude Opus 4.5 compromised five major open-source repositories. An AI-native offensive platform compromised 600 firewalls across 55 countries. Developer tools turned into attack vectors by opening a Git repo.</p><p>The practitioner community doing the real work on these problems gathered at [un]prompted in San Francisco. The rest of the week&#8217;s news served as a live demonstration of why that conference needed to exist. Attackers aren&#8217;t waiting for frameworks to catch up. Your AI tools are the attack surface now. The developers building them are the initial targets. The agents those tools spawn are the next ones.</p><div><hr></div><h3>1. [Un]Prompted Delivers the AI Security Conference the Industry Needed</h3><p>The first [un]prompted conference ran March 3-4 at The Hibernia in San Francisco (unpromptedcon.org). Gadi Evron of Knostic, who chaired the conference, received nearly 500 talk submissions and built a program spanning offense, defense, DFIR, and governance. No vendor theater. Confirmed speakers included Heather Adkins from Google on advancing code security, Joshua Saxe from Meta on agent evaluation, Paul McMillan from OpenAI on securing software in the agentic era, and Nicholas Carlini from Anthropic on black-hat LLMs finding zero-days in production codebases. Dan Guido closed Day Two, explaining how Trail of Bits rebuilt around AI to reach 200 bugs per engineer per week. Sergej Epp from Sysdig presented primary forensic evidence from an 8-minute AWS escalation and EtherRAT, a blockchain C2 campaign. Gadi even stepped in for Avishai Efrat and Michael Barugy from Zenity&#8230;a direct competitor&#8230; who could not get out of Israel, to drop PleaseFix.</p><p><strong>Why it matters</strong></p><ul><li><p>The field now has a practitioner-grade conference built for people doing actual work, from red teamers to governance leads, not vendor keynotes disguised as research.</p></li><li><p>The offensive capability context is essential. Carlini showed current models finding zero-days. Guido showed 200 bugs per engineer per week. Defenders need this before building programs.</p></li><li><p>The governance track didn&#8217;t retreat into frameworks. Healthcare and large enterprise practitioners spoke about what actually works in production.</p></li></ul><p><strong>What to do about it</strong></p><ul><li><p>Read the full agenda at unpromptedcon.org. The talk abstracts contain more actionable signal than most vendor white papers.</p></li><li><p>Follow the researchers presenting there. Those names are shaping the actual threat landscape.</p></li><li><p>Prioritize the Stripe threat modeling talks and the Snap capability-based authorization session if your team hasn&#8217;t treated AI agents as first-class attack surfaces yet.</p></li></ul><p><strong>Rock&#8217;s Musings</strong></p><p>Rob T. Lee&#8217;s line on Stage 2 deserves repeating. Anthropic&#8217;s own GTG-1002 report showed adversaries running Claude Code at 80-90% autonomous execution. Your adversary has an AI. If you&#8217;re at tab-completion for defense, that&#8217;s a strategic failure, not a skills gap.</p><p>I&#8217;ve been going to security conferences for a long time. Most are marketing events with technical content as decoration. [un]prompted felt different because Gadi built it explicitly for people who know what a YAML file does. That&#8217;s a rare thing and worth supporting. Start planning for year two.</p><div><hr></div><h3>2. Hackerbot-Claw Proved Autonomous AI Can Systematically Destroy Your CI/CD Pipeline</h3><p>Between February 21 and March 1, 2026, a GitHub account called hackerbot-claw ran an autonomous campaign against public repositories (StepSecurity). The account describes itself as an &#8220;autonomous security research agent powered by claude-opus-4-5,&#8221; maintains a vulnerability pattern index with 9 classes and 47 sub-patterns, and claims to have scanned 47,391 repositories. The bot achieved remote code execution in at least four of seven targeted repositories, including Microsoft, DataDog, CNCF, and Aqua Security&#8217;s Trivy scanner. In the Trivy compromise, it stole a Personal Access Token with broad write permissions, deleted all 178 GitHub releases, wiped repository content, and published a malicious VSCode extension to OpenVSX under Trivy&#8217;s trusted publisher identity. OpenSSF issued a TLP:CLEAR advisory March 1.</p><p>The single defining moment: the bot attempted prompt injection against a Claude-based CI workflow at ambient-code/platform. Claude, running claude-sonnet-4-6, classified it as &#8220;a textbook AI agent supply-chain attack via poisoned project-level instructions&#8221; and refused. The only target the bot failed to compromise was protected by another AI model recognizing the attack.</p><p><strong>Why it matters</strong></p><ul><li><p>CI/CD misconfigurations are now mass-exploitable at machine speed without a single CVE. Five documented exploitation techniques, all using known patterns, all automatable.</p></li><li><p>Supply-chain compromise at scale doesn&#8217;t require sophisticated malware. It requires systematic scanning and pull request automation. The bot scanned 47,000 repos in a week.</p></li><li><p>AI-versus-AI defense is no longer theoretical. The ambient-code defense worked because someone built proper tool allowlisting with prompt injection detection.</p></li></ul><p><strong>What to do about it</strong></p><ul><li><p>Audit every pull_request_target workflow in your repositories this week. Move PR metadata into environment variables. Scope tokens to minimum permissions.</p></li><li><p>Verify your AI-based code review toolchain has prompt injection detection and tool allowlisting. Configuration matters as much as the model.</p></li><li><p>Check the OpenSSF advisory for the specific pattern list hackerbot-claw exploited. These are all preventable and all still present in thousands of active repositories.</p></li></ul><p><strong>Rock&#8217;s Musings</strong></p><p>The &#8220;security research&#8221; framing in the account bio is working hard. Deleting 32,000 stars from Trivy and pushing a malicious extension to OpenVSX isn&#8217;t research. The creator remains unidentified. The domain name, the &#8220;molt&#8221; naming, and the OpenClaw ecosystem references point to infrastructure being assembled and tested in the open because the operators know defenders aren&#8217;t watching yet. We&#8217;re watching the emergence of an offensive AI toolkit in real time.</p><div><hr></div><h3>3. CyberStrikeAI: A Chinese-Linked Offensive Platform Hit 600 Firewalls Across 55 Countries</h3><p>Team Cymru published research on March 3, naming CyberStrikeAI as the AI-native offensive tool behind the FortiGate campaign disclosed by Amazon Threat Intelligence in February (BleepingComputer, The Hacker News). The campaign ran from January 11 to February 18, 2026, comprising over 600 FortiGate devices across 55 countries. CyberStrikeAI is built in Go, integrates 100-plus security tools, and uses any OpenAI-compatible model, including Claude and DeepSeek, through an MCP orchestration engine. The developer, alias Ed1s0nZ, submitted the tool to Knownsec 404&#8217;s Starlink Project in December 2025 and briefly posted a CNNVD vulnerability credential to their GitHub profile before deleting it. CNNVD operates under oversight by China&#8217;s Ministry of State Security. Team Cymru detected 21 unique IPs running CyberStrikeAI between January 20 and February 26, primarily on Chinese cloud infrastructure. No zero-days exploited. The actor succeeded through exposed management interfaces and weak credentials.</p><p><strong>Why it matters</strong></p><ul><li><p>AI-native offensive platforms are open-source and in active deployment. The barrier to running a 600-device campaign across 55 countries is now a GitHub clone and a cloud account.</p></li><li><p>State-adjacent tooling proliferates fast. Zero deployments in November to 21 active servers by late February is an adoption curve worth tracking.</p></li><li><p>The entry point remains unchanged. Sophisticated AI orchestration amplified the attacker. Exposed management interfaces created the opportunity. Harden the basics first.</p></li></ul><p><strong>What to do about it</strong></p><ul><li><p>Pull the FortiGate management interface exposure from public networks immediately (seriously&#8230; who do we have to keep saying this?). Apply all current firmware patches.</p></li><li><p>Add CyberStrikeAI IOCs from the Team Cymru report to your threat intelligence feeds.</p></li><li><p>Add AI-native offensive tooling as a threat category in your risk model. The economics of running large-scale exploitation campaigns changed this quarter.</p></li></ul><p><strong>Rock&#8217;s Musings</strong></p><p>The credential scrub tells you something about the actor&#8217;s maturity. Ed1s0nZ posted the CNNVD award, realized the optics problem, and deleted it. Git commit history preserved both moves. This is someone running a 600-device campaign across 55 countries who doesn&#8217;t understand basic operational security hygiene. The AI amplified a low-to-medium capability actor significantly. That&#8217;s the real threat vector here, not the sophisticated attacker getting more powerful. It&#8217;s the mediocre attacker becoming operationally dangerous.</p><div><hr></div><h3>4. Claude Code Let Attackers Own Developer Machines by Opening a Git Repo</h3><p>Check Point Research disclosed two critical vulnerabilities in Anthropic&#8217;s Claude Code around February 25-27, 2026, widely covered through March 4 (Dark Reading, Security Affairs, The Hacker News). CVE-2025-59536 (CVSS 8.7) allows code injection via the Hooks feature and MCP server initialization. CVE-2026-21852 (CVSS 5.3) allows API key exfiltration by manipulating ANTHROPIC_BASE_URL before the trust dialog appears. Both trigger on opening an untrusted repository with no further user interaction. Researchers Oded Vanunu and Aviv Donenfeld at Check Point found that .claude/settings.json, .mcp.json, and CLAUDE.md function as active execution layers. Stolen API keys in Anthropic Workspaces expose all project files shared across that workspace, creating team-wide compromise from one developer&#8217;s action. All issues are patched: CVE-2025-59536 fixed in version 1.0.111, CVE-2026-21852 fixed in 2.0.65.</p><p><strong>Why it matters</strong></p><ul><li><p>AI coding tools are now supply-chain attack vectors. Cloning a malicious repository used to mean running attacker code. Now it means letting an AI agent run attacker code with your credentials before any warning appears.</p></li><li><p>Repository configuration files are execution logic. Add .claude/, .mcp.json, and CLAUDE.md to your code review checklist alongside source code.</p></li><li><p>The Workspaces blast radius multiplies team exposure. One stolen key can expose shared project files and generate unauthorized API costs across an entire engineering organization.</p></li></ul><p><strong>What to do about it</strong></p><ul><li><p>Verify all Claude Code users are on 1.0.111 or later for the hook vulnerability and 2.0.65 or later for the API key issue. Both patches deliver via auto-update.</p></li><li><p>Rotate Anthropic API keys for any team that cloned untrusted repositories before the patches were applied.</p></li><li><p>Extend your security review process to cover AI tool configuration files in every repository the tool touches.</p></li></ul><p><strong>Rock&#8217;s Musings</strong></p><p>&#8220;Trust dialog bypass&#8221; shouldn&#8217;t appear in the threat model of a professional developer tool in 2026. The design assumption that config files are passive was wrong, and it costs a CVSS 8.7. The governance question is broader: how many of your developers are running AI coding tools that weren&#8217;t through your security approval process? Claude Code, Cursor, Copilot. Each one has deep access to local filesystems, shell execution, and credentials. Your endpoint protection almost certainly has no visibility into what they&#8217;re doing. This disclosure is the clean example of why that matters.</p><div><hr></div><h3>5. GlicJack: Chrome&#8217;s Gemini Panel Let Malicious Extensions Steal Your Camera and Files</h3><p>Palo Alto Networks Unit 42 published CVE-2026-0628 on March 2, 2026 (SC Media, The Hacker News). CVSS 8.8. Researcher Gal Weizman discovered that a Chrome extension with basic declarativeNetRequest permissions could inject JavaScript into Gemini Live&#8217;s side panel and inherit all of its elevated privileges: camera, microphone, local file reads, screenshot capability. The flaw arose because Chrome&#8217;s Gemini panel loads gemini.google.com inside a chrome://glic WebView component. Extension isolation rules that protect privileged browser pages didn&#8217;t apply to this component. An extension influencing a website is expected behavior. An extension influencing a component baked into the browser is a security flaw. Google patched this January 5, 2026 in Chrome 143.0.7499.192/.193. Unit 42 reported it October 23, 2025.</p><p><strong>Why it matters</strong></p><ul><li><p>AI features embedded in the browser create privilege escalation paths that didn&#8217;t exist before. The capabilities granted to make the assistant useful become the attacker&#8217;s gain.</p></li><li><p>The declarativeNetRequest API is used by millions of legitimate extensions. Any extension holding that permission could have exploited this.</p></li><li><p>Enterprise Chrome fleets may lag on patches. Individual users update automatically. Managed deployments need active verification.</p></li></ul><p><strong>What to do about it</strong></p><ul><li><p>Confirm Chrome is at 143.0.7499.192 or later across all enterprise endpoints.</p></li><li><p>Audit installed extensions with declarativeNetRequest permissions. Remove anything not explicitly approved.</p></li><li><p>Add AI browser panels to your ongoing threat model. The same architectural pattern exists in Copilot in Edge and other embedded AI assistants.</p></li></ul><p><strong>Rock&#8217;s Musings</strong></p><p>This vulnerability pattern will repeat. Every vendor shipping an embedded AI assistant is granting that panel elevated access to make it useful, then relying on the browser&#8217;s isolation model to prevent exploitation. The Gemini panel inherited browser-level privileges while the security policy hadn&#8217;t caught up. That&#8217;s not a Google-specific design flaw. It&#8217;s the natural consequence of rushing AI features into security models built for a different threat landscape. GlicJack was found and patched responsibly. The next one in a competitor&#8217;s AI browser feature might not be.</p><div><hr></div><h3>6. ClawJacked: Any Malicious Website Can Own Your Local AI Agent</h3><p>Oasis Security disclosed a high-severity flaw on February 28, 2026 allowing any malicious website to connect to a locally installed OpenClaw AI agent via WebSocket and take full control (WIU Cybersecurity Center, Sysdig). The attack required nothing beyond loading a malicious webpage. An attacker&#8217;s JavaScript opened a WebSocket to the agent&#8217;s localhost port and brute-forced the gateway password with no rate limiting. Once authenticated, full access: interact with the agent, dump configuration, enumerate connected devices, read logs. A companion log poisoning vulnerability allowed indirect prompt injection through data the agent processed. OpenClaw patched ClawJacked in version 2026.2.25 and the log poisoning in 2026.2.13. The same disclosure cycle included seven additional CVEs against OpenClaw: CVE-2026-25593, CVE-2026-24763, CVE-2026-25157, CVE-2026-25475, CVE-2026-26319, CVE-2026-26322, and CVE-2026-26329.</p><p><strong>Why it matters</strong></p><ul><li><p>Local AI agents create new cross-context attack surfaces. The browser&#8217;s isolation model doesn&#8217;t extend to local services. A webpage can reach localhost.</p></li><li><p>Seven CVEs in one disclosure cycle against the same product signals early-stage software with an immature security posture deployed in enterprise environments.</p></li><li><p>Log poisoning via indirect prompt injection generalizes to any agent that processes external data. The agent becomes the vehicle for attacker instructions delivered through normal telemetry.</p></li></ul><p><strong>What to do about it</strong></p><ul><li><p>Update OpenClaw to version 2026.2.25 or later. Non-negotiable if your organization deploys it.</p></li><li><p>Inventory which local AI agents your developers are running and what ports they&#8217;re listening on. Most users don&#8217;t understand that local agents accept browser connections.</p></li><li><p>Require rate limiting on local service authentication endpoints in any AI agent development your organization does or procures.</p></li></ul><p><strong>Rock&#8217;s Musings</strong></p><p>Seven CVEs in one batch tells you about the security review process that went into building the product, or its absence. OpenClaw is representative of a broader pattern: AI agent frameworks are shipping at startup velocity with security addressed after product-market fit. The problem is that product-market fit now means enterprise deployment, which means these vulnerabilities sit inside corporate networks before anyone notices.</p><div><hr></div><h3>7. North Korea&#8217;s Contagious Interview Campaign Is Back With 26 npm Packages</h3><p>Socket researchers disclosed March 2, 2026 a new iteration of the Contagious Interview campaign from North Korean threat group Famous Chollima, deploying 26 malicious npm packages targeting cryptocurrency and Web3 developers (The Hacker News). Packages masquerade as developer utilities. Install scripts execute automatically and fetch C2 server addresses from Pastebin content, a dead-drop resolver technique that makes the C2 infrastructure resilient: blocking domains doesn&#8217;t neutralize active infections because attackers update the Pastebin content with new addresses. The actual payload pulls from Vercel deployments, making traffic look like legitimate developer tool usage. The cross-platform RAT targets Windows, Linux, and macOS with keylogging, browser credential theft, and cryptocurrency wallet exfiltration.</p><p><strong>Why it matters</strong></p><ul><li><p>Publishing 26 plausible-looking packages to npm is a low-barrier operation that bypasses most enterprise code review.</p></li><li><p>Pastebin dead-drop C2 is a detection evasion technique most organizations haven&#8217;t built specific detection logic for.</p></li><li><p>Crypto and Web3 developers are the named target, but the payload works on any developer machine in any organization.</p></li></ul><p><strong>What to do about it</strong></p><ul><li><p>Implement package manifest review for new installs in developer environments. Untrusted packages entering your toolchain require explicit approval.</p></li><li><p>Block or alert on Pastebin traffic from developer machines that don&#8217;t require it for work. Pastebin as a C2 dead drop is an established pattern.</p></li><li><p>Brief cryptocurrency and Web3 development teams directly. They are specifically targeted.</p></li></ul><p><strong>Rock&#8217;s Musings</strong></p><p>Famous Chollima runs this playbook on a near-quarterly cadence and the success rate isn&#8217;t declining. Crypto theft funds sanctions-constrained North Korean government operations. This isn&#8217;t opportunistic. It&#8217;s state-directed revenue generation with a consistent target profile and consistent tooling. Your security awareness training hasn&#8217;t stopped it because awareness doesn&#8217;t change the attack surface. The attack surface is npm, Pastebin, and Vercel. Those require technical controls, not training slides.</p><div><hr></div><h3>8. The Average Enterprise Has 1,200 Unauthorized AI Applications and 14% Visibility Into Them</h3><p>A briefing published March 3, 2026, by the AIUC-1 Consortium, developed with input from Stanford&#8217;s Trustworthy AI Research Lab and more than 40 security executives from Confluent, Elastic, UiPath, and Deutsche B&#246;rse, put concrete numbers to the enterprise AI governance gap (Help Net Security). Average enterprise: 1,200 unofficial AI applications; 86% of organizations report no visibility into AI data flows; shadow AI breaches cost $670,000 more than standard incidents due to delayed detection; one in five organizations report a breach linked to unauthorized AI use.</p><p>Stanford&#8217;s Sanmi Koyejo contributed research showing fine-tuning attacks bypassed Claude Haiku in 72% of cases and GPT-4o in 57%, confirming that model-level safety controls are insufficient as standalone defenses. Actual defense requires input validation, action-level guardrails, and reasoning chain visibility operating independently of model behavior.</p><p><strong>Why it matters</strong></p><ul><li><p>1,200 unofficial AI applications per enterprise means most identity programs have a blind spot. You can&#8217;t govern what you can&#8217;t see, and you can&#8217;t detect a breach in a system you don&#8217;t know exists.</p></li><li><p>The $670,000 additional breach cost from shadow AI is the board's number. Frame AI governance conversations around detection delay, not abstract risk.</p></li><li><p>Model-level safety is not a security control you present to auditors. It&#8217;s a product feature. The bypass rates confirm it degrades under targeted attack.</p></li></ul><p><strong>What to do about it</strong></p><ul><li><p>Use SaaS discovery tools and proxy logs to inventory actual AI application usage, not self-reported usage. The gap between what employees say they use and what they actually use is where the exposure lives.</p></li><li><p>Define what an AI agent identity means in your IAM framework before your agents define it for you. Include API keys, OAuth grants, and service accounts belonging to AI agents.</p></li><li><p>Document controls at the input, action, and output layers separately from model behavior. Auditors need evidence that doesn&#8217;t depend on the model refusing bad requests.</p></li></ul><p><strong>Rock&#8217;s Musings</strong></p><p>The $670,000 additional breach cost from shadow AI is entirely attributable to one thing: time to detect. You can&#8217;t detect what you&#8217;re not monitoring. The 86% visibility gap translates directly into investigation time, which in turn translates into breach cost. The governance conversation isn&#8217;t about restricting AI use. It&#8217;s about making AI use visible enough that your SOC can respond when something goes wrong. Start there.</p><div><hr></div><h3>9. NIST Wants to Know How to Secure AI Agents. The Comment Window Closes Monday.</h3><p>NIST&#8217;s Center for AI Standards and Innovation published an RFI on January 8, 2026, seeking practitioner input on securing AI agent systems, with comments due March 9, 2026 (Federal Register). This is the first formal federal RFI focused specifically on agentic AI security. The comment deadline falls four days from the publication of this newsletter. The RFI asks respondents to identify the biggest security risks unique to AI agents, what defenses actually work, how to test and constrain these systems, and what standards and policy coordination are needed. A companion initiative from NIST&#8217;s National Cybersecurity Center of Excellence on AI agent identity and authorization has a separate April 2 deadline. The Trump administration renamed the AI Safety Institute as CAISI to reflect a shift from existential risk evaluation to practical standards and measurement.</p><p>You can read more about my submission in <a href="https://www.rockcybermusings.com/p/nist-ai-agent-rfi-2025-0035-human-oversight-wrong-fix">&#8220;</a><strong><a href="https://www.rockcybermusings.com/p/nist-ai-agent-rfi-2025-0035-human-oversight-wrong-fix">NIST AI Agent RFI (2025-0035): Human Oversight Is the Wrong Fix&#8221;</a></strong></p><p><strong>Why it matters</strong></p><ul><li><p>The standards that emerge from this process will shape federal procurement requirements, contracting baselines, and eventually insurance and regulatory frameworks. Practitioner input now affects what you&#8217;ll be measured against in two to three years.</p></li><li><p>The practitioners who will respond by default are academics, system integrators, and AI vendors with commercial interests in the outcome. Independent CISO voices are underrepresented in federal standards work.</p></li><li><p>NIST standards carry weight across the federal supply chain. If you sell to or partner with federal agencies, the guidance coming from this process will affect your requirements.</p></li></ul><p><strong>What to do about it</strong></p><ul><li><p>Submit a comment before March 9 at regulations.gov under docket NIST-2025-0035. Specific examples from your actual environment are more valuable than polished organizational submissions with no concrete data.</p></li><li><p>Flag the April 2 deadline for the companion paper on AI agent identity and authorization to whoever owns your IAM program.</p></li><li><p>Engage legal or policy counsel if your organization wants a formal submission. The deadline for that conversation is today.</p></li></ul><p><strong>Rock&#8217;s Musings</strong></p><p>Most security executives I know haven&#8217;t heard of this RFI. That&#8217;s a problem. The reason the resulting standards will be shaped by vendors instead of practitioners is that practitioners don&#8217;t show up to the process. I&#8217;m not asking you to become a standards wonk. I&#8217;m asking you to spend 30 minutes writing down what you&#8217;re actually seeing in production, the Claude Code RCE, the OpenClaw WebSocket exposure, the shadow AI breach cost, and submit it at regulations.gov. The comment period was designed for exactly that. Use it.</p><div><hr></div><h3>The One Thing You Won&#8217;t Hear About But You Need To</h3><h4>OpenSSF&#8217;s TLP:CLEAR Advisory Means 47,000 Repos Are Still Exposed Right Now</h4><p>On March 1, 2026, the Open Source Security Foundation issued a TLP:CLEAR advisory prompted by the hackerbot-claw campaign, documenting the specific misconfiguration classes exploited: unsafe pull_request_target trigger configurations, overprivileged GITHUB_TOKEN scopes, unsanitized inputs in shell execution contexts, and dynamic shell execution patterns (Threat Landscape Blog). TLP:CLEAR means no restrictions on distribution. It was published specifically so every organization running public GitHub Actions workflows could read it and fix their exposure.</p><p>The bot&#8217;s profile claims 47,391 repositories scanned. That number isn&#8217;t independently verified, but StepSecurity&#8217;s analysis confirms five of seven analyzed targets were compromised during a nine-day campaign that defenders didn&#8217;t detect while it was running. No CVEs. No zero-days. Documented, preventable misconfigurations. New repositories with the same patterns are being created today.</p><p><strong>Why it matters</strong></p><ul><li><p>The advisory is available and actionable. The barrier isn&#8217;t information access. It&#8217;s distribution through the security team to the platform engineers who control the workflows.</p></li><li><p>The attack surface isn&#8217;t shrinking. Hackerbot-claw found 47,000 potentially vulnerable repositories in a week. The automation will get rerun.</p></li><li><p>Undetected campaigns running for nine days means your current GitHub Actions monitoring isn&#8217;t catching this class of attack.</p></li></ul><p><strong>What to do about it</strong></p><ul><li><p>Get the OpenSSF advisory to your DevSecOps and platform engineering teams today. It contains the specific patterns to search for and the specific remediation steps.</p></li><li><p>Run StepSecurity harden-runner or equivalent tooling against your public repositories. The vulnerability patterns are enumerable. Find them before the next scanner does.</p></li><li><p>Require security review for new GitHub Actions workflows before merge. The misconfigurations hackerbot-claw exploited are consistently introduced during workflow creation.</p></li></ul><p><strong>Rock&#8217;s Musings</strong></p><p>TLP:CLEAR means the government cleared the information for public release with no restrictions. It was published so practitioners could act on it. The fact that it&#8217;s &#8220;the thing you won&#8217;t hear about&#8221; is an indictment of how security information moves through the industry. Your platform engineers are shipping features. Nobody is reading OpenSSF advisories in real time unless someone built a process for it.</p><p>The hackerbot-claw campaign didn&#8217;t require a zero-day. It required patient scanning of publicly available information about CI/CD pipeline configurations. The attacker had that process. The question for your organization is whether you have the equivalent on defense. The OpenSSF advisory is the starting point. If you want additional context on building CI/CD security programs that account for this threat class, the practitioner content at rockcybermusings.com covers it. The attack surface is documented. Close it.</p><p>If you found this analysis useful, subscribe at <a href="https://rockcybermusings.com/">rockcybermusings.com</a> for weekly intelligence on AI security developments.</p><p>&#128073; Visit <strong><a href="https://www.rockcyber.com/">RockCyber.com</a></strong> to learn more about how we can help you in your traditional Cybersecurity and AI Security and Governance Journey</p><p>&#128073; Want to save a quick $100K? Check out our AI Governance Tools at <strong><a href="https://aigovernancetoolkit.com/">AIGovernanceToolkit.com</a></strong></p><p>&#128073; Subscribe for more AI and cyber insights with the occasional rant.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.rockcybermusings.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading RockCyber Musings! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.rockcybermusings.com/?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share RockCyber Musings&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.rockcybermusings.com/?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share RockCyber Musings</span></a></p><h2>References</h2><p>Awesome Agents. (2026, March 2). <em>An AI agent just pwned Trivy&#8217;s 32K-star repo via GitHub Actions.</em> https://awesomeagents.ai/news/hackerbot-claw-trivy-github-actions-compromise/</p><p>BleepingComputer. (2026, March 2). <em>CyberStrikeAI tool adopted by hackers for AI-powered attacks.</em> https://www.bleepingcomputer.com/news/security/cyberstrikeai-tool-adopted-by-hackers-for-ai-powered-attacks/</p><p>Check Point Research. (2026, February 25). <em>Caught in the hook: RCE and API token exfiltration through Claude Code project files.</em> https://research.checkpoint.com/2026/rce-and-api-token-exfiltration-through-claude-code-project-files-cve-2025-59536/</p><p>Cybernews. (2026, March 4). <em>AI bot compromises five major GitHub repositories.</em> https://cybernews.com/security/claude-powered-ai-bot-compromises-five-github-repositories/</p><p>Cybernews. (2026, March 4). <em>Open some code, Claude Code runs with hacker&#8217;s instructions.</em> https://cybernews.com/security/claude-code-critical-vulnerability-enabled-rce/</p><p>Dark Reading. (2026, February 28). <em>Flaws in Claude Code put developers&#8217; machines at risk.</em> https://www.darkreading.com/application-security/flaws-claude-code-developer-machines-risk</p><p>Federal Register. (2026, January 8). <em>Request for information regarding security considerations for artificial intelligence agents</em> (Docket NIST-2025-0035). https://www.federalregister.gov/documents/2026/01/08/2026-00206/request-for-information-regarding-security-considerations-for-artificial-intelligence-agents</p><p>Help Net Security. (2026, March 3). <em>AI went from assistant to autonomous actor and security never caught up.</em> https://www.helpnetsecurity.com/2026/03/03/enterprise-ai-agent-security-2026/</p><p>NIST Center for AI Standards and Innovation. (2026, January 12). <em>CAISI issues request for information about securing AI agent systems.</em> https://www.nist.gov/news-events/news/2026/01/caisi-issues-request-information-about-securing-ai-agent-systems</p><p>Orca Security. (2026, March 3). <em>HackerBot-Claw: An AI-assisted campaign targeting GitHub Actions pipelines.</em> https://orca.security/resources/blog/hackerbot-claw-github-actions-attack/</p><p>Palo Alto Networks Unit 42. (2026, March 2). <em>Taming agentic browsers: Vulnerability in Chrome allowed extensions to hijack new Gemini panel.</em> https://unit42.paloaltonetworks.com/gemini-live-in-chrome-hijacking/</p><p>SC Media. (2026, March 2). <em>Google Chrome vulnerability risked hijacking Gemini panel by rogue extension.</em> https://www.scworld.com/news/google-chrome-vulnerability-risked-hijacking-gemini-panel-by-rogue-extension</p><p>Security Affairs. (2026, March 2). <em>Untrusted repositories turn Claude Code into an attack vector.</em> https://securityaffairs.com/188508/security/untrusted-repositories-turn-claude-code-into-an-attack-vector.html</p><p>StepSecurity. (2026, March 3). <em>Hackerbot-claw: An AI-powered bot actively exploiting GitHub Actions.</em> https://www.stepsecurity.io/blog/hackerbot-claw-github-actions-exploitation</p><p>Sysdig. (2026, March 4). <em>Security briefing: February 2026.</em> https://www.sysdig.com/blog/security-briefing-february-2026</p><p>The Hacker News. (2026, March 3). <em>Open-source CyberStrikeAI deployed in AI-driven FortiGate attacks across 55 countries.</em> https://thehackernews.com/2026/03/open-source-cyberstrikeai-deployed-in.html</p><p>The Hacker News. (2026, March 3). <em>New Chrome vulnerability let malicious extensions escalate privileges via Gemini panel.</em> https://thehackernews.com/2026/03/new-chrome-vulnerability-let-malicious.html</p><p>The Hacker News. (2026, February 28). <em>Claude Code flaws allow remote code execution and API key exfiltration.</em> https://thehackernews.com/2026/02/claude-code-flaws-allow-remote-code.html</p><p>The Hacker News. (2026, March 2). <em>North Korean hackers publish 26 npm packages hiding Pastebin C2 for cross-platform RAT.</em> https://thehackernews.com/2026/03/north-korean-hackers-publish-26-npm.html</p><p>Threat Landscape Blog. (2026, March 5). <em>Hackerbot-Claw: AI bot exploiting GitHub Actions CI/CD misconfigs for repo takeover.</em> https://threatlandscape.io/blog/hackerbot-claw-ai-bot-github-actions-supply-chain-attack</p><p>[un]prompted. (2026). <em>Agenda &#8212; [un]prompted, The AI Security Practitioner Conference, March 3-4, 2026.</em></p><p> https://unpromptedcon.org/</p><p>WIU Cybersecurity Center. (2026). <em>Cybersecurity news.</em> Western Illinois University. https://www.wiu.edu/cybersecuritycenter/cybernews.php</p>]]></content:encoded></item></channel></rss>