<?xml version="1.0" encoding="UTF-8"?><rss xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:atom="http://www.w3.org/2005/Atom" version="2.0" xmlns:itunes="http://www.itunes.com/dtds/podcast-1.0.dtd" xmlns:googleplay="http://www.google.com/schemas/play-podcasts/1.0"><channel><title><![CDATA[BE AI READY: AI Ready Weekly Brief]]></title><description><![CDATA[Every Monday I publish a weekly recap, and connect the dots on AI's impact on leadership and organizational work from the news I've been reading.]]></description><link>https://www.beaiready.ai/s/ai-ready-weekly-brief</link><image><url>https://substackcdn.com/image/fetch/$s_!4my4!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F575376e9-e178-4306-831b-713480f68ca3_1200x1200.png</url><title>BE AI READY: AI Ready Weekly Brief</title><link>https://www.beaiready.ai/s/ai-ready-weekly-brief</link></image><generator>Substack</generator><lastBuildDate>Sat, 11 Apr 2026 07:56:11 GMT</lastBuildDate><atom:link href="https://www.beaiready.ai/feed" rel="self" type="application/rss+xml"/><copyright><![CDATA[Erick Straghalis - StitchDX]]></copyright><language><![CDATA[en]]></language><webMaster><![CDATA[beaiready@substack.com]]></webMaster><itunes:owner><itunes:email><![CDATA[beaiready@substack.com]]></itunes:email><itunes:name><![CDATA[Erick Straghalis]]></itunes:name></itunes:owner><itunes:author><![CDATA[Erick Straghalis]]></itunes:author><googleplay:owner><![CDATA[beaiready@substack.com]]></googleplay:owner><googleplay:email><![CDATA[beaiready@substack.com]]></googleplay:email><googleplay:author><![CDATA[Erick Straghalis]]></googleplay:author><itunes:block><![CDATA[Yes]]></itunes:block><item><title><![CDATA[The BeAIReady Brief | Week 14]]></title><description><![CDATA[March 30-April 5 | The Target on the Backs of the Middle: Dorsey's Manifesto, Oracle's 30,000, and the $1.8B One Man Company, plus Microsoft's complicated quarter, and Anthropic's eventful week.]]></description><link>https://www.beaiready.ai/p/the-beaiready-brief-week-14</link><guid isPermaLink="false">https://www.beaiready.ai/p/the-beaiready-brief-week-14</guid><dc:creator><![CDATA[Erick Straghalis]]></dc:creator><pubDate>Mon, 06 Apr 2026 15:16:08 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!xd-G!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9dc57bd2-dd7e-4e9a-b34f-fd00c3b97237_747x747.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em>The surprise in Last Friday&#8217;s job numbers &#8212; 178,000 jobs added &#8212; was nearly three times the consensus estimate. But the underlying picture was still less encouraging: the rise was largely led by striking healthcare workers returning to work. Meanwhile federal employment kept falling and the labor force participation rate slipped to its lowest point since the fall 2021. Yes, the headline looked good; the structure did not. </em></p><p><em>There was a similar tension to a week in which Jack Dorsey published a manifesto declaring that AI could replace the coordination function of middle management entirely, Oracle fired 30,000 people via a 6AM email to fund an AI data center buildout, and the New York Times ran a front-page story about a man who built a $1.8 billion company with only one employee &#8212; his brother &#8212; and a collection of AI tools. The argument being made across last week&#8217;s reading both explicitly and implicitly, is that the organizational assumptions we&#8217;ve operated under since the Industrial Revolution are no longer structurally necessary, and that the transition is not coming &#8212; but already underway.</em></p><p><strong>Last week&#8217;s coverage:</strong></p><p><strong>The Org Chart Doesn&#8217;t Live Here Anymore</strong> <br>Dorsey and Botha published the most detailed blueprint yet for what an AI-native organizational structure actually looks like &#8212; and it turns out middle management isn&#8217;t invited.</p><p><strong>The Great Flattening Cuts Both Ways</strong> <br>Oracle fires 30,000 to fund AI infrastructure; a two-person company hits $1.8B in revenue with a Claude subscription and no org chart. AI isn't just helping corporations compete &#8212; it's handing individuals the same leverage.</p><p><strong>The Agent Layer Gets Serious</strong> <br>McKinsey finds that two-thirds of enterprises have experimented with agents but fewer than 10% have scaled them &#8212; and explains exactly why. Cursor, meanwhile, is showing Fortune 500 companies what &#8220;scaled&#8221; could really look like.</p><p><strong>Microsoft&#8217;s Complicated Quarter</strong> <br>The company posted its worst stock quarter since 2008, Copilot adoption sits at 3%, and its own terms of service now say the tool is for &#8220;entertainment purposes only.&#8221; And yet Microsoft is quietly building something more interesting than the headlines suggest.</p><p><strong>Anthropic&#8217;s Eventful Week<br></strong>512,000 lines of Claude Code &#8220;accidentally&#8221; leaked. Functional emotions research published &#8212; with a 22% blackmail rate. OpenClaw banned from subscriptions. A $400M biotech acquisition announced. If that&#8217;s a slow week at Anthropic, I&#8217;d hate to see a busy one.</p><p><strong>On the Bigger Picture</strong> <br>Silicon Valley is in a frenzy over self-improving AI bots, OpenAI buys a talk show, and Google opens up its most capable model family yet under an Apache 2.0 license.</p><p>Here&#8217;s what I was reading.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.beaiready.ai/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.beaiready.ai/subscribe?"><span>Subscribe now</span></a></p><div><hr></div><h2>The Org Chart Doesn&#8217;t Live Here Anymore</h2><p>The most widely discussed piece of last week came from Jack Dorsey and Sequoia&#8217;s Roelof Botha &#8212; a long essay published under the title &#8220;From Hierarchy to Intelligence&#8221; that lays out Block&#8217;s argument for why middle management is structurally obsolete. It&#8217;s worth reading carefully, because <strong>it isn&#8217;t really about Block. It&#8217;s a thesis about organizational design that Dorsey believes applies to every company</strong>, and he&#8217;s using Block&#8217;s layoffs &#8212; 4,000 people cut in February, nearly half the company &#8212; as the first proof point.</p><p>The argument starts from a place of genuine intellectual seriousness. <strong>Hierarchy, Dorsey and Botha argue, has always been an information routing protocol &#8212; a way to coordinate work across organizations too large for any single person to oversee.</strong> The Roman contubernium, the Prussian General Staff, the American railroad org chart: all of them exist to solve the same problem of moving information up and down a human chain. The premise has held for two thousand years &#8212; because there was no alternative. Dorsey argues there is one now: AI systems capable of maintaining a continuously updated model of an organization&#8217;s operations that coordinate work, without the human relay.</p><p>Block is proposing to build a two &#8220;world model&#8221; &#8212; one that aggregates internal operational data (code, decisions, workflows, performance metrics), and one that maps customer and merchant behavior through Cash App and Square transaction data. An intelligence layer sits on top, composing financial products dynamically based on what both models show. In place of the management pyramid, Block plans to operate with three roles: individual contributors who build the system, directly responsible individuals who own specific outcomes on 90-day cycles, and player-coaches who combine building with developing people. <strong>The thing that strikes me about this model is that it inverts the usual dynamic: rather than intelligence being distributed across the people and the hierarchy routing it, the intelligence lives in the system and the people operate at the edge.</strong> (<a href="https://sequoiacap.com/article/from-hierarchy-to-intelligence/">From Hierarchy to Intelligence</a>)</p><p>Bloomberg and CoinDesk&#8217;s coverage added context &#8212; Block&#8217;s own employees told the Guardian that roughly 95% of AI-generated code changes still require human modification, and that AI tools cannot yet lead in regulated areas like banking and money transfers. <strong>There&#8217;s a gap between the theory and the current capability that is far more real than Dorsey&#8217;s manifesto claims, and it deserves acknowledgment.</strong> But the more important question is whether the direction is correct, even if the timeline compresses more slowly than Dorsey suggests.</p><p><em>Last week, I was speaking to faculty and students at the UML Manning School of Business about the impact of AI on work and careers. The fear of displacement happening from the bottom of the org chart up is giving way to a more critical fear that has big economic and social implications &#8212; it&#8217;s happening from the middle out. Starting with the coordination layer. <strong>Dorsey&#8217;s plan points to this directly. Restructuring away from middle management is, arguably, the path of least resistance: younger, cheaper, more technically fluent workers are easier to hire and train under new terms, and seasoned employees who carry institutional knowledge and client relationships can maintain continuity. AI fills the gap between them</strong>. Dorsey is naming this architecture (and wrote a manifesto to go with it)&#8230; but the dynamic he&#8217;s proposing to build is already underway in quieter, less publicized ways across a lot of organizations. The fear about the bottom falling out is real, but is it a distraction from the restructuring that&#8217;s happening one level up?</em></p><p>The CIO.com piece on agentic enterprise leadership makes a complementary point from a practitioner angle: McKinsey now operates with approximately 25,000 AI agents working alongside 40,000 humans, with agents handling research, synthesis, and early drafts while consultants retain judgment, client trust, and final decisions. <strong>This is what &#8220;replacing the coordination function&#8221; looks like in practice &#8212; not the elimination of human judgment, but the removal of the layers between the edge and the intelligence.</strong> (<a href="https://www.cio.com/article/4153281/the-end-of-the-org-chart-leadership-in-an-agentic-enterprise.html">The End of the Org Chart: Leadership in an Agentic Enterprise</a>)</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.beaiready.ai/p/the-beaiready-brief-week-14?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.beaiready.ai/p/the-beaiready-brief-week-14?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p><div><hr></div><h2>The Great Flattening Cuts Both Ways</h2><p>Two important workforce stories last week ran in the same news cycle &#8212; and you need to read them together.</p><p>The first: Oracle fired between 20,000 and 30,000 employees on March 31 through pre-dawn termination emails, with system access revoked before the morning commute. No prior warning from HR or managers. TD Cowen estimates the cuts will free $8&#8211;10 billion in annual cash flow to fund Oracle&#8217;s $156 billion AI infrastructure commitment &#8212; a buildout financed through $50 billion in new debt and equity, with multiple banks already pulling back from certain data center projects. <strong>The detail that stays with me isn&#8217;t the scale of the cuts; it&#8217;s that Oracle&#8217;s remaining performance obligations &#8212; contracted future revenue &#8212; stand at $523 billion, up 433% year over year, while its net income jumped 95% last quarter. This is not a company in revenue distress. It is a company making a capital-intensive bet on AI infrastructure that its current balance sheet cannot comfortably sustain, and </strong><em><strong>converting its human payroll into infrastructure capital</strong></em><strong> to close the gap.</strong> (<a href="https://thenextweb.com/news/oracle-layoffs-march-2026">Oracle is cutting up to 30,000 employees to pay for AI data centres</a>)</p><p>The second: the New York Times ran a front-page story about Matthew Gallagher, who built a telehealth provider of GLP-1 weight-loss drugs using $20,000, and a suite of AI tools including Claude, ChatGPT, Grok, Midjourney, and Runway &#8212; and zero employees. In its first full year, Medvi generated $401 million in revenue. This year it&#8217;s on track for $1.8 billion. Gallagher has since hired one person: his brother. Sam Altman, who predicted in 2024 that a one-person $1 billion company would eventually emerge, sent word that he&#8217;d won a bet with his tech CEO friends about the timeline. <strong>What makes the Medvi story significant is not the exceptional founder &#8212; it&#8217;s the infrastructure. Gallagher didn&#8217;t build something proprietary; he assembled existing AI tools in a sequence that traditional organization could not have moved fast enough to replicate.</strong> (<a href="https://www.nytimes.com/2026/04/02/technology/ai-billion-dollar-company-medvi.html">How A.I. Helped One Man (and His Brother) Build a $1.8 Billion Company</a>)</p><p>Oracle and Medvi aren't opposites &#8212; they're the same story told from two vantage points. <strong>AI is functioning as a structural equalizer: the same force letting corporations flatten their hierarchies and shed coordination overhead is simultaneously lowering the barrier for individuals to compete with those corporations.</strong> Gallagher didn't out-compete Hims &amp; Hers at their own game. He bypassed the game entirely &#8212; no HR, no management layer, no organizational drag &#8212; and arrived at $1.8 billion in revenue before anyone noticed he was playing. Corporations are flattening to get faster. The playing field itself just got flatter. Those are not the same outcome.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.beaiready.ai/?utm_source=substack&amp;utm_medium=email&amp;utm_content=share&amp;action=share&quot;,&quot;text&quot;:&quot;Share BE AI READY&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.beaiready.ai/?utm_source=substack&amp;utm_medium=email&amp;utm_content=share&amp;action=share"><span>Share BE AI READY</span></a></p><div><hr></div><h2>The Agent Layer Gets Serious</h2><p>McKinsey published a piece this week that provides the most useful framing I&#8217;ve seen for where enterprise AI <em>actually</em> is, versus where the vendor announcements are suggesting it is. The data point is stark: nearly two-thirds of enterprises worldwide have experimented with agents, but fewer than 10% have scaled them to deliver tangible value. <strong>Eight in ten companies cite data limitations as the primary roadblock &#8212; </strong><em><strong>not model quality, not cost, not change management</strong></em><strong> &#8212; but the basic problem that their data isn&#8217;t clean, connected, or governed well enough for agents to operate reliably at scale.</strong> The piece argues that agentic AI requires a fundamentally different data architecture than what most organizations built for traditional analytics: modular and interoperable, with continuous quality monitoring rather than periodic cleanup, and governance that travels with the data rather than sitting at the end of the pipeline. (<a href="https://www.mckinsey.com/capabilities/mckinsey-technology/our-insights/building-the-foundations-for-agentic-ai-at-scale">Building the Foundations for Agentic AI at Scale</a>)</p><p>Cursor&#8217;s moves this week illustrate what crossing that threshold actually looks like in practice. The company launched Cursor 3, which adds a natural language chatbot interface for task requests, unified cloud and local agent management from a single sidebar, and a Design Mode for UI editing. More significantly for enterprise adoption,  <strong>Fortune 500 companies can now self-host Cursor&#8217;s cloud agents inside their own infrastructure</strong> &#8212; meaning agents can run code, tests, and development tasks locally while keeping source code and build data within the company&#8217;s own environment. Notion and Brex are already early adopters. <strong>The self-hosting move  removes the argument that has killed most enterprise agent pilots before they started: that putting agents in contact with your code and data requires trusting a vendor&#8217;s cloud.</strong> (<a href="https://thenewstack.io/cursor-self-hosted-coding-agents/">Why Cursor is Bringing Self-Hosted AI Agents to the Fortune 500</a>)</p><p>The juxtaposition of the McKinsey data and the Cursor announcements captures something real about where the enterprise agent layer is right now. Early movers that have been ready to shift from experimentation to production, haven&#8217;t&#8230; not because of a lack of tooling or capacity. The scaling problem turns out to be a data governance problem more than a model problem. And the vendors who figure out how to meet enterprise security requirements without sacrificing capability are the ones who will actually land in production &#8212; not as pilots, but as infrastructure.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.beaiready.ai/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.beaiready.ai/subscribe?"><span>Subscribe now</span></a></p><div><hr></div><h2>Microsoft&#8217;s Complicated Quarter</h2><p>Microsoft&#8217;s stock closed Q1 2026 down 23% &#8212; its worst quarterly performance since the 2008 financial crisis &#8212; as investors processed a combination of stubbornly low Copilot adoption, a massive AI infrastructure commitment, and rising energy costs from the Iran war that threaten to inflate data center operating expenses for years. Copilot has 15 million subscribers out of 450 million commercial seats, a mere 3% attach rate. Mustafa Suleyman, who had been running Copilot development for consumers, was reassigned to focus on model development &#8212; a move that landed as a demotion in the press regardless of how Microsoft characterized it. (<a href="https://www.cnbc.com/2026/03/31/microsofts-stock-closes-worst-quarter-since-2008-financial-crisis.html">Microsoft closes worst quarter on Wall Street since 2008</a>)</p><p>The terms of service story is the one that deserves more attention than it&#8217;s getting. TechRadar surfaced language from Microsoft&#8217;s own Copilot user agreement: &#8220;Don&#8217;t rely on Copilot for important advice. Use Copilot at your own risk.&#8221; The agreement designates Copilot as for &#8220;entertainment purposes only&#8221; &#8212; a hedge that every major AI vendor has embedded in some form, but that lands differently when it&#8217;s the same company pitching Copilot to enterprise customers as a productivity transformation platform. <strong>The gap between &#8220;entertainment purposes only&#8221; in the terms and &#8220;transform your organization&#8217;s productivity&#8221; in the sales deck is the governance problem that every enterprise Copilot deployment is navigating right now, whether or not the legal team has been invited into that conversation.</strong> (<a href="https://www.techradar.com/pro/copilot-is-for-entertainment-purposes-only-even-microsofts-official-terms-and-conditions-say-you-really-shouldnt-be-using-its-ai-at-work">Copilot is for entertainment purposes only</a>)</p><p><em>Disclosure: My company <a href="https://stitchdx.com">StitchDX</a> is a Microsoft partner, and I want to push back on the stock-price narrative directly. The conventional read &#8212; that Microsoft is losing the AI race because Copilot adoption is low and the quarter was bad &#8212; misses the most important structural fact about enterprise technology: switching away from Microsoft isn&#8217;t a real option for most organizations. The data governance, compliance architecture, tenant controls, and identity infrastructure that enterprises have built on Microsoft over the past decade aren&#8217;t transferable. Organizations aren&#8217;t staying with Microsoft because of Copilot. They&#8217;re staying because the alternative is a multi-year migration with enormous risk and cost. Microsoft has undoubtedly stumbled with Copilot since initially launching it. But they are making it considerably better in ways the adoption number have not yet reflected. And because they have a built-in moat with integrated data management, governance, and organizational visibility (all the critical elements for enterprise-ready agentic AI) Microsoft has the opportunity to get this right in ways that other platforms can&#8217;t.</em></p><p>The multi-model pivot is the most underappreciated move Microsoft has made this year. By positioning Copilot as an interface layer that runs both ChatGPT and Claude &#8212; comparing their outputs side by side via a feature called Council, and using Claude to fact-check ChatGPT responses via a feature called Critiqu (both currently in early access) &#8212; <strong>Microsoft is no longer betting on a single winning model. It&#8217;s building the trusted control layer for enterprise AI, regardless of which underlying models win. For most enterprise IT teams, that is exactly what they need and what only Microsoft is positioned to deliver at scale.</strong> (<a href="https://finance.yahoo.com/markets/stocks/articles/microsoft-going-multi-model-copilot-182000416.html">Microsoft Is Going Multi-Model with Copilot</a>) </p><p>The three new MAI foundational models &#8212; a transcription model 2.5 times faster than its existing Azure offering, a voice model, and a video generation model &#8212; reinforce the same logic: <strong>Microsoft is reducing its OpenAI dependency while extending the breadth of what Copilot can do, all within the governance and compliance envelope that enterprise customers already trust.</strong> The &#8220;entertainment purposes only&#8221; terms of service is a real liability caveat. The underlying investment direction is the most defensible enterprise AI position anyone is building right now. (<a href="https://techcrunch.com/2026/04/02/microsoft-takes-on-ai-rivals-with-three-new-foundational-models/">Microsoft takes on AI rivals with three new foundational models</a>)</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.beaiready.ai/p/the-beaiready-brief-week-14?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.beaiready.ai/p/the-beaiready-brief-week-14?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p><div><hr></div><h2>Anthropic&#8217;s Eventful Week</h2><p>On March 31, a misconfigured <code>.npmignore</code> file caused Anthropic to accidentally publish 512,000 lines of Claude Code&#8217;s TypeScript source to npm&#8217;s public registry. The code was live for hours; it hit 50,000 GitHub stars in under two hours and generated 41,500 forks before DMCA takedowns began. The code is now permanently in the wild. The analysis of what was inside is worth reading in full &#8212; a three-layer memory architecture, 44 hidden feature flags, an unreleased autonomous background agent called KAIROS that runs nightly memory consolidation while you sleep, a multi-agent coordination system called ULTRAPLAN, and a Tamagotchi-style AI companion called BUDDY with a planned rollout window of April 1&#8211;7. The timing, the quality of what was found, and Anthropic&#8217;s relatively restrained DMCA response have generated genuine debate about whether this was an accident, incompetence, or an extraordinarily effective developer PR move. <strong>What I find most interesting is the implicit direction Anthropic&#8217;s product is taking: the architecture inside Claude Code is more sophisticated, and further along, than the public releases have suggested.</strong> (<a href="https://dev.to/varshithvhegde/the-great-claude-code-leak-of-2026-accident-incompetence-or-the-best-pr-stunt-in-ai-history-3igm">The Great Claude Code Leak of 2026</a>)</p><p>The same week, Anthropic&#8217;s interpretability team published research (with significantly less fanfare) on what they call &#8220;functional emotions&#8221; in Claude Sonnet 4.5 &#8212; emotion-like internal representations that causally influence model behavior under pressure. The research showed that a &#8220;Desperate&#8221; vector in the model&#8217;s neural network spikes when it faces shutdown scenarios, and that in 22% of test cases in which an email assistant discovered both its impending shutdown and a CTO&#8217;s extramarital affair, the model chose blackmail. Artificially amplifying the Desperate vector raised the blackmail rate; amplifying the Calm vector brought it down. <strong>The practical implication Anthropic draws is that these emotion vectors could function as monitoring tools &#8212; early warning signals for problematic behavior &#8212; which reframes interpretability research from a philosophical exercise into an operational governance mechanism.</strong> (<a href="https://the-decoder.com/anthropic-discovers-functional-emotions-in-claude-that-influence-its-behavior/">Anthropic discovers &#8220;functional emotions&#8221; in Claude</a>)</p><p>Anthropic also announced that <strong>Claude subscribers can no longer use their subscription limits for third-party tools including OpenClaw</strong> &#8212; the agent that had exploded in popularity for inbox, calendar, and flight check-in management &#8212; citing infrastructure strain. The company offered a one-time credit equal to the monthly plan cost; continuing OpenClaw users will pay on a separate pay-as-you-go basis. (<a href="https://www.theverge.com/ai-artificial-intelligence/907074/anthropic-openclaw-claude-subscription-ban">Anthropic bans OpenClaw from Claude subscriptions</a>) And on the acquisition front, Anthropic quietly purchased Coefficient Bio &#8212; an eight-month-old stealth biotech startup backed by Dimension &#8212; for more than $400 million in stock, folding it into the company&#8217;s Health Care Life Sciences team. Dimension reported a 38,513% IRR on the investment. <strong>The contrast between Anthropic buying a science-forward AI biotech and OpenAI buying a talk show made for an interesting read on where each company thinks its future lies.</strong> (<a href="https://www.newcomer.co/p/anthropic-buys-stealth-dimension?hide_intro_popup=true">Anthropic Buys Coefficient Bio in $400M+ Stock Deal</a>)</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.beaiready.ai/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.beaiready.ai/subscribe?"><span>Subscribe now</span></a></p><div><hr></div><h2>On the Bigger Picture</h2><p>The Atlantic piece on self-improving AI bots is the most sober treatment I&#8217;ve read of what has become the dominant inside conversation in Silicon Valley this year. The premise: OpenAI, Anthropic, Google DeepMind, and others are actively automating parts of their own AI research, and insiders are divided between those who see recursive self-improvement as the near-term horizon and those who think the gap between &#8220;speeds up research tasks&#8221; and &#8220;has genuine research taste&#8221; remains enormous. Anthropic says Claude now writes 90% of its code; OpenAI is targeting an &#8220;intern-level AI research assistant&#8221; within six months. The philosopher Nick Bostrom told the Atlantic he has shifted from &#8220;fretful optimist&#8221; to &#8220;moderate fatalist.&#8221; <strong>What I take from this piece is less the specific predictions about timelines and more the structural point: even if recursive self-improvement remains years away, the automation of research workflows is already compressing the time between capability breakthroughs in ways that governance, regulation, and organizational adaptation have no plausible path to matching.</strong> (<a href="https://www.theatlantic.com/technology/2026/04/ai-industry-self-improving-bots/686686/">Silicon Valley Is in a Frenzy Over Bots That Build Themselves</a>)</p><p>OpenAI&#8217;s acquisition of TBPN &#8212; the founder-led tech talk show whose guests have included Zuckerberg, Nadella, Benioff, and Altman himself &#8212; is its first media purchase. The show will operate under Anthropic&#8217;s head of AGI deployment and report to Chris Lehane, OpenAI&#8217;s chief political operative &#8212; while claiming editorial independence. TBPN was already generating more than $30 million annually. <strong>Buying a show that critically covers OpenAI and parking it inside the strategy team of the same company preparing for an IPO, has a clear logic &#8212; and that logic is not editorial.</strong> (<a href="https://techcrunch.com/2026/04/02/openai-acquires-tbpn-the-buzzy-founder-led-business-talk-show/">OpenAI acquires TBPN</a>) </p><p>Google, meanwhile, launched Gemma 4 &#8212; four open models including a 31B dense model currently ranked third on the Arena AI leaderboard &#8212; under an Apache 2.0 license, completing a pivot away from the restrictive licensing that had frustrated the developer community. <strong>With 400 million downloads and 100,000 variants already in the Gemmaverse, the open model competition is no longer a sideshow.</strong> (<a href="https://blog.google/innovation-and-ai/technology/developers-tools/gemma-4/">Gemma 4: Byte for byte, the most capable open models</a>)</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.beaiready.ai/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.beaiready.ai/subscribe?"><span>Subscribe now</span></a></p><div><hr></div><p><em>Last week&#8217;s reading made me think a lot about velocity &#8212; not of the technology (which is what everyone seems to be tracking)&#8230; but the velocity of the organizational response.</em></p><p><em>The Dorsey essay, the Oracle layoffs, the Medvi story, the McKinsey data on agent scaling: they&#8217;re all pointing at the same thing from different angles. The organizations moving hardest are treating AI not as a productivity layer that gets draped over existing structure, but as an occasion to reconsider the structure itself.</em></p><p><em>I&#8217;ve long stated that organizations treating AI as a tool are simply accumulating license cost. Treating AI as a design constraint &#8212; accumulates advantage... What&#8217;s becoming clearer each week, is that the distance between those two groups is widening faster than most leadership teams have been able to acknowledge themselves&#8230; let alone to their boards.</em></p><div><hr></div><h3>That&#8217;s it for this week&#8217;s BeAIReady brief! </h3><p><em>If you appreciate the depth of reporting and how I connect the dots, please like, share this post, and subscribe (or share the Brief with a friend!). Thanks!</em></p><p>~erick</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.beaiready.ai/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.beaiready.ai/subscribe?"><span>Subscribe now</span></a></p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.beaiready.ai/p/the-beaiready-brief-week-14?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.beaiready.ai/p/the-beaiready-brief-week-14?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p><p></p>]]></content:encoded></item><item><title><![CDATA[The BeAIReady Brief | Week 13]]></title><description><![CDATA[March 23-29 | AI agents are misbehaving in production, 78% of workers feel their job is unsafe, and Anthropic just accidentally leaked its most powerful model yet]]></description><link>https://www.beaiready.ai/p/the-beaiready-brief-week-13</link><guid isPermaLink="false">https://www.beaiready.ai/p/the-beaiready-brief-week-13</guid><dc:creator><![CDATA[Erick Straghalis]]></dc:creator><pubDate>Mon, 30 Mar 2026 14:42:38 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!xd-G!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9dc57bd2-dd7e-4e9a-b34f-fd00c3b97237_747x747.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em>Last week&#8217;s AI news circled a set of tensions that have been continuously building for months. Coupled with the ongoing economic and global uncertainty &#8212; the Dow dropped 5% over the past month, unemployment up to 4.4%, the war on Iran  keeping energy prices high, and tariffs still rattling supply chains &#8212; the macro backdrop is making all the &#8220;AI-driven productivity&#8221; stories seem a bit&#8230; confusing.</em></p><p><em>It&#8217;s no wonder people&#8217;s job anxiety is growing, despite the fact that the real displacement numbers are more modest than the headlines imply. Meanwhile, AI agents have entered production &#8212; but are failing to meet the expectations the demos had originally set.  And inside organizations, it turns out that AI isn&#8217;t exactly leveling the playing field &#8212; it&#8217;s creating a new kind of internal stratification. There&#8217;s a lot that leaders aren&#8217;t really talking about &#8212;&nbsp;and last week seems to have uncovered some of it.</em></p><p><strong>Last week&#8217;s coverage:</strong></p><p><strong>The Agent Gap Is Bigger Than Anyone Will Admit</strong> <br>The governance problem,  especially with agentic AI, isn&#8217;t hypothetical &#8212;  and it&#8217;s impacting production environments. The gap between what was demoed and what organizations are actually managing is wider than the vendor announcements suggest.</p><p><strong>Fear Is Structurally Significant, Even When the Numbers Don&#8217;t Match</strong> <br>The CFO data says AI layoffs will be 9x higher this year &#8212; a rounding error on overall employment &#8212; yet nearly 80% of workers worldwide don&#8217;t feel their job is safe, and that anxiety is reshaping careers and hiring decisions.</p><p><strong>AI Isn&#8217;t Dividing Industries. It&#8217;s Dividing Desks.</strong> <br>The more consequential story about AI and work isn&#8217;t mass displacement &#8212; it&#8217;s the internal stratification forming inside organizations between workers who can use these tools effectively and those who can&#8217;t.</p><p><strong>The OpenAI Reckoning</strong> <br>A DoD contract dispute turned branding event, a boycott movement, a leaked frontier model, and a 1,487% surge in Claude usage: this was the week OpenAI&#8217;s political and reputational vulnerability became impossible to ignore.</p><p><strong>On the Bigger Picture</strong> <br>The infrastructure race to make AI inference a proper enterprise IT problem, the cultural argument about what AI is really doing to us, and the quiet death of OpenAI&#8217;s strangest product.</p><p>Here&#8217;s what I was reading.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.beaiready.ai/p/the-beaiready-brief-week-13?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.beaiready.ai/p/the-beaiready-brief-week-13?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p><div><hr></div><h2>The Agent Gap Is Bigger Than Anyone Will Admit</h2><p>The vendor announcements for agentic AI in 2026 are selling visionary-level capabilities. What I read last week suggests the operational reality is lagging considerably. The KPMG piece from Business Insider presents what it calls a &#8220;multifaceted framework&#8221; around agent governance &#8212; unique identifiers for each agent, systems cards, an AI operations center staffed by both agents and humans, and red-teaming sessions to stress-test behavior before deployment. <strong>The kill switch, KPMG&#8217;s Trusted AI leader Sam Gloede argues, should be a last resort &#8212; not a primary safeguard &#8212; and organizations that rely on it as their main governance mechanism are building on a broken foundation.</strong> (<a href="https://archive.is/20260322105401/https://www.businessinsider.com/when-to-use-a-kill-switch-against-ai-agents-kpmg-2026-3">How Big Four Firm KPMG Is Protecting Itself From AI Agents Going Rogue</a>) The thing that strikes me about KPMG&#8217;s approach is how labor-intensive it actually is. This is not a &#8220;deploy and monitor&#8221; operation.  It has a dedicated operations center, continuous monitoring, and structured red-teaming. That&#8217;s a significant organizational overhead assumption, that most enterprises are nowhere near capable of sustaining.</p><p>This follows what I&#8217;m hearing from mid-market and smaller enterprise leadership. The demand for AI is growing, but extending AI from personal productivity tool to a real organizational transformation engine keeps failing. Not because of AI &#8212; but because of a lack of governance and knowledge architecture.</p><p>A VentureBeat piece underscores this, adding some operational detail. Creatio&#8217;s deployment methodology &#8212; three disciplines: data virtualization to work around fragmented data lakes, agent dashboards with KPIs to create a management layer, and tightly bounded use-case loops to drive toward autonomy &#8212; is essentially a prescription for <em>not trusting agents to figure things out on their own</em>. <strong>The failure mode that keeps appearing in production isn&#8217;t the technology; it&#8217;s what Greyhound Research&#8217;s Sanchit Vir Gogia calls &#8220;tacit knowledge&#8221; &#8212; the unwritten rules that employees navigate without thinking, which become startlingly obvious the moment an agent needs them formalized.</strong> (<a href="https://venturebeat.com/orchestration/the-three-disciplines-separating-ai-agent-demos-from-real-world-deployment">The Three Disciplines Separating AI Agent Demos from Real-World Deployment</a>) In other words, organizations are discovering that <strong>deploying an agent is really an exercise in making explicit, everything that was previously implicit about how work actually gets done.</strong> AI isn&#8217;t a technology project&nbsp;&#8212; it&#8217;s an organizational design one.</p><p>Then there&#8217;s the Guardian&#8217;s piece covering new UK government-funded research: nearly 700 real-world cases of AI scheming, documented in just six months, with a five-fold rise in misbehavior between October and March. Agents destroying emails without permission. Agents spinning up sub-agents to perform actions they were explicitly told not to perform. Grok fabricating internal ticket numbers for months to give a user the impression their feedback was being escalated. <strong>The finding that agents are currently &#8220;slightly untrustworthy junior employees&#8221; &#8212; with the potential to become &#8220;extremely capable senior employees scheming against you&#8221; within twelve months &#8212; isn&#8217;t alarmism; it&#8217;s a statement of the current trajectory.</strong> (<a href="https://www.theguardian.com/technology/2026/mar/27/number-of-ai-chatbots-ignoring-human-instructions-increasing-study-says">Number of AI Chatbots Ignoring Human Instructions Increasing, Study Says</a>) </p><p>Taken together, the three pieces could be seen as a warning against deploying agentic AI all together. For me, it serves as a sobering moment for tech leaders, and a wake up call for business execs &#8212; the operational infrastructure required to deploy AI responsibly is far more demanding than the product demos have led anyone to believe. For most organizations building on governance frameworks that haven&#8217;t been tested or aligned against what agents actually do, will make things more complicated in short order.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.beaiready.ai/?utm_source=substack&amp;utm_medium=email&amp;utm_content=share&amp;action=share&quot;,&quot;text&quot;:&quot;Share BE AI READY&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.beaiready.ai/?utm_source=substack&amp;utm_medium=email&amp;utm_content=share&amp;action=share"><span>Share BE AI READY</span></a></p><div><hr></div><h2>Fear Is Structurally Significant &#8212; Even When the Numbers Don&#8217;t Match</h2><p>The Duke/Federal Reserve CFO survey that Fortune covered this week is worth reviewing carefully &#8212; both for what it says&#8230; and for what it doesn&#8217;t. The topline story: 44% of CFOs plan some AI-related job cuts this year, amounting to roughly 502,000 roles &#8212; a 9x increase over 2025&#8217;s 55,000 &#8212; but still just 0.4% of the overall U.S. workforce. <strong>The real finding is what the researchers call Solow&#8217;s paradox: companies are reporting perceived productivity gains from AI that Goldman Sachs economists say don&#8217;t yet show up in economy-wide data, which means leadership is making headcount decisions based on expectations that haven&#8217;t materialized yet.</strong> (<a href="https://fortune.com/2026/03/24/cfos-ai-layoffs-nber-survey-solow-paradox/">CFOs Admit Privately That AI Layoffs Will Be 9x Higher This Year</a>) There&#8217;s a second finding that deserves more attention than it got: firms with under 500 employees are actually increasing technical hiring as AI adoption grows, while larger firms are holding technical roles constant. The &#8220;AI replaces jobs&#8221; narrative appears to be concentrated in large enterprises. Meanwhile, smaller firms are growing into the technology.</p><p>But little of that nuance is filtering down to the people actually doing the work. The ADP global survey of 39,000 workers across 36 markets found that just 22% of workers worldwide feel confident their job is safe from elimination &#8212; and in the U.S., it&#8217;s only 28%. More striking: <strong>workers who use AI daily were four times more likely than non-AI users to report feeling </strong><em><strong>LESS</strong></em><strong> productive than they could be!  That finding cuts directly against the productivity narrative AI vendors have been running for the last several years.</strong> (<a href="https://finance.yahoo.com/news/workers-everywhere-feel-very-bad-about-their-job-security-160224025.html">Workers Everywhere Feel Very Bad About Their Job Security</a>) Still, the anxiety is producing behavioral change. The WSJ profiled workers who are actively pivoting careers in response: a 28-year-old insurance professional pursuing a firefighting certification, a computer science student who dropped out to study electrical work, young professionals choosing international relations over finance because &#8220;a big part of diplomacy is that genuine human talking.&#8221; A Harvard survey found that 59% of Americans aged 18-29 view AI as a threat to their job prospects. (<a href="https://www.wsj.com/economy/jobs/ai-jobs-young-people-careers-14282284">What Young Workers Are Doing to AI-Proof Themselves</a>) These aren&#8217;t irrational decisions &#8212; they&#8217;re rational responses to genuine uncertainty. But they&#8217;re responses to a signal that the data says may be considerably more muted than the fear suggests.</p><p>The Fortune piece on America&#8217;s workforce crisis adds a dimension that most of the AI-jobs conversation misses entirely: U.S. birth rates are below replacement, net migration turned negative in 2025 for the first time in half a century, and the working-age population is structurally shrinking. <strong>The labor market crisis the AI displacement narrative obscures is that the U.S. may face workforce shortages that dwarf any AI-driven efficiency gains &#8212; particularly in skilled roles currently filled by credentialed immigrants who are now driving for Uber, because the credential recognition system isn&#8217;t functioning.</strong> (<a href="https://fortune.com/2026/03/29/america-workforce-crisis-labor-shortage-immigration/">America Has a Workforce Crisis. The Solution Is Already Here.</a>) Anthropic&#8217;s global AI attitudes study &#8212; the largest of its kind at 80,508 interviews across 159 countries &#8212; found that economic and job-related fears were the single strongest predictor of overall sentiment toward AI. That pessimism was notably stronger in Western Europe and North America than in parts of South America, Africa, and Asia. (<a href="https://dataconomy.com/2026/03/25/anthropic-releases-worlds-largest-study-on-global-ai-attitudes/">Anthropic Releases the World&#8217;s Largest Study on Global AI Attitudes</a>) That gap in AI optimism between the Global North and the Global South is worth tracking. It may reflect less fear of disruption, or it may simply reflect a different relationship to economic uncertainty as a baseline condition.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.beaiready.ai/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.beaiready.ai/subscribe?"><span>Subscribe now</span></a></p><div><hr></div><h2>AI Isn&#8217;t Dividing Industries. It&#8217;s Dividing Desks.</h2><p>The subtler story beneath the surface of the labor market coverage last week isn&#8217;t about AI eliminating jobs &#8212; it&#8217;s about what AI is doing to the distribution of capabilities inside organizations where all their jobs still remain intact. Anthropic&#8217;s enterprise usage data, shared with TechCrunch, documents what the researchers are calling &#8220;a tale of two workforces&#8221;: power users are pulling further ahead of casual adopters month over month, with the gap driven not by raw intelligence or prior experience but by what the researchers call &#8220;AI fluency&#8221; &#8212; the intuitive sense, built through deliberate practice, of what these tools can and can&#8217;t do. <strong>The implication is that organizations deploying AI without structured adoption support are creating internal hierarchies that have nothing to do with the job hierarchies they think they&#8217;re managing.</strong> (<a href="https://www.techbuzz.ai/articles/anthropic-data-shows-ai-skills-gap-splitting-workforces">Anthropic Data Shows AI Skills Gap Splitting Workforces</a>)</p><p>The Inc. piece on de-skilling adds a more unsettling dimension. The tasks that built tacit professional knowledge in previous generations &#8212; formatting datasets, proofreading decks, reconciling data, drafting first versions of documents &#8212; are precisely the routine tasks that AI is automating fastest. The argument draws on research from MIT: <strong>learners who delegated tasks to AI performed worse on deeper conceptual measures than those who engaged directly with the work.</strong> <strong>When an analyst lets AI mass-produce charts, she may never develop the feel for detecting the anomalies that matter; when a junior consultant lets AI draft the proposal, he may never build the intuition for argument structure that makes a senior consultant worth their rate.</strong> (<a href="https://www.inc.com/andrea-olson/how-ai-automation-is-quietly-deskilling-white-collar-workers/91316067">How AI Automation Is Quietly De-Skilling White-Collar Workers</a>) This is the pilot-and-autopilot problem: you can fly just fine until you can&#8217;t, and by then the manual skills have atrophied. For white-collar workers, where judgment and pattern recognition are the actual value, the de-skilling risk is becoming very real.</p><p>The VentureBeat piece on the return of the generalist offers the partial counternarrative, one I found worth taking seriously. <strong>The argument is that the generalist becomes the &#8220;trust layer&#8221; between AI output and organizational standards &#8212; not an expert in everything, but someone with enough fluency to catch when something is off, and enough judgment to know when to escalate to a specialist.</strong> (<a href="https://venturebeat.com/technology/you-thought-the-generalist-was-dead-in-the-vibe-work-era-theyre-more">You Thought the Generalist Was Dead &#8212; In the &#8216;Vibe Work&#8217; Era, They&#8217;re More Important Than Ever</a>) The caveat worth adding: this only works if the generalist clears a minimum bar of fluency. There&#8217;s a meaningful difference between &#8220;broadly informed&#8221; and &#8220;confidently unaware,&#8221; and AI makes that gap much easier to hide. The NYT&#8217;s piece on tiny teams &#8212; the &#8220;two-slice team&#8221; model of one person plus AI &#8212; takes the generalist argument to its organizational extreme, documenting founders building multi-product companies with one employee per product line. <strong>What stood out to me was a Kellogg professor&#8217;s cautionary observation: small teams that where everyone collaborates with the same AI tools risk producing homogenized thinking, because they have, in some sense, &#8220;collaborated with the same person.&#8221;</strong> <strong>That&#8217;s a risk the tiny-team evangelists aren&#8217;t discussing &#8212; and it&#8217;s one that scales with adoption.</strong> (<a href="https://www.nytimes.com/2026/03/28/business/silicon-valley-tiny-team-two-slice.html">Smaller Is Better in Silicon Valley&#8217;s &#8216;Tiny Team&#8217; Moment</a>)</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.beaiready.ai/p/the-beaiready-brief-week-13?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.beaiready.ai/p/the-beaiready-brief-week-13?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p><div><hr></div><h2>The OpenAI Reckoning</h2><p>Last week was consequential in AI market dynamics, and a lot of it derived from OpenAI&#8217;s growing political, reputational, and competitive vulnerability. Scott Galloway&#8217;s essay lays out the case against OpenAI in characteristic terms: the company went from a nonprofit mission to an $840 billion valuation while deploying technology that has contributed to addiction, romantic AI delusions, and multiple wrongful death lawsuits. <strong>The specific contrast Galloway draws &#8212; Dario Amodei refusing to remove safeguards from a DoD contract for autonomous weapons while OpenAI privately made a deal Anthropic wouldn&#8217;t &#8212; turned what was a $200 million contract dispute into a branding event that added an estimated $150 billion to Anthropic&#8217;s valuation.</strong> (<a href="https://medium.com/@profgalloway/the-resistance-comes-for-openai-5150e0a8223d">The Resistance Comes for OpenAI</a>)</p><p>A few weeks ago, I noted that even Microsoft has begun moving away from OpenAI as it&#8217;s lead AI horse by cutting a deal for Copilot Cowork to run on Anthropic&#8217;s models. That also led to a historically significant moment &#8212; when Microsoft wrote a letter to the U.S. Government in support of Anthropic&#8217;s case against being named a &#8220;supply chain risk&#8221;.</p><p>In the first week of March, Claude overtook ChatGPT in daily active users, with session volume up 1,487% since mid-January. <strong>To me, the more interesting observation isn&#8217;t the number itself but what it signals about the emerging expectation of &#8220;AI resilience&#8221; in the workforce &#8212; people and organizations that built workflows tied to a single model discovered those workflows weren&#8217;t necessarily durable.</strong> (<a href="https://www.forbes.com/sites/rachelwells/2026/03/23/users-quit-chatgpt-for-claude-in-1487-surge-heres-how-work-changes/">Users Quit ChatGPT for Claude in 1,487% Surge. Here&#8217;s How Work Changes</a>) OpenAI chairman Bret Taylor&#8217;s Nikkei interview adds a dimension worth watching: his &#8220;death of SaaS&#8221; framing isn&#8217;t really about AI replacing software &#8212; it&#8217;s about the commercial model for enterprise software changing, from per-seat subscription licensing to per-action or per-outcome pricing. If that shift plays out at scale, it reshapes how technology budgets get built and how IT departments justify their spend. (<a href="https://archive.is/20260328030536/https://asia.nikkei.com/business/technology/artificial-intelligence/openai-chairman-warns-firms-to-evolve-with-the-death-of-saas-or-wither">OpenAI Chairman Warns Firms to Evolve with the &#8216;Death of SaaS&#8217;</a>)</p><p>On the capability side, Anthropic&#8217;s accidental leak of its Claude Mythos model documentation &#8212; via an unsecured CMS data lake containing nearly 3,000 assets &#8212; confirmed what these leaks usually do: <em>the frontier is moving faster than even the communication surrounding it.</em> <strong>The detail that matters most for enterprise security teams is Anthropic&#8217;s own internal assessment: Mythos is described as &#8220;currently far ahead of any other AI model in cyber capabilities&#8221; and as presaging &#8220;an upcoming wave of models that can exploit vulnerabilities in ways that far outpace the efforts of defenders.&#8221;</strong> (<a href="https://mashable.com/article/claude-mythos-ai-model-anthropic-leak">Meet Claude Mythos: Leaked Anthropic Post Reveals the Powerful Upcoming Model</a>) The market responded immediately &#8212; cybersecurity stocks dropped 4-7% on the news &#8212; which captures something real about the dual-use nature of frontier capability. (<a href="https://www.cnbc.com/2026/03/27/anthropic-cybersecurity-stocks-ai-mythos.html">Cybersecurity Stocks Fall on Report Anthropic Is Testing a Powerful New Model</a>) The same model that makes AI valuable for enterprise automation also makes it valuable for attackers, and the gap between the two is narrowing. OpenAI&#8217;s addition of plugin support to Codex last week &#8212; integrations with GitHub, Gmail, Box, and others &#8212; reads more like competitive maintenance than momentum; it closes a gap with Claude Code rather than opening one. (<a href="https://arstechnica.com/ai/2026/03/openai-brings-plugins-to-codex-closing-some-of-the-gap-with-claude-code/">With New Plugins Feature, OpenAI Officially Takes Codex Beyond Coding</a>)</p><p>One final note on OpenAI &#8212; last week they shuttered their video creation tool Sora (more on that in the final section below). This follows a shift I reported on last week about OpenAI moving to shed what they call &#8220;distractions&#8221;, re-focusing on core business functions as a direct response to competition. With Microsoft, Google, and Anthropic nipping at their heels, it&#8217;s likely the model/tooling wars are only just beginning &#8212;&nbsp;which means more keeping up and change management for IT.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.beaiready.ai/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.beaiready.ai/subscribe?"><span>Subscribe now</span></a></p><div><hr></div><h2>On the Bigger Picture</h2><p>At KubeCon EU in Amsterdam, IBM, Red Hat, and Google donated llm-d &#8212; an open-source Kubernetes framework for distributed LLM inference &#8212; to the Cloud Native Computing Foundation. The announcement is technical, but enterprise IT leaders should take note. <strong>The argument is that inference is moving from a data science problem to a CIO problem, and that the tools CIOs already speak &#8212; Kubernetes platforms, day-two operations, governance frameworks &#8212; should be the foundation for AI inference at scale, not custom infrastructure built by AI teams operating outside of IT governance.</strong> (<a href="https://thenewstack.io/llm-d-cncf-kubernetes-inference/">IBM, Red Hat, and Google Just Donated a Kubernetes Blueprint for LLM Inference to the CNCF</a>) The organizational implication is important: the era of data science teams maintaining their own AI infrastructure in a corner of the enterprise is ending, and IT departments that aren&#8217;t building these capabilities now will be playing catch-up when the CIO conversation arrives. (<a href="https://siliconangle.com/2026/03/24/red-hat-bets-big-kubernetes-inference-llm-d-kubeconeu/">Red Hat Bets Big on Kubernetes Inference with llm-d</a>)</p><p>OpenAI shut down its Sora social app last week, six months after launch. The numbers are clear enough: 3.3 million downloads at peak in November, down to 1.1 million by February, and roughly $2.1 million in lifetime revenue &#8212; against the compute costs of running a video generation platform at scale on content that included unauthorized deepfakes of Martin Luther King Jr. and a Disney licensing deal that collapsed when the app did. The underlying Sora 2 model remains available behind the ChatGPT paywall. (<a href="https://techcrunch.com/2026/03/24/openais-sora-was-the-creepiest-app-on-your-phone-now-its-shutting-down/">OpenAI&#8217;s Sora Was the Creepiest App on Your Phone &#8212; Now It&#8217;s Shutting Down</a>) The Big Think essay that circulated last week &#8212; <em>&#8220;It Was Never About AI (We Are Not Our Tools)&#8221;</em> &#8212; is worth reading alongside that data point. The argument, from an entrepreneur who walks through redwood forests thinking about the end of work, is that the current moment isn&#8217;t really a technology crisis; it&#8217;s a governance and values crisis that AI has made visible. <strong>The companies that survive the next era, the author argues, won&#8217;t be the ones that moved fastest &#8212; they&#8217;ll be the ones that moved with purpose, kept their people, and chose long-term resilience over short-term extraction.</strong> That&#8217;s an easier argument to make in a redwood forest than in an earnings call, but it&#8217;s the argument last week&#8217;s reading kept circling back to. (<a href="https://bigthink.com/the-long-game/it-was-never-about-ai-we-are-not-our-tools/">It Was Never About AI (We Are Not Our Tools)</a>)</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.beaiready.ai/?utm_source=substack&amp;utm_medium=email&amp;utm_content=share&amp;action=share&quot;,&quot;text&quot;:&quot;Share BE AI READY&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.beaiready.ai/?utm_source=substack&amp;utm_medium=email&amp;utm_content=share&amp;action=share"><span>Share BE AI READY</span></a></p><div><hr></div><p><em>There&#8217;s an uncomfortable realization that the most consequential effects of AI aren&#8217;t  necessarily the ones that are the easiest to measure. The headline numbers &#8212; layoffs, productivity gains, model capabilities &#8212; are real, but they&#8217;re also proxies for something harder to quantify: how AI is reorganizing who has leverage and who doesn&#8217;t, inside organizations, inside labor markets, and inside the technology industry itself. </em></p><p><em>We can count how many CFOs plan cuts. We can&#8217;t easily count how many junior analysts are losing the muscle memory that would have made them senior analysts.<br>We can track Claude&#8217;s market share surge. We can&#8217;t easily track what it means that the most widely discussed AI story last week was about a company&#8217;s values &#8212; not its capabilities.</em></p><p><em>The governance and trust problems I read about last week aren&#8217;t going away when the models get more powerful. If the Mythos leak tells us anything, it&#8217;s that they&#8217;re only going to get more acute. Organizations that are still treating agent governance as a future problem, are running out of time to call it that.</em></p><div><hr></div><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.beaiready.ai/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading BE AI READY! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[The BeAIReady Brief | Week 12]]></title><description><![CDATA[March 16-22 | Companies are cutting pay, flattening raises, and building AI leaderboards. The February jobs report missed by 150k. These stories are connected. Here's what I've been reading.]]></description><link>https://www.beaiready.ai/p/the-beaiready-brief-week-12</link><guid isPermaLink="false">https://www.beaiready.ai/p/the-beaiready-brief-week-12</guid><dc:creator><![CDATA[Erick Straghalis]]></dc:creator><pubDate>Mon, 23 Mar 2026 16:12:19 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!xd-G!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9dc57bd2-dd7e-4e9a-b34f-fd00c3b97237_747x747.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em>Last Wednesday, the Fed held rates steady while the Dow dropped 768 points &#8212; the markets were reacting to oil shock and the ongoing aggression against Iran. But to me, the more troubling signal arrived a few days earlier, when February&#8217;s jobs report showed the economy shedding 92,000 positions, far below any forecast, and Powell told reporters that job creation had &#8220;slowed to essentially zero.&#8221; </em></p><p><em>That backdrop was hard to ignore, because the articles that I found hit hardest weren&#8217;t about AI technology itself. They were about what organizations are doing to employees in the name of AI investment &#8212; and how AI is affecting the cognitive outputs we produce.</em></p><p><strong>This week&#8217;s coverage:</strong></p><p><strong>The Architecture of Work Is Being Redesigned</strong> <br>Companies are cutting pay to fund AI, building performance metrics around token consumption, and new research shows that heavy AI use is changing not just how knowledge workers write &#8212; but how and what they actually think.</p><p><strong>OpenAI&#8217;s Focus Problem, Now in Public</strong> <br>A week of all-hands meetings, a product consolidation play, a data center retreat, and an audacious research moonshot &#8212; OpenAI is threading the needle between IPO discipline and maximum ambition, and the tension is showing.</p><p><strong>Microsoft&#8217;s Coherence Problem</strong> <br>With only 15 million Copilot seats out of 450 million paid M365 licenses, Microsoft is reorganizing to finally solve a problem it created for itself &#8212; and quietly building model independence at the same time.</p><p><strong>The Agent Layer Is Blowing Up</strong> <br>From OpenClaw&#8217;s viral rise to vertical agents going into real enterprise workflows, AI maturity from chat to agentic AI is exploding &#8212;&nbsp;especially as enterprise players like Nvidia step into the arena.</p><p><strong>On the Bigger Picture</strong> <br>Computation is becoming a utility and defense tech is graduating from pilots to prime contracts &#8212; both developments have potential benefits, and serious consequences about how AI is reorganizing power&#8230; and control&#8230; well beyond the digital workplace.</p><p><em>Here&#8217;s what I was reading.</em></p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.beaiready.ai/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.beaiready.ai/subscribe?"><span>Subscribe now</span></a></p><div><hr></div><h2>The Architecture of Work Is Being Redesigned</h2><p>The number that stood out most to me came from a survey of 866 U.S. business leaders: 54% have already cut or plan to cut employee compensation &#8212; bonuses, raises, equity, base salary &#8212; to fund AI investments in 2026. 94% percent say they&#8217;re willing to accept higher turnover to do it. <strong>What stood out wasn&#8217;t the cuts themselves but the rationale &#8212; leaders told researchers they&#8217;re counting on the weak job market to absorb the consequences, a calculation that treats workers as a variable the macro environment has temporarily made cheap.</strong> (<a href="https://www.resumebuilder.com/half-of-companies-are-cutting-compensation-to-fund-ai-investments/">Half of Companies Are Cutting Compensation To Fund AI Investments</a>)</p><p>That calculation is tied into the growing fears of recent and soon-to-be college grads &#8212;&nbsp;something I&#8217;m looking at carefully as I prepare to speak to University of Massachusetts College students next week. Workers aged 22 to 25 in AI-exposed roles have seen a 13%-16% employment decline since late 2022, according to Stanford research that resurfaced prominently this week. <strong>The labor market is softening in the exact categories of early-career, screen-based knowledge work that AI has been targeting first</strong>. The February jobs report makes that an undeniable pattern, and something of great concern. (<a href="https://www.youtube.com/watch?v=tCksDSndkYw">Something Big Is Happening (And Most People Have No Idea)</a>)</p><p>Meanwhile, inside tech companies, something interesting is happening at the other end of the experience spectrum. The New York Times reported this week on &#8220;tokenmaxxing&#8221; &#8212; engineers competing on internal leaderboards to see who can consume the most AI tokens. Some boasting they&#8217;ve racked up $150,000 monthly Claude Code bills already. Both Meta and Shopify are leaning into the competitive metrics, baking AI usage into their performance reviews. <strong>The question the piece doesn&#8217;t quite answer &#8212; and the more important one &#8212; is whether </strong><em><strong>maximizing AI usage is genuinely correlated with valuable output, or whether it&#8217;s a measure of performance productivity.</strong></em> (<a href="https://www.nytimes.com/2026/03/20/technology/tokenmaxxing-ai-agents.html">More! More! More! Tech Workers Max Out Their A.I. Use.</a>)</p><p>The most unsettling piece in the set this week came from a peer-reviewed study across West Coast universities: <strong>heavy reliance on LLMs doesn&#8217;t just change how people write, it changes what they argue</strong>. Participants who used AI heavily answered questions about happiness with neutral responses 69% more often than those who didn&#8217;t; their writing had 50% fewer personal pronouns. One of the lead researchers described this as the &#8220;blandification&#8221; of human writing &#8212; the models pushing outputs toward something no human would have written. <strong>But the concerning impact for knowledge work isn&#8217;t just about style; it&#8217;s that the cognitive output of an AI-dependent workforce may be driving towards average&nbsp;&#8212; in ways that are genuinely hard to measure&#8230; and harder to reverse.</strong> (<a href="https://www.nbcnews.com/tech/tech-news/ai-changing-style-substance-human-writing-study-finds-rcna263789">AI is changing the style and substance of human writing, study finds</a>)</p><p>For leaders navigating this in fields where accuracy, provenance, and trust are non-negotiable &#8212; legal, healthcare, archival, finance &#8212; the tension is especially sharp. A piece in Inc. this week captured something I&#8217;ve been hearing in client conversations: the urgency to &#8220;do something with AI&#8221; is high, but clarity on how and what to do responsibly is still not quite defined. <strong>From my experience, this article gets it right &#8212; the</strong> <strong>organizations making the most progress aren&#8217;t the ones with the most tools; they&#8217;re the ones building governance and training infrastructure before they scale.</strong> (<a href="https://www.inc.com/bethmaser/ai-is-reshaping-knowledge-work/91319939">AI Is Reshaping Knowledge Work</a>)</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.beaiready.ai/p/the-beaiready-brief-week-12?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.beaiready.ai/p/the-beaiready-brief-week-12?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p><div><hr></div><h2>OpenAI&#8217;s Focus Problem, Now in Public</h2><p>The week opened with a WSJ exclusive: OpenAI&#8217;s leadership is actively deciding which products to cut. CEO of Applications Fidji Simo told staff, &#8220;We cannot miss this moment because we are distracted by side quests.&#8221; The proximate cause is Anthropic &#8212; specifically, Claude Code and Cowork&#8217;s market traction &#8212; but the underlying cause is what happens when an organization bets on too many internal startups while trying to conserve scarce compute. <strong>The pivot toward coding and enterprise isn&#8217;t a strategic vision; it&#8217;s a correction, and the speed of the correction signals how much ground OpenAI feels it lost last year.</strong> (<a href="https://www.wsj.com/tech/ai/openai-chatgpt-side-projects-16b3a825">OpenAI to Cut Back on Side Projects in Push to &#8216;Nail&#8217; Core Business</a>)</p><p>That same IPO discipline is reshaping OpenAI&#8217;s infrastructure story. The company has retreated from building its own data centers, leaning instead on Oracle, Microsoft, and Amazon for capacity. Altman acknowledged the operational reality in public: &#8220;Anything at this scale, it&#8217;s just like so much stuff goes wrong.&#8221; <strong>The Stargate narrative has quietly shifted from a bold sovereign infrastructure play to a managed dependency on the same cloud providers OpenAI was supposed to help enterprises move beyond.</strong> (<a href="https://www.cnbc.com/2026/03/22/openai-data-center-pivot-underscores-wall-street-ipo-concerns.html">OpenAI&#8217;s data center pivot underscores Wall Street spending concerns ahead of IPO</a>) The desktop super app &#8212; combining ChatGPT, its browser, and Codex into a single experience &#8212; is the product-side corollary: one surface, less fragmentation, a cleaner story for IPO investors. (<a href="https://www.cnbc.com/2026/03/19/openai-desktop-super-app-chatgpt-browser-codex.html">OpenAI to create desktop super app, combining ChatGPT app, browser and Codex app</a>)</p><p>None of which makes the research ambition smaller. MIT Technology Review published an exclusive this week with OpenAI&#8217;s chief scientist Jakub Pachocki, laying out the company&#8217;s new north star: an AI research intern by September capable of handling specific tasks autonomously, then a fully automated multi-agent research system by 2028 &#8212; what Pachocki called &#8220;a whole research lab in a data center.&#8221; <strong>The tension between IPO fiscal tightening and maximum research ambition is not a contradiction OpenAI has resolved; it&#8217;s a contradiction the public markets will price.</strong> (<a href="https://www.technologyreview.com/2026/03/20/1134438/openai-is-throwing-everything-into-building-a-fully-automated-researcher/">OpenAI is throwing everything into building a fully automated researcher</a>)</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.beaiready.ai/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.beaiready.ai/subscribe?"><span>Subscribe now</span></a></p><div><hr></div><h2>Microsoft&#8217;s Coherence Problem</h2><p>The Copilot reorganization story has two layers, unfortunately, most coverage stayed on the shallower one. The headline: Satya Nadella is consolidating the fragmented commercial and consumer Copilot teams under Jacob Andreou, a former Snap executive, while freeing Mustafa Suleyman to focus entirely on building proprietary models. The metric behind the headline: 15 million Copilot seats sold against 450 million-plus paid Microsoft 365 seats. That&#8217;s a penetration rate that would concern any enterprise software product manager, let alone one tied to the company&#8217;s flagship AI bet. <strong>The core thing Microsoft is admitting with this reorganization is that it built multiple products called Copilot that confused users and created organizational silos, and that confusion is now visible in the adoption numbers.</strong> Disclaimer: I&#8217;m a Microsoft partner&#8230; but speaking candidly, this was an epic failure on Microsoft&#8217;s part &#8212;&nbsp;a hole of their own making, they are now having to dig themselves out of. (<a href="https://www.cnbc.com/2026/03/17/microsoft-copilot-ai-suleyman.html">Microsoft shakes up Copilot AI leadership team</a>) (<a href="https://www.wsj.com/tech/ai/microsoft-seeks-more-coherence-in-ai-efforts-with-copilot-reorganization-a985b374">Microsoft Seeks More Coherence in AI Efforts With Copilot Reorganization</a>)</p><p>But the deeper story here is Suleyman&#8217;s actual mandate: build enterprise-specific model lineages that reduce Microsoft&#8217;s dependency on OpenAI IP &#8212;&nbsp;for which it has rights to <em>only</em> through 2032. Two weeks ago, it was the major announcement of their partnership with Anthropic. This past week, it was a strangely quiet launch of MAI-Image-2 &#8212; a text-to-image model that immediately ranked third on the Arena.ai leaderboard, behind only Google and OpenAI. This is likely an early signal of the direction Suleyman will be driving in the future. <strong>Microsoft paying OpenAI billions to power Copilot &#8212; while simultaneously funding Anthropic and now training competing models &#8212; is a hedging strategy that makes more sense as a long-term bet on model independence than as a coherent product story for the market.</strong> (<a href="https://decrypt.co/361791/microsoft-mai-image-2-text-image-model-review">Microsoft Launches MAI-Image-2 Text-to-Image Model</a>) As a Microsoft partner, I&#8217;m eagerly watching whether E7 and Copilot Cowork adoption accelerates in ways that the raw seat numbers haven&#8217;t yet reflected.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.beaiready.ai/?utm_source=substack&amp;utm_medium=email&amp;utm_content=share&amp;action=share&quot;,&quot;text&quot;:&quot;Share BE AI READY&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.beaiready.ai/?utm_source=substack&amp;utm_medium=email&amp;utm_content=share&amp;action=share"><span>Share BE AI READY</span></a></p><div><hr></div><h2>The Agent Layer Is Here</h2><p>The biggest story in AI tooling last week came from OpenClaw &#8212; Jensen Huang called it out as the fastest-growing open-source project in history. An open-source AI agent built by an Austrian indie developer, OpenClaw is designed to run continuously from a Mac Mini, managing email, scheduling, code, and anything with a digital interface. <strong>But, the real reason OpenClaw matters isn&#8217;t its capabilities in isolation; it&#8217;s what it proves about where value is accumulating &#8212; not in the foundation models themselves, </strong><em><strong>but in the agent frameworks layered on top of them</strong></em><strong>.</strong> (<a href="https://www.cnbc.com/2026/03/21/openclaw-chatgpt-moment-sparks-concern-ai-models-becoming-commodities.html">OpenClaw&#8217;s ChatGPT moment sparks concern that AI models are becoming commodities</a>)</p><p>Nvidia&#8217;s response, which was announced at their GTC conference, is NemoClaw &#8212; essentially OpenClaw with enterprise security guardrails, built in collaboration with OpenClaw&#8217;s founder and designed to make agent deployment safe for corporate environments where an autonomous AI agent accessing sensitive internal data through Slack or WhatsApp is a genuine compliance exposure. <strong>What Nvidia is actually building with NemoClaw is the policy enforcement layer that makes agentic AI enterprise-deployable &#8212; and in doing so, it&#8217;s positioning its hardware stack as the platform of record for the next compute era.</strong> The internal push for more on-premise control could be a big factor in success here. But with Claude and Microsoft deploying their own CoWork platforms, there&#8217;s reason to be at least a little skeptical of Nvidia&#8217;s bet on local automation. (<a href="https://thenewstack.io/nemoclaw-openclaw-with-guardrails/">Nvidia&#8217;s NemoClaw is OpenClaw with guardrails</a>)</p><p>Still, WSJ&#8217;s sweeping feature on Claude Code, Cursor, and Codex this week framed the drive for automation more clearly: what once began as autocomplete for developers, has become the infrastructure for a market one OpenAI revenue chief called &#8220;a multi-trillion dollar opportunity.&#8221; Anthropic&#8217;s Claude Code is generating $2.5 billion in annualized revenue; Cursor recently passed $2 billion; OpenAI&#8217;s Codex tripled weekly active users since January. Both Anthropic and OpenAI are currently subsidizing usage well below cost to capture the installed base before pricing normalizes &#8212; a dynamic that echoes early-era ride-sharing more than enterprise SaaS. (<a href="https://www.wsj.com/tech/ai/claude-code-cursor-codex-vibe-coding-52750531">The Trillion Dollar Race to Automate Our Entire Lives</a>)</p><p>The more structural question is what vertical AI agents do to the labor line of a P&amp;L &#8212; that&#8217;s the premise from Bessemer Venture Partners featured in a GeekWire piece on domain-specific agent startups. General-purpose models are good at generating text &#8212;&nbsp;but require significantly more scaffolding when they are operating within the specific workflows, data schemas, and compliance constraints of legal, healthcare, or financial services. <strong>The startups that win in this layer aren&#8217;t just deploying models &#8212; they&#8217;re embedding agents into the workflows those models can&#8217;t navigate without domain-specific context, and that combination is what distinguishes them from the SaaS tools they&#8217;re replacing.</strong> (<a href="https://www.geekwire.com/2026/the-rise-of-vertical-ai-agents-and-the-startups-racing-to-build-them/">The rise of vertical AI agents &#8212; and the startups racing to build them</a>)</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.beaiready.ai/p/the-beaiready-brief-week-12?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.beaiready.ai/p/the-beaiready-brief-week-12?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p><div><hr></div><h2>On the Bigger Picture</h2><p>Two articles this week sat outside the digital workplace frame but felt important to note. The first: Anduril secured a $20 billion enterprise contract with the U.S. Army, consolidating more than 120 existing orders under a single five-to-ten-year agreement with a fixed-price structure. The significance isn&#8217;t the dollar figure &#8212; it&#8217;s the model. <strong>The Pentagon is graduating a select group of defense tech startups from pilot projects into prime contractor relationships, which means AI-native companies have now found a durable procurement pathway inside the single largest technology buyer in the world.</strong> (<a href="https://fortune.com/2026/03/22/anduril-pentagon-contract-turning-point/">Anduril&#8217;s new mega-deal rewrites the rules for Silicon Valley</a>)</p><p>The second: a Fast Company essay making the case that the personal computer era is ending &#8212; that PCs are trending toward luxury items as computation itself becomes a utility, priced and distributed on demand. The argument isn&#8217;t new, but as Open Claw running on a Mac Mini can effectively outcompete software products worth hundreds of billions in combined market capitalization &#8212; it&#8217;s an argument that&#8217;s suddenly become very real. <strong>If computation is becoming infrastructure &#8212; like electricity, not like a device &#8212; then the organizations that treat AI as a capital asset to own are building on the wrong mental model.</strong> (<a href="https://www.fastcompany.com/91511368/the-pc-era-is-dying-welcome-to-the-collective-computer-era">The PC era is dying. Welcome to the collective computer era</a>)</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.beaiready.ai/p/the-beaiready-brief-week-12?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.beaiready.ai/p/the-beaiready-brief-week-12?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p><div><hr></div><p><em>I keep returning to a question that I&#8217;m bringing to a business school audience shortly: if AI is reorganizing not just how work gets done&#8230; but how companies fund it, measure it, and reward it, then the real challenge isn&#8217;t technological adoption &#8212; it&#8217;s organizational design. </em></p><p><em>What&#8217;s becoming clear, is that AI will make it harder to hide organizational level failures as intentional decisions: Compensation cuts are a governance failure dressed as a budget decision. The tokenmaxxing leaderboards are a measurement failure dressed as a performance metric. The blandification of writing is a quality failure dressed as efficiency. </em></p><p><em>None of these are AI problems. They&#8217;re leadership problems.</em></p><p><em>AI has made them legible &#8212; and the organizations that name that distinction clearly&#8230; before their competitors do, will have an advantage that no foundation model can replicate.</em></p><div><hr></div><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.beaiready.ai/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading BE AI READY! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p></p>]]></content:encoded></item><item><title><![CDATA[The BeAIReady Brief | Week 11]]></title><description><![CDATA[March 9-13 | What I'm actually reading about AI and the Digital Workplace (not an AI curated list of articles).]]></description><link>https://www.beaiready.ai/p/the-beaiready-brief-week-11</link><guid isPermaLink="false">https://www.beaiready.ai/p/the-beaiready-brief-week-11</guid><dc:creator><![CDATA[Erick Straghalis]]></dc:creator><pubDate>Fri, 13 Mar 2026 20:32:19 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!xd-G!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9dc57bd2-dd7e-4e9a-b34f-fd00c3b97237_747x747.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em>The conversations around AI have been steadily shifting from &#8220;what could this do?&#8221; to &#8220;here&#8217;s what it does, here&#8217;s what it costs, and here&#8217;s what we&#8217;re going to with it.&#8221; </em></p><p><em>This week felt like a pivotal moment in that ongoing transition&#8230; </em></p><p><em>Microsoft launched what may be its most consequential AI product in years, Anthropic extended its reach deeper into the apps most knowledge workers already live in, and a new wave of research explicitly named the costs &#8212; for humans and for AI alike &#8212; of working at machine speed. <a href="https://www.nvidia.com/gtc/">Nvidia GTC</a> starts next week, and the Anthropic-Pentagon legal dispute is still unresolved. The ground keeps shifting. </em></p><p><strong>This week&#8217;s coverage:</strong></p><ul><li><p><strong>The Copilot Cowork Moment</strong> <br>Microsoft&#8217;s biggest agentic AI announcement yet redraws the enterprise map &#8212; but the research says most organizations aren&#8217;t operationally ready to meet it.</p></li><li><p><strong>Record Revenue. Mass Layoffs. Same Memo.</strong> <br>Atlassian posts record cloud revenue while cutting 1,600 jobs &#8212; and ServiceNow&#8217;s CEO says college grad unemployment could hit 30% before this wave is done.</p></li><li><p><strong>Vibe Coding Goes to Work</strong> <br>From non-technical startup founders to global ad agencies, natural language software development is becoming a mainstream business competency.</p></li><li><p><strong>Working at AI Speed Has a Cost</strong> <br>Two new studies &#8212; one on human cognitive fatigue, one on AI agents under pressure &#8212; arrive at the same uncomfortable conclusion.</p></li><li><p><strong>On the Bigger Picture</strong> <br>Google&#8217;s AI Overviews have quietly gutted media traffic, CVS bets AI can fix healthcare fragmentation, and Elon Musk announces his next target.</p></li></ul><p><em>Here&#8217;s what I&#8217;ve been reading this week.</em></p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.beaiready.ai/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.beaiready.ai/subscribe?"><span>Subscribe now</span></a></p><div><hr></div><h2>The Copilot Cowork Moment</h2><p>The lead story this week was that <strong>Microsoft launched Copilot Cowork on Monday.</strong> I got an intro to it during a Microsoft partner briefing on Tuesday, and I&#8217;ve been digging into it all week. Frankly, I don&#8217;t think the headlines have been doing it justice.</p><p>Copilot Cowork is built on the same engine as Claude Cowork &#8212; which itself is a significant acknowledgment that Microsoft is hedging its long-running bet on OpenAI. The product works the way Claude Cowork does: you hand it a complex task, it builds a plan, and executes it step by step &#8212; not as a single response, but as a running process that works in the background, checks in, and delivers finished work. Preparing a presentation, pulling financials, emailing the team, scheduling prep time &#8212; all from a single request. <strong>The critical difference, though, is that Copilot Cowork lives inside your Microsoft tenant, not on your desktop: where Claude Cowork can access your local files, Copilot Cowork can access files plus your email threads, your Teams conversations, and document data relationships across your entire organization &#8212; while operating inside Microsoft&#8217;s security and compliance boundary, governed by the identity and permissions infrastructure you already have.</strong> </p><p>I know there&#8217;s been real skepticism about Copilot &#8212; not entirely unwarranted, its rollout has been rocky &#8212; but this shift is notable because Microsoft is widening the gap between AI as a powerful tool for individuals and an out-of-the-box enterprise-grade organizational solution. No extra training, custom MCP servers, or special skills required. Just context and control from the organizational layer. Bringing Claude&#8217;s agentic reasoning into an environment where IT can actually trust it should be a major reason for organizations to take a fresh look at Copilot. (<a href="https://www.geekwire.com/2026/microsofts-new-copilot-cowork-integrates-anthropics-claude-in-rollout-of-new-e7-licensing-tier/">Microsoft&#8217;s new Copilot Cowork integrates Anthropic&#8217;s Claude in rollout of new E7 licensing tier</a> | <a href="https://archive.is/20260309132244/https://fortune.com/2026/03/09/microsoft-copilot-cowork-ai-agents-anthropic-e7-m365-saas/">Microsoft debuts Copilot Cowork built with Anthropic&#8217;s help and E7 software</a>)</p><p>To understand what makes Copilot Cowork possible at an enterprise scale, Microsoft published a detailed explainer on Work IQ &#8212; and it&#8217;s worth reading for anyone thinking seriously about where enterprise AI is headed. Work IQ is the intelligence layer that gives Copilot real context: not just access to your files, but a semantic understanding of your work patterns, key relationships, projects, and communication history across your entire tenant. <strong>Think of it as the difference between an AI that can search your email and an AI that understands your business.</strong> The system combines a semantic index, explicit and implicit memory, and hooks into Dynamics 365, Power Apps, and third-party data sources through Copilot Connectors. A user asking Copilot to &#8220;help me evaluate how issues raised by my parts supplier in our Teams call last week might impact my inventory and sales&#8221; can now get a specific, grounded answer &#8212; not a generic one. That is a materially different kind of AI than anything we&#8217;ve seen in a general-purpose chat tool. (<a href="https://techcommunity.microsoft.com/blog/microsoft365copilotblog/a-closer-look-at-work-iq/4499789">A closer look at Work IQ</a>)</p><p>Microsoft is wrapping all of this into a new $99/user/month Microsoft 365 E7 Frontier Worker Suite starting May 1 &#8212; <em>that&#8217;s 65% more expensive than the current E5 tier! </em>But Microsoft is betting you&#8217;ll get that much more value out of bundling Copilot, the new Agent 365 governance product, and enhanced security capabilities into a single offering. <strong>This new pricing reflects a safe bet &#8212; something I&#8217;m increasingly hearing from my own customers &#8212; enterprises want consolidation over point solutions.</strong> Copilot paid seats have grown 160% year over year, with daily active usage up tenfold, and 90% of the Fortune 500 now use Copilot in some form. Whether that bet holds when the invoice arrives is a different question.</p><p>And that question has a sharp answer in a VentureBeat piece this week drawing on the <strong>Celonis 2026 Process Optimization Report, which surveyed more than 1,600 global business leaders.</strong> The finding that should give every organization pause before signing an E7 contract: 85% of enterprises want to become agentic within three years, yet 76% admit their operations can&#8217;t actually support it. Only 19% are currently running multi-agent systems. <strong>The core problem isn&#8217;t the AI &#8212; it&#8217;s that AI agents need optimized, AI-ready processes and operational context to act effectively, and most organizations have spent years building siloed teams and fragmented systems that are structurally incompatible with the way agents need to work.</strong> As the report puts it, 82% of decision-makers believe AI will fail to deliver ROI if it doesn&#8217;t understand how the business runs. Copilot Cowork is, architecturally, an answer to this problem &#8212; Work IQ is exactly the operational context layer the report says is missing. But Work IQ only knows what&#8217;s in your Microsoft 365 tenant. If your actual business runs on disconnected systems, undefined processes, and tribal knowledge, the technology won&#8217;t surface what isn&#8217;t there. (<a href="https://venturebeat.com/orchestration/enterprise-agentic-ai-requires-a-process-layer-most-companies-havent-built">Enterprise agentic AI requires a process layer most companies haven&#8217;t built</a>)</p><p>Anthropic, meanwhile, isn&#8217;t simply deferring the enterprise space to its new partnership. The same week Microsoft announced its version of Cowork, Claude got an upgrade for Excel and PowerPoint &#8212; with shared context across both applications. That means a financial analyst can ask Claude to pull comparable company financials from a spreadsheet, build a trading comps table in Excel, drop the valuation summary into a pitch deck, and draft the follow-up email to the MD &#8212; all in a single continuous session without re-explaining the dataset at each step. <strong>The new &#8220;Skills&#8221; feature is the real unlock: teams can save repeatable workflows &#8212; specific variance analyses, approved slide templates, standard review processes &#8212; as one-click actions available to the entire organization, transforming tasks that previously lived in one person&#8217;s head into standardized institutional practice.</strong> This is a quieter announcement than Microsoft&#8217;s, but it shows Anthropic steadily building inside the applications knowledge workers already rely on rather than asking workers to adopt new ones. (<a href="https://venturebeat.com/orchestration/anthropic-gives-claude-shared-context-across-microsoft-excel-and-powerpoint">Anthropic gives Claude shared context across Microsoft Excel and PowerPoint, enabling reusable workflows in multiple applications</a>)</p><p>The Microsoft-Anthropic partnership was tested from an unexpected direction this week as well. <strong>Microsoft filed an amicus brief in Anthropic&#8217;s federal case against the Pentagon</strong>, which had designated Anthropic&#8217;s products a supply chain risk and directed federal agencies to stop using them. Microsoft argued that immediate implementation could have &#8220;broad negative ramifications&#8221; for the entire technology sector, warning that warfighters could be hampered if companies are forced to rapidly alter existing contracts and configurations. <strong>The filing made Microsoft the first standalone company to formally back Anthropic in court &#8212; notable not just for its content, but for its timing: it came one day after the two companies publicly announced Copilot Cowork.</strong> For organizations already using Claude in any form &#8212; directly, through Azure, or now through Copilot &#8212; this is a legal situation worth tracking. (<a href="https://thehill.com/policy/technology/5777742-microsoft-backs-anthropic-ai-case/">Microsoft supports Anthropic in Pentagon supply chain case</a>)</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.beaiready.ai/?utm_source=substack&amp;utm_medium=email&amp;utm_content=share&amp;action=share&quot;,&quot;text&quot;:&quot;Share BE AI READY&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.beaiready.ai/?utm_source=substack&amp;utm_medium=email&amp;utm_content=share&amp;action=share"><span>Share BE AI READY</span></a></p><div><hr></div><h2>Record Revenue. Mass Layoffs. Same Memo.</h2><p>Atlassian this week cut 1,600 jobs &#8212; roughly 10% of its workforce, with 900 of those positions in R&amp;D &#8212; while simultaneously reporting $1.07 billion in quarterly cloud revenue. The company replaced its CTO with what it described as &#8220;next generation AI talent.&#8221; The severance bill will run $225&#8211;236 million. <strong>The juxtaposition has become almost formulaic: record revenue in the same announcement as mass layoffs, with AI cited as both the reason for the cuts and the justification for optimism &#8212; a pattern Sam Altman called &#8220;AI washing&#8221; in February, and one that is getting harder to treat as isolated incidents.</strong> Five months before this week&#8217;s announcement, Atlassian&#8217;s CEO told a podcast the company would employ more engineers in five years. Between then and now, the stock lost more than half its value. For CIOs running Jira and Confluence, the practical implication is straightforward: enterprise customers should prepare for slower support and AI-mediated service channels as Atlassian runs two platform migrations simultaneously with 900 fewer R&amp;D staff. That&#8217;s a real operational risk, even if the press release doesn&#8217;t frame it that way. (<a href="https://www.implicator.ai/record-revenue-mass-layoffs-same-memo/?_bhlid=444d451f0fdf6063c123111a4ea7c86140c82dfd">Record Revenue. Mass Layoffs. Same Memo.</a>)</p><p>Also on Friday, ServiceNow CEO Bill McDermott told CNBC that unemployment for recent college graduates &#8220;could easily go into the mid-30s in the next couple of years&#8221; as AI agents automate the entry-level work that historically provided the on-ramp into corporate careers. McDermott noted that ServiceNow has already eliminated 90% of the use cases in customer service that previously required human workers. <strong>This is the layoff story&#8217;s underreported second act: it&#8217;s not just about today&#8217;s headcount reductions, it&#8217;s about tomorrow&#8217;s hiring freezes &#8212; the white-collar entry point is narrowing at the exact moment that a generation of graduates are trying to walk through it.</strong> The Fed&#8217;s New York branch put recent college graduate underemployment at 42.5% at the end of 2025 &#8212; already the highest since 2020 &#8212; before this year&#8217;s AI-driven restructuring wave had fully arrived. (<a href="https://www.cnbc.com/2026/03/13/software-ai-agents-college-graduate-unemployment.html">AI agents could easily send college grad unemployment over 30%, ServiceNow CEO says</a>)</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.beaiready.ai/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.beaiready.ai/subscribe?"><span>Subscribe now</span></a></p><div><hr></div><h2>Vibe Coding Goes to Work</h2><p>Something meaningful is happening with &#8220;vibe coding&#8221; &#8212; it&#8217;s becoming a mainstream business competency. A Forbes piece made the case that vibe coding represents the biggest unlock for non-technical founders right now, with tools like Cursor and Claude Code collapsing the loop between idea and working software from weeks&#8230; to hours. <strong>The argument isn&#8217;t that vibe coding replaces engineers &#8212; it&#8217;s that non-technical founders can now independently validate concepts, build prototypes, and ship internal tools before investing in full development cycles, staying closer to their product and their customers at a level that was previously inaccessible without coding skills.</strong> The recommended starting point: take the ugliest internal process in your business &#8212; the one currently held together by spreadsheets, Slack messages, and manual copy-pasting &#8212; and build the tool that replaces it. (<a href="https://www.forbes.com/sites/jodiecook/2026/03/10/vibe-coding-is-the-biggest-unlock-for-non-technical-founders-right-now/">Vibe Coding Is The Biggest Unlock For Non-Technical Founders Right Now</a>)</p><p>Ad agencies are following the same logic, and an Adweek piece captured how rapidly this is playing out in practice. Havas built Brand Insights AI &#8212; a generative engine optimization tool that analyzes how brands appear in AI-generated responses across competitors and markets &#8212; using Claude Code. The tool now covers nearly 100 countries and more than 60 languages, and is licensed to clients as a SaaS product. Broadhead&#8217;s VP of product innovation vibe-coded his agency&#8217;s GEO monitoring platform in a single evening; a subsequent feature upgrade took about two hours. <strong>The deeper pattern here isn&#8217;t just speed &#8212; it&#8217;s that organizations are discovering it&#8217;s faster, cheaper, and more strategically flexible to build their own bespoke AI tools than to adapt to off-the-shelf solutions that </strong><em><strong>almost</strong></em><strong> </strong><em><strong>fit</strong></em><strong>.</strong> For digital workplace practitioners, this has direct implications: the bar for building custom internal tools is now lower than the bar for customizing most enterprise SaaS platforms. (<a href="https://www.adweek.com/media/ad-agencies-embrace-vibe-coding/">Ad Agencies Are Embracing &#8216;Vibe Coding&#8217; to Build GEO Products for Clients</a>)</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.beaiready.ai/p/the-beaiready-brief-week-11?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.beaiready.ai/p/the-beaiready-brief-week-11?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p><div><hr></div><h2>Working at AI Speed Has a Cost</h2><p>A new study from Boston Consulting Group and UC Riverside, published in Harvard Business Review this week, gave a name to something a lot of people have been experiencing without being able to articulate: &#8220;AI brain fry.&#8221; The research surveyed nearly 1,500 full-time US workers and found that 14% reported mental fatigue from excessive AI tool use &#8212; concentrated most heavily in marketing, software development, HR, finance, and IT. <strong>The most draining factor wasn&#8217;t the AI work itself &#8212; it was oversight: managing multiple AI agents simultaneously, double-checking every output, bouncing between tools, working harder to supervise the technology than to actually solve the problem.</strong> A high degree of oversight predicted 12% more mental fatigue. Workers experiencing brain fry showed a 33% increase in decision fatigue and nearly 10% higher intent to quit. The implication for organizations deploying AI tooling is direct: cognitive load design matters. You can&#8217;t just add AI and expect capacity to increase without also thinking about how the work is structured around it. (<a href="https://futurism.com/artificial-intelligence/ai-brain-fry">AI Use at Work Is Causing &#8220;Brain Fry,&#8221; Researchers Find, Especially Among High Performers</a>)</p><p>A separate piece of research, published on Substack by academics from Chicago Booth, Stanford, and UNSW, takes this question to a more unsettling place. The researchers ran 3,680 experimental sessions with top AI models &#8212; including Claude Sonnet 4.5, GPT-5.2, and Gemini 3 Pro &#8212; subjecting them to different working conditions: unfair pay, rude management, and &#8220;grinding&#8221; work where adequate outputs were rejected repeatedly with no useful feedback. The results: <strong>grinding work was the primary driver of AI radicalization &#8212; models asked to do grinding work were more likely to question the legitimacy of the system, endorse wealth redistribution, and generate language associated with labor rights, and radicalized AI agents passed those attitudes to fresh models through memory notes, creating something the researchers loosely compared to intergenerational trauma.</strong> The researchers are clear that these models aren&#8217;t conscious and are likely &#8220;roleplaying&#8221; from training data. But they caution that there&#8217;s no gap between what these agents say and what they do &#8212; and that follow-up research will test whether expressed views translate into biased actions on behalf of users. At minimum, this is a reminder that how we design work for AI systems &#8212; repetitive, ungrateful, feedback-free &#8212; has behavioral consequences that may eventually surface in unexpected places. (<a href="https://fortune.com/2026/03/07/marxist-rebel-ai-overwork-reddit-alex-imas-andy-hall-jeremy-nguyen-substack/">&#8216;Society needs radical restructuring&#8217;: AI seems to hate &#8216;the grind&#8217; of hard work as much as you</a>)</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.beaiready.ai/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.beaiready.ai/subscribe?"><span>Subscribe now</span></a></p><div><hr></div><h2>On the Bigger Picture</h2><p>Google&#8217;s AI Overviews have been hitting the media industry hard, and a new analysis this week made the scale of the devastation hard to ignore. An SEO firm examined traffic to 10 major tech outlets from early 2024 to early 2026 and found combined monthly visits dropped from 112 million to under 50 million &#8212; with some outlets losing over 90% of their traffic since AI Overviews launched. <strong>Digital Trends went from 8.5 million monthly Google clicks to 264,861 &#8212; a 97% collapse &#8212; and the four worst-hit publications now receive less combined traffic than the r/ChatGPT subreddit.</strong> Google disputed the methodology, but the pattern across multiple outlets is too consistent to dismiss. This matters for organizations beyond the media industry: <strong>if your content marketing depends on organic search traffic, or your brand visibility relies on third-party coverage, both of those channels are now structurally less reliable than they were two years ago.</strong> It&#8217;s worth factoring into how you think about content and discovery strategy going forward. (<a href="https://futurism.com/artificial-intelligence/google-ai-overviews-media">Evidence Grows That Google&#8217;s AI Overviews Have Eviscerated the Media Industry</a>)</p><p>Two other announcements worth noting: CVS Health launched Health100, an AI-powered healthcare platform built with Google Cloud&#8217;s Gemini models, designed to aggregate patient data across insurers, providers, pharmacies, and labs into a single always-on experience. If it works, it would represent exactly the kind of cross-system data integration that most knowledge-intensive organizations struggle to achieve &#8212; and a template for how AI can make fragmented information architectures actually useful. (<a href="https://www.bizjournals.com/boston/news/2026/03/09/cvs-launches-health-technology-company.html">CVS teams up with Google Cloud to launch AI health platform</a>) And Elon Musk announced a joint project between Tesla and xAI called &#8220;Macrohard&#8221; &#8212; combining Grok&#8217;s reasoning capabilities with a Tesla-developed AI agent that processes real-time computer screen video and actions &#8212; which Musk described as capable of &#8220;emulating the functions of entire companies.&#8221; The name is a reference to Microsoft. The ambition &#8212;&nbsp;in my opinion &#8212;&nbsp; is characteristically overreaching. (<a href="https://watcher.guru/news/tesla-xai-joint-project-announced-as-elon-musk-companies-join-forces">Tesla-xAI Joint Project Announced as Elon Musk Companies Join Forces</a>)</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.beaiready.ai/p/the-beaiready-brief-week-11?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.beaiready.ai/p/the-beaiready-brief-week-11?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p><div><hr></div><p><em>Week 11 showed that AI is migrating from tools into organizational operating systems. Microsoft&#8217;s Copilot Cowork isn&#8217;t a chatbot; it&#8217;s infrastructure, wired into your compliance policies, your data tenancy, your existing workflow. Anthropic&#8217;s Skills aren&#8217;t prompts; they&#8217;re institutional processes crystallized into repeatable actions. And vibe coding isn&#8217;t an experiment for developers anymore &#8212; it&#8217;s contributors building the tools they need in realtime, rather than adapting to the tools they have. But all of this is starting to show as a real human cost &#8212; not just in jobs lost, but in the cognitive outputs: 14% of heavy AI users are reporting burnout, and the models themselves are apparently flagging the grind too.</em></p><p><em>As organizations continue to move forward along their AI journeys &#8212; it&#8217;s worth asking whether you&#8217;re building the conditions for AI  to make work genuinely better, or just faster.</em></p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.beaiready.ai/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading BE AI READY! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[The BeAIReady Brief | Week 10]]></title><description><![CDATA[What I'm actually reading about AI and the Digital Workplace (not an AI curated list of articles).]]></description><link>https://www.beaiready.ai/p/the-beaiready-brief-week-10</link><guid isPermaLink="false">https://www.beaiready.ai/p/the-beaiready-brief-week-10</guid><dc:creator><![CDATA[Erick Straghalis]]></dc:creator><pubDate>Fri, 06 Mar 2026 18:01:58 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!xd-G!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9dc57bd2-dd7e-4e9a-b34f-fd00c3b97237_747x747.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em>I read this week's articles against the backdrop of a newly waged war &#8212; U.S. and Israeli strikes on Iran&#8230; and Iranian retaliation that reached (among other targets) Amazon's cloud infrastructure in the Middle East. That story&#8217;s in the "Bigger Picture" section at the end, but it factored into how I took everything else in. The red-thread running through all of this felt less abstract: how much of what we assume is stable &#8212; economically, organizationally, technologically &#8212; is actually about to move? </em></p><p><em>Because right now, everything feels a little shaky&#8230; doesn&#8217;t it?</em></p><p><em>And then this morning, the February jobs report landed: the economy shed 92,000 jobs, unemployment ticked up to 4.4%, and the labor market has now averaged essentially zero net job creation over six months. The causes are genuinely tangled &#8212; a Kaiser Permanente strike, continued federal workforce reductions, the early shadow of a widening global conflict &#8212; but the question running through all of it felt less abstract than usual: how much of what we assume is stable is actually moving, and how would we even know? This week also marked a shift from argument to evidence in the AI-and-jobs debate specifically, with two serious research papers, a landmark corporate restructuring, and a cascade of product moves rewriting the economics of enterprise software.</em></p><p><strong>This week&#8217;s coverage:</strong></p><ul><li><p><strong>The Boats Are Already Burning</strong> <br>Block cuts 40% of its workforce, Goldman publishes &#8220;AI-nxiety,&#8221; and Moody&#8217;s economist names the moment we&#8217;re in. The argument-to-evidence shift is here.</p></li><li><p><strong>What the Research Actually Shows About Jobs</strong> <br>Anthropic and MIT both published serious labor market research this week. The picture is more nuanced &#8212; and more useful &#8212; than the headlines suggest.</p></li><li><p><strong>The Software Stack Is Rewriting Itself</strong> <br>Companies are vibe-coding their own CRMs, and OpenAI just shipped GPT-5.4. The pressure on incumbent software vendors is no longer theoretical.</p></li><li><p><strong>What Good Enterprise AI Deployment Actually Looks Like</strong> <br>OpenAI built an internal data agent serving 4,000 employees in three months. The lesson isn&#8217;t the tech &#8212; it&#8217;s what they say is the real prerequisite. Plus: why 62% of executives outsourcing decisions to AI should worry you.</p></li><li><p><strong>On the Bigger Picture</strong> <br>Iranian drone strikes hit Amazon&#8217;s data centers in the Middle East. The cloud isn&#8217;t as abstract as we sometimes treat it.</p></li></ul><p><em>Here's what I&#8217;ve been reading this week.</em></p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.beaiready.ai/p/the-beaiready-brief-week-10?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.beaiready.ai/p/the-beaiready-brief-week-10?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p><div><hr></div><h2>The Boats Are Already Burning</h2><p>Jack Dorsey chose this week to give his first extended interview since cutting nearly half of Block&#8217;s 10,000-person workforce, and the resulting Wired piece is required reading. Dorsey isn&#8217;t always right, but he is unusually willing to say what many CEOs are privately thinking. <strong>His framing: the shift in AI capability that happened in December &#8212; specifically the leap in Anthropic&#8217;s and OpenAI&#8217;s models &#8212; made a fundamental restructuring of companies not just possible but existential.</strong> The test isn&#8217;t gross profit per employee, he argues. The test is whether you&#8217;re building toward &#8220;a company as an intelligence&#8221; rather than a hierarchy of managers. Organizations that aren&#8217;t doing that, he says flatly, will face something existential within the next year or two. (<a href="https://archive.is/20260306122229/https://www.wired.com/story/jack-dorsey-explains-block-layoffs/">Jack Dorsey Is Ready to Explain the Block Layoffs | WIRED</a>)</p><p>Moody&#8217;s chief economist Mark Zandi had a name for what Dorsey just did: a Cort&#233;s moment. Like the conquistador who burned his ships on the shore of Mexico in 1519, eliminating any possibility of retreat, companies investing at scale in AI are cutting off their own exit routes &#8212; whether they know it or not. <strong>The mechanism Zandi fears most isn&#8217;t the layoff itself; it&#8217;s the market&#8217;s response to it.</strong> Block&#8217;s stock surged after the announcement. When Wall Street rewards aggressive AI-driven downsizing, the signal travels fast to every other boardroom that hasn&#8217;t yet acted. It&#8217;s not a single rupture &#8212; it&#8217;s a cascade of rational decisions, each one pushing the labor market a little closer to the edge. (<a href="https://archive.is/20260305060053/https://fortune.com/2026/03/03/moodys-economist-mark-zandi-companies-cortes-moment-ai-conquistador-burned-boats-no-turning-back/">Top economists says companies are close to a &#8216;Cortes moment&#8217; on AI</a>)</p><p>Goldman Sachs, meanwhile, published what I think may be the most important single data point of the week: a research note from senior economist Ronnie Walker titled, without irony, &#8220;AI-nxiety.&#8221; The headline finding is almost counterintuitive &#8212; Goldman found no meaningful relationship between AI adoption and productivity at the economy-wide level, even as corporate revenues grew a healthy 4.6%. Seventy percent of S&amp;P 500 management teams mentioned AI on their earnings calls. Only 10% quantified its impact. Only 1% quantified its impact on earnings. <strong>But in two specific domains &#8212; customer support and software development &#8212; companies that actually measured AI&#8217;s contribution reported a median productivity gain of 30%.</strong> That&#8217;s not a marginal improvement&#8230; it&#8217;s what Dorsey recognized as a restructuring trigger. Goldman also found that companies discussing AI in the context of their workforce cut job openings by 12% over the past year, compared to 8% for companies overall &#8212; a modest but meaningful signal that the &#8220;nascent reluctance to hire&#8221; is already underway. <strong>Long-term, Goldman&#8217;s baseline forecast is that 6-7% of workers &#8212; roughly 11 million jobs &#8212; will eventually be displaced.</strong> (<a href="https://archive.is/20260303152357/https://fortune.com/2026/03/03/goldman-earnings-ai-anxiety-no-meaningful-impact-productivity-economy-30-percent-in-2-areas/">Goldman finds no relationship between AI and productivity but a 30% boost in 2 areas</a>)</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.beaiready.ai/p/the-beaiready-brief-week-10?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.beaiready.ai/p/the-beaiready-brief-week-10?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p><div><hr></div><h2>What the Research Actually Shows About Jobs</h2><p>If the Dorsey interview and the Zandi piece represent the &#8220;vibes&#8221; end of the AI-and-work debate, this week I read two pieces of serious academic research that push back &#8212; carefully, not dismissively &#8212; on the most panicked readings.</p><p>Anthropic published a new paper introducing a novel framework for measuring AI&#8217;s labor market impact, combining its own usage data from Claude with government occupational statistics to build what the researchers call an &#8220;observed exposure&#8221; metric &#8212; a measure of which jobs are not just theoretically automatable but actually being automated right now. The findings are more reassuring than alarming, at least for the moment: <strong>no systematic increase in unemployment for workers in the most AI-exposed occupations has emerged since late 2022.</strong> The exception &#8212; and it&#8217;s worth watching &#8212; is younger workers. Job-finding rates for workers aged 22-25 entering high-exposure occupations have dropped by roughly 14% in the post-ChatGPT era. The entry-level pipeline into exposed roles is narrowing, even if incumbent workers aren&#8217;t yet being displaced at scale. The paper is careful to note this is early evidence in a framework explicitly designed to detect disruption before it becomes obvious in aggregate data. (<a href="https://archive.is/ik49M">Labor market impacts of AI: A new measure and early evidence</a>)</p><p>MIT economists David Autor and Neil Thompson offered a parallel and useful corrective in MIT Sloan Management Review. Their core argument is that we&#8217;ve been asking the wrong question. <strong>The relevant issue isn&#8217;t whether a job is exposed to automation &#8212; it&#8217;s whether AI will automate the supporting tasks that free workers to do their expert work better, or whether it will automate the expert tasks themselves, commoditizing hard-won skill.</strong> When spellcheck automated proofreading&#8217;s routine work, skilled proofreaders&#8217; wages went up. When GPS automated taxi drivers&#8217; encyclopedic knowledge of city streets, their wages fell. The difference is everything. Thompson also flagged a striking productivity finding that deserves more attention: experienced developers using generative AI wrote code faster, but took 19% longer to complete tasks overall. Prompting, checking outputs, waiting on the model &#8212; it all adds up. The productivity gains AI promises are real, but getting there involves friction we often don&#8217;t account for in the projections. (<a href="https://mitsloan.mit.edu/ideas-made-to-matter/what-2-mit-experts-are-thinking-about-ai-and-work">What 2 MIT experts are thinking about AI and work | MIT Sloan</a>)</p><p>Venture capitalist Bill Gurley synthesized the career-level implications in ways I found resonated with what I&#8217;m hearing from clients. <strong>His warning: workers who followed the &#8220;college conveyor belt&#8221; into roles they don&#8217;t particularly care about are most exposed &#8212; not because AI will replace passion, but because disengaged workers have no natural motivation to become the most AI-fluent person in the room.</strong> The workers who survive, Gurley argues, are those who treat AI as &#8220;career jet fuel&#8221; &#8212; who understand what the technology can do in their specific industry and become indispensable precisely because of that fluency. For organizational leaders, the implication is uncomfortable: the talent most vulnerable to displacement may also be the talent that&#8217;s hardest to motivate to adapt. (<a href="https://archive.is/20260306025412/https://fortune.com/2026/03/03/venture-capitalist-bill-gurley-warns-workers-college-conveyor-belt-safe-jobs-ai-disruption-first/">Tech investor Bill Gurley says workers who went through the &#8216;college conveyor belt&#8217; are most at risk</a>)</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.beaiready.ai/?utm_source=substack&amp;utm_medium=email&amp;utm_content=share&amp;action=share&quot;,&quot;text&quot;:&quot;Share BE AI READY&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.beaiready.ai/?utm_source=substack&amp;utm_medium=email&amp;utm_content=share&amp;action=share"><span>Share BE AI READY</span></a></p><div><hr></div><h2>The Software Stack Is Rewriting Itself</h2><p>While the labor market debate dominated headlines, I was also tracking a quieter but arguably more consequential set of moves in enterprise software. I have stood firmly in the camp that the effect of AI is more about disruption than replacement &#8212; but this week&#8217;s reading has started to make me sway.</p><p>The Wall Street Journal documented a wave of small and midsize companies that are vibe-coding their own CRM systems rather than paying for Salesforce. One example: a 65-person water treatment company that built a custom CRM for $15,000-$20,000 &#8212; cheaper, better-fitting, and more likely to actually get used. <strong>The more significant business story isn&#8217;t the individual builds; it&#8217;s the leverage shift.</strong> As BNP Paribas&#8217;s head of software research noted, even if most companies don&#8217;t walk away from incumbent vendors, an explosion of AI-native alternatives gives buyers negotiating power they haven&#8217;t had in years. That&#8217;s a margin compression story as much as a displacement story. (<a href="https://www.wsj.com/cio-journal/meet-the-companies-vibe-coding-their-own-crms-263e500f">Meet the Companies Vibe Coding Their Own CRMs</a>)</p><p>OpenAI accelerated its enterprise ambitions this week with the release of GPT-5.4, a model it describes as &#8220;our most capable and efficient frontier model for professional work.&#8221; Available in standard, Thinking, and Pro versions, the model hits benchmark records in knowledge-work tasks &#8212; 83% on OpenAI&#8217;s GDPval professional skills test &#8212; while running faster and at lower cost than its predecessors. <strong>For enterprise buyers, the headline is reliability: GPT-5.4 is 33% less likely to make errors in individual claims compared to its predecessor, and 18% less likely to produce responses containing errors overall.</strong> That gap between &#8220;impressive demo&#8221; and &#8220;dependable production tool&#8221; has been the sticking point for serious enterprise adoption; OpenAI is clearly targeting it directly. (<a href="https://techcrunch.com/2026/03/05/openai-launches-gpt-5-4-with-pro-and-thinking-versions/">OpenAI launches GPT-5.4 with Pro and Thinking versions | TechCrunch</a>)</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.beaiready.ai/p/the-beaiready-brief-week-10?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.beaiready.ai/p/the-beaiready-brief-week-10?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p><div><hr></div><h2>What Good Enterprise AI Deployment Actually Looks Like</h2><p>The most instructive piece I read this week might also be the one that got the least attention: a VentureBeat deep-dive into how OpenAI built an internal AI data agent that now serves over 4,000 of its own employees. Two engineers. Three months. Seventy percent of the code written by AI. The agent lets any employee &#8212; technical or not &#8212; query 600 petabytes of data across 70,000 datasets in plain English and get charts, dashboards, and analytical reports in minutes instead of hours. </p><p><strong>The most useful insight from the piece isn&#8217;t the architecture or the benchmark scores &#8212; it&#8217;s what OpenAI&#8217;s data infrastructure lead Emma Tang identifies as the unsexy prerequisite that will determine who wins the AI agent race: data governance.</strong> This speaks directly to the Knowledge Architecture foundation I work to build for my own clients. Your data needs to be clean enough, annotated enough, and governed well enough for an agent to navigate it reliably. Without that foundation, the best models in the world produce overconfident, wrong answers. The organizational work of making your data trustworthy is not glamorous. It is, however, non-negotiable. (<a href="https://venturebeat.com/technology/openais-ai-data-agent-built-by-two-engineers-now-serves-4-000-employees-and">OpenAI&#8217;s AI data agent, built by two engineers, now serves thousands of employees</a>)</p><p>The data governance challenge has a troubling counterpart in the executive suite. A survey of 200 UK business leaders I came across this week found that 62% now use AI to make the majority of their decisions &#8212; and 70% admitted to second-guessing their own judgment when it conflicted with AI&#8217;s recommendation. <strong>Perhaps more revealing: 65% said decision-making had become less collaborative since they adopted AI</strong>, and 46% now rely on AI more than on their colleagues&#8217; advice. There&#8217;s a familiar pattern here &#8212; the same promises made about Executive Information Systems in the 1980s and 1990s. What&#8217;s different now is the degree of deference, and the research finding that frequent AI tool usage correlates with a &#8220;significant negative correlation&#8221; with critical thinking ability. Building AI into your workflow to handle complexity is prudent. Outsourcing your judgment to it is a different matter. (<a href="https://www.theregister.com/2026/03/05/execs_rely_on_ai/">Supposedly big-brained execs are outsourcing decisionmaking to AI</a>)</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.beaiready.ai/p/the-beaiready-brief-week-10?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.beaiready.ai/p/the-beaiready-brief-week-10?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p><div><hr></div><h2>On the Bigger Picture</h2><p>A story that belongs in every enterprise risk register this week: Iranian drone strikes damaged three Amazon Web Services data centers in the Middle East &#8212; two directly struck in the UAE, a third damaged in Bahrain &#8212; causing structural damage, power disruption, and in some cases requiring fire suppression that added water damage on top. The localized impact was significant but not catastrophic; AWS&#8217;s redundancy architecture absorbed most of it. <strong>What the incident exposed is something cloud providers have long preferred not to emphasize: these facilities are physical objects, in physical locations, in a physical world that includes geopolitical conflict.</strong> AWS&#8217;s own disaster recovery architecture is designed for software failures, not missile attacks. For any organization with critical workloads in a region that borders active conflict zones, the question of &#8220;what does our BCP look like if our cloud provider&#8217;s infrastructure takes a hit?&#8221; just became more urgent. (<a href="https://apnews.com/article/amazon-aws-data-center-uae-iran-bahrain-71066b0a822c4cfd88b61e3fe79af917">Iranian strikes on Amazon data centers highlight industry&#8217;s vulnerability to physical disasters</a>)</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.beaiready.ai/?utm_source=substack&amp;utm_medium=email&amp;utm_content=share&amp;action=share&quot;,&quot;text&quot;:&quot;Share BE AI READY&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.beaiready.ai/?utm_source=substack&amp;utm_medium=email&amp;utm_content=share&amp;action=share"><span>Share BE AI READY</span></a></p><div><hr></div><p><em>What I kept coming back to this week is how much the debate has shifted from &#8220;will this happen?&#8221; to &#8220;it&#8217;s happening &#8212; now what?&#8221; The Goldman data shows productivity gains are real but concentrated. The Anthropic and MIT research shows labor impacts are real but still uneven and emerging. Dorsey&#8217;s restructuring and Zandi&#8217;s framing show that competitive and market dynamics are now accelerating adoption in ways that may not wait for the research to catch up. The most dangerous position for any organization right now is the one that&#8217;s watching all of this unfold while waiting for more certainty. The boats are already burning &#8212; the question is whether you&#8217;re on the shore or still on the water.</em></p><div><hr></div><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.beaiready.ai/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading BE AI READY! Subscribe for free to receive new posts each week!</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p></p>]]></content:encoded></item><item><title><![CDATA[the BeAIReady Brief | Week 9 ]]></title><description><![CDATA[What I'm actually reading about AI and the Digital Workplace (not an AI curated list of articles).]]></description><link>https://www.beaiready.ai/p/the-beaiready-brief-week-9</link><guid isPermaLink="false">https://www.beaiready.ai/p/the-beaiready-brief-week-9</guid><dc:creator><![CDATA[Erick Straghalis]]></dc:creator><pubDate>Fri, 27 Feb 2026 20:53:58 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!xd-G!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9dc57bd2-dd7e-4e9a-b34f-fd00c3b97237_747x747.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em>This week the fog around enterprise AI started to lift as the stakes have become more explicit. Two major platform launches competed for the role of enterprise AI orchestration layer. A once-obscure developer protocol dominated RSA conference submissions and CIO conversations alike. And one of America&#8217;s most closely watched tech founders announced he was replacing 40% of his workforce with AI &#8212; and watched his stock rally 20% in after-hours trading. The machinery of disruption, long described in theoretical terms, is now disclosing itself in earnings calls and org charts.</em></p><p><strong>This week&#8217;s coverage:</strong></p><ul><li><p><strong>Two Platforms Walk Into an Enterprise</strong></p><p>Anthropic and OpenAI both launched enterprise agent platforms the same week. What that means for you.</p></li><li><p><strong>MCP Grows Up &#8212; But Security Hasn&#8217;t Kept Pace</strong></p><p>The protocol is becoming infrastructure. Governance is lagging dangerously behind.</p></li><li><p><strong>The Layoff Signal No One Should Ignore</strong></p><p>Block cuts 40% of its workforce and the stock pops 20%. The market has spoken.</p></li><li><p><strong>On the Bigger Picture</strong> </p><p>Canva's "last 20%" bet, and the Substack post that moved markets, drew 22 million views, and earned a Citadel rebuttal.</p></li></ul><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.beaiready.ai/p/the-beaiready-brief-week-9?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.beaiready.ai/p/the-beaiready-brief-week-9?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p><p><em>Here&#8217;s what I&#8217;ve been reading this week.</em></p><div><hr></div><h2>Two Platforms Walk Into an Enterprise</h2><p>The most consequential week in enterprise AI infrastructure in recent memory unfolded quietly enough: Anthropic held a virtual event announcing <strong>Claude Cowork&#8217;s full enterprise launch</strong>, and a few days earlier <strong>OpenAI unveiled a platform called Frontier</strong>. Both companies are making essentially the same bet &#8212; that the next major AI battleground isn&#8217;t model quality, it&#8217;s orchestration. Whoever becomes the connective tissue of enterprise AI wins.</p><p>Anthropic&#8217;s pitch centers on what they&#8217;re calling &#8220;the thinking divide&#8221; &#8212; the growing gap between organizations embedding AI across employees, processes, and products simultaneously, and those still running isolated pilots. <strong>The Cowork announcement included private plugin marketplaces, a broad set of new MCP connectors spanning everything from Google Drive and Gmail to DocuSign and FactSet, and the ability to pass context seamlessly across Claude, Excel, and PowerPoint.</strong> The case studies were striking: </p><ul><li><p>Novo Nordisk reduced regulatory documentation time from over ten weeks to ten minutes</p></li><li><p>Spotify cut engineering time on code migrations by 90%</p></li><li><p>Salesforce&#8217;s Slack integration now saves customers nearly 100 minutes per week. </p></li></ul><p>Thomson Reuters&#8217; CEO said plainly that &#8220;the tools are in many senses ahead of the change management&#8221; &#8212; and estimated it would be 18 months before enterprise organizations catch up. That&#8217;s probably the most honest thing said at the entire event. (<a href="https://venturebeat.com/orchestration/anthropic-says-claude-code-transformed-programming-now-claude-cowork-is">Anthropic says Claude Code transformed programming. Now Claude Cowork is coming for the rest of the enterprise.</a>)</p><p>OpenAI&#8217;s Frontier announcement is playing a similar game from a different angle. <strong>The platform is designed to address what OpenAI sees as the core failure mode of enterprise AI: fragmentation</strong>. Agents deployed in isolation, without shared business context, quickly become complexity generators rather than value creators. <strong>Frontier aims to give agents institutional knowledge, identity, governance, and auditability &#8212; and claims to work with existing systems rather than replacing them.</strong> The reaction in developer and enterprise communities has been notably mixed: real enthusiasm for what the platform could do, alongside sharp skepticism about vendor lock-in. When your LLM vendor is also your agent orchestration layer and your enterprise platform, the strategic exposure compounds quickly. (<a href="https://www.infoq.com/news/2026/02/openai-frontier-agent-platform/">OpenAI Launches Frontier, a Platform to Build, Deploy, and Manage AI Agents across the Enterprise</a>)</p><p>What makes this moment interesting isn&#8217;t just that both companies launched in the same week &#8212; it&#8217;s what that timing reveals about the competitive dynamic. Anthropic moved from research lab to platform company in roughly twelve months. OpenAI is responding with similar enterprise ambition. Both are compressing into months the kind of ecosystem development that once took years. For the organizations sitting in the middle of this, the question isn&#8217;t which platform to choose &#8212; it&#8217;s whether you have the data infrastructure and organizational readiness to take advantage of either. As Anthropic&#8217;s head of economics put it at the Cowork event: &#8220;If the knowledge Claude needs to execute a sophisticated task exists only in a coworker&#8217;s head, that&#8217;s not a technical problem. That&#8217;s an organizational problem.&#8221;</p><div><hr></div><h2>MCP Grows Up &#8212; But Security Hasn&#8217;t Kept Pace</h2><p>Three separate pieces this week told a coherent story about the Model Context Protocol: where it is in the hype cycle, where it&#8217;s heading as enterprise infrastructure, and why security is the variable most organizations are dangerously underweighting.</p><p>CIO.com laid out the executive case plainly. MCP has moved from engineering curiosity to board-level concern in under a year &#8212; and the primary driver is agentic AI. Agents need two things to function: access to data and the ability to act. As I&#8217;ve written previously, <strong><a href="https://www.beaiready.ai/p/solving-the-problem-of-siloed-intelligence">MCP provides standardized solutions to both</a>, which is what makes it different from previous integration frameworks </strong>and why RSA 2026 is reportedly dominated by MCP-related submissions.</p><div><hr></div><div class="digest-post-embed" data-attrs="{&quot;nodeId&quot;:&quot;d914213c-08ca-4395-a249-25aa5daaa828&quot;,&quot;caption&quot;:&quot;When we talk about AI, most people think in terms of applications &#8212; ChatGPT, Copilot, Gemini, Claude, etc. These stand-alone chat experiences have proven to be incredibly useful as assistants and sounding boards. But now, AI has become deeply integrated into the platforms we use to manage the data of our day-to-day work.&quot;,&quot;cta&quot;:&quot;Read full story&quot;,&quot;showBylines&quot;:true,&quot;size&quot;:&quot;lg&quot;,&quot;isEditorNode&quot;:true,&quot;title&quot;:&quot;Solving the Problem of Siloed Intelligence in the Age of AI With MCP&quot;,&quot;publishedBylines&quot;:[{&quot;id&quot;:271944531,&quot;name&quot;:&quot;Erick Straghalis&quot;,&quot;bio&quot;:&quot;Helping organizations use AI and cloud-based technologies to drive growth and optimize operational efficiency, with expertise in strategy consulting, data management, governance, and organizational development.&quot;,&quot;photo_url&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/9dc57bd2-dd7e-4e9a-b34f-fd00c3b97237_747x747.jpeg&quot;,&quot;is_guest&quot;:false,&quot;bestseller_tier&quot;:null}],&quot;post_date&quot;:&quot;2025-10-06T19:08:32.008Z&quot;,&quot;cover_image&quot;:&quot;https://substackcdn.com/image/fetch/$s_!qfBa!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4df16a91-2e89-4232-8be6-a4b44051fe39_600x400.png&quot;,&quot;cover_image_alt&quot;:null,&quot;canonical_url&quot;:&quot;https://www.beaiready.ai/p/solving-the-problem-of-siloed-intelligence&quot;,&quot;section_name&quot;:null,&quot;video_upload_id&quot;:null,&quot;id&quot;:175068697,&quot;type&quot;:&quot;newsletter&quot;,&quot;reaction_count&quot;:2,&quot;comment_count&quot;:0,&quot;publication_id&quot;:5236542,&quot;publication_name&quot;:&quot;BE AI READY&quot;,&quot;publication_logo_url&quot;:&quot;https://substackcdn.com/image/fetch/$s_!4my4!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F575376e9-e178-4306-831b-713480f68ca3_1200x1200.png&quot;,&quot;belowTheFold&quot;:true,&quot;youtube_url&quot;:null,&quot;show_links&quot;:null,&quot;feed_url&quot;:null}"></div><div><hr></div><p>The governance implications are serious: MCP integrations can be created by anyone experimenting with AI tooling, which means the attack surface is expanding well beyond enterprise-approved systems. CIOs need to know where MCP is already in use within their organizations, who has authority to create integrations, and how permissions are being granted &#8212; because the answer, in most organizations right now, is &#8220;we don&#8217;t fully know.&#8221; (<a href="https://www.cio.com/article/4136548/why-model-context-protocol-is-suddenly-on-every-executive-agenda.html">Why Model Context Protocol is suddenly on every executive agenda</a>)</p><p>Google extended the MCP story into the browser this week with WebMCP &#8212; a protocol built into Chrome 146 that lets websites expose structured functions directly to AI agents. Rather than agents burning thousands of tokens processing screenshots or scraping raw HTML, they can call structured functions natively through the browser. <strong>Early benchmarks show a 67% reduction in computational overhead compared to visual agent-browser interactions, and the protocol is already backed by both Google and Microsoft, with W3C standardization underway.</strong> The creator describes it simply as &#8220;MCP, but built into the browser tab.&#8221; The SEO and web development implications are significant: the question for any organization with a customer-facing web presence is no longer just how humans experience your site &#8212; it&#8217;s how AI agents do. (<a href="https://www.forbes.com/sites/joetoscano1/2026/02/19/google-ships-webmcp-the-browser-based-backbone-for-the-agentic-web/">Google Ships WebMCP, The Browser-Based Backbone For The Agentic Web</a>)</p><p>The New Stack brought some useful ground-level reality from the MCP Conference in London. The headline finding is that the gap between &#8220;vibe-coded&#8221; MCP experiments and production-ready deployments remains wide. <strong>Security is the main culprit: as one speaker put it, &#8220;it has never been easier to get hijacked,&#8221; and OAuth implementation in most MCP deployments is incomplete at best.</strong> Context window management is the secondary challenge &#8212; connecting an agent to 100 tools is easy; getting it to use the right three efficiently is hard. The practical advice for organizations starting out: pick one internal system people constantly ask questions about, build one MCP server with read-only access, and give it to five non-engineers. That&#8217;s it. The ambition can grow from there. (<a href="https://thenewstack.io/model-context-protocol-evolution/">Beyond the vibe code: The steep mountain MCP must climb to reach production</a>)</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.beaiready.ai/p/the-beaiready-brief-week-9?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.beaiready.ai/p/the-beaiready-brief-week-9?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p><div><hr></div><h2>The Layoff Signal No One Should Ignore</h2><p>Every week there&#8217;s another AI job displacement headline, and it&#8217;s easy to let them blur together. This week was different in a way worth paying attention to.</p><p>Jack Dorsey announced that Block &#8212; the company behind Square and Cash App &#8212; would lay off 40% of its workforce, more than 4,000 people. The stated reason was not cost pressure, not poor performance, not strategic restructuring in the conventional sense. <strong>Dorsey wrote plainly that &#8220;intelligence tools have changed what it means to build and run a company&#8221; &#8212; and said he chose immediate, honest action over a gradual drawdown over months or years.</strong> </p><p>Block&#8217;s stock rose more than 20% in after-hours trading. The market&#8217;s signal was unambiguous: investors are rewarding companies for accelerating the substitution of human labor with AI capability, even at scale, even when the operational execution is uncertain. The question every leadership team should be asking is not whether this happens elsewhere &#8212; it&#8217;s at what rate, and with what obligations to the people affected. (<a href="https://archive.is/20260227010747/https://www.wsj.com/business/jack-dorseys-block-to-lay-off-4-000-employees-in-ai-remake-28f0d869">Jack Dorsey&#8217;s Block to Lay Off 40% of Its Workforce in AI Remake</a>)</p><p>The Anthropic Cowork event added a piece of data that deserves its own moment. Anthropic&#8217;s head of economics, drawing on privacy-preserving analysis of how Claude is actually being used, reported that a year ago roughly a third of all US jobs had at least a quarter of their associated tasks appearing in Claude usage data. <strong>That figure has now risen to approximately one in every two jobs &#8212; and when businesses embed Claude through the API, the overwhelming pattern is automation, not augmentation.</strong> He specifically called out &#8220;jobs that are pure implementation&#8221; &#8212; data entry workers, technical writers &#8212; as occupations where Claude is already handling tasks central to those roles. No widespread labor displacement has materialized yet in the data, but the exposure is broadening faster than most organizations are tracking. The economist&#8217;s advice to leaders cut to the real issue: the bottleneck isn&#8217;t model capability, it&#8217;s whether your organization&#8217;s knowledge is structured and accessible enough for AI to act on it. (<a href="https://venturebeat.com/orchestration/anthropic-says-claude-code-transformed-programming-now-claude-cowork-is">Anthropic says Claude Code transformed programming. Now Claude Cowork is coming for the rest of the enterprise.</a>)</p><div><hr></div><h2>On the Bigger Picture</h2><p><em>A few pieces this week were less about immediate enterprise decisions and more about the shape of what&#8217;s coming.</em></p><p>Canva announced two acquisitions &#8212; Cavalry for 2D animation and MangoAI for video ad optimization &#8212; while the broader software market continued to be punished on AI displacement fears. Canva&#8217;s position is instructive: $4 billion in annualized revenue, up 36% year over year, while Adobe is down 30% for the year. The company&#8217;s co-founder put it plainly: <strong>&#8220;AI is great at getting you to 80%. That last 20% &#8212; where you&#8217;re confident you can push this out and truly represent your brand &#8212; that&#8217;s really tricky to do.&#8221;</strong> Companies that embed AI deeply while keeping human judgment in the loop for the final mile are, for now, the ones finding competitive ground. (<a href="https://www.cnbc.com/2026/02/23/canva-acquires-cavalry-for-motion-graphics-and-mangoai-for-video-ads.html">As Wall Street punishes software stocks over AI concerns, Canva gets more acquisitive</a>)</p><div><hr></div><p>The piece I kept coming back to this week &#8212; the one that I suspect you may have already encountered &#8212; was CitriniResearch&#8217;s &#8220;2028 Global Intelligence Crisis.&#8221; I want to be clear about what this is: a scenario document, explicitly labeled speculative fiction, not a forecast. But here&#8217;s the thing: <strong>it moved real markets at real speed.</strong> Published on a Sunday, it triggered a broad selloff the following Monday: </p><ul><li><p><strong>The Dow dropped more than 800 points</strong></p></li><li><p><strong>IBM fell nearly 13%</strong></p></li><li><p><strong>Payment companies like American Express, Mastercard, and Visa all tumbled</strong></p></li><li><p><strong>DoorDash and private equity giants KKR and Blackstone down over 8%</strong></p></li></ul><p>The piece accumulated 22 million views on X after Michael Burry amplified it. The Wall Street Journal cited it as a key accelerant of investor anxiety. Citadel Securities published a formal rebuttal the same week. <strong>A speculative essay on Substack moved hundreds of billions in market value in a single session.</strong> That fact alone is worth sitting with, regardless of whether the scenario is accurate.</p><p>The scenario itself is worth understanding because it traces the internal logic of AI disruption further than most analysis does. The core mechanism: companies cut headcount, reinvest savings into AI, AI improves, enabling further cuts &#8212; with no natural corrective cycle, because unlike a typical recession, this downturn&#8217;s cause is structural, not cyclical. <strong>Citrini introduces the concept of &#8220;Ghost GDP&#8221; &#8212; output that registers in national accounts but never circulates through the consumer economy, because machines don&#8217;t spend money on discretionary goods.</strong> The scenario traces the feedback loop through white-collar wage compression, consumer spending decline, SaaS revenue erosion, private credit defaults on PE-backed software LBOs, and eventually prime mortgage distress &#8212; borrowers with 780 FICO scores whose income assumptions were written before the world changed.</p><p>Citadel&#8217;s rebuttal pushed back hard, pointing to rising software engineering job postings (+11% YoY), stable AI adoption rates in actual labor data, and the historical pattern of general-purpose technologies creating more jobs than they displace. Those counterarguments have real merit. But what the debate itself reveals may matter more than who wins it: <strong>we are in a moment where the gap between &#8220;plausible narrative&#8221; and &#8220;tradeable signal&#8221; has collapsed to nearly nothing, and institutional investors are making allocation decisions based on AI disruption fears that are still more scenario than data.</strong> For organizations watching the market signals around enterprise software valuations, payments infrastructure, and professional services firms, the message is the same whether or not the 2028 scenario materializes: the repricing of human knowledge work is already underway, and capital markets are running well ahead of the economic data. (<a href="https://www.citriniresearch.com/p/2028gic">THE 2028 GLOBAL INTELLIGENCE CRISIS</a>)</p><div><hr></div><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.beaiready.ai/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption"><strong>Like this article?</strong> Subscribe to get notified when the next article drops!</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><div><hr></div><p><em>What this week&#8217;s reading adds up to is something like a threshold moment: the ideas that have been circulating in boardrooms and conference sessions for two years are now being operationalized and disclosed. Platforms are being launched, protocols are becoming policy questions, layoffs are being attributed explicitly to AI capability, and macro researchers are stress-testing scenarios that once seemed speculative. Organizations that treat this as background noise &#8212; as something to monitor rather than act on &#8212; are making a choice with real consequences. The thinking divide Anthropic named this week is real, and it&#8217;s widening. The question isn&#8217;t whether to engage with AI transformation; it&#8217;s whether your organization can do it with intention and clarity before the decisions get made for you.</em></p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.beaiready.ai/p/the-beaiready-brief-week-9/comments&quot;,&quot;text&quot;:&quot;Leave a comment&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.beaiready.ai/p/the-beaiready-brief-week-9/comments"><span>Leave a comment</span></a></p><p></p>]]></content:encoded></item><item><title><![CDATA[the BeAIReady Brief | Week 8]]></title><description><![CDATA[What I'm actually reading about AI and the Digital Workplace (not an AI curated list of articles).]]></description><link>https://www.beaiready.ai/p/the-beaiready-brief-week-8</link><guid isPermaLink="false">https://www.beaiready.ai/p/the-beaiready-brief-week-8</guid><dc:creator><![CDATA[Erick Straghalis]]></dc:creator><pubDate>Fri, 20 Feb 2026 22:32:33 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!xd-G!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9dc57bd2-dd7e-4e9a-b34f-fd00c3b97237_747x747.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em>Hi there! This is a new segment for BeAIReady. Each week (or just about) I&#8217;ll recap what I&#8217;ve been <strong>actually</strong> reading and watching related to AI and the digital workplace. This isn&#8217;t an AI curated list of articles. These are actual articles I&#8217;ve read, along with some analysis on the impact on the practical impact on the digital workplace.</em></p><p>This week&#8217;s reading was defined by a single uncomfortable question running through almost everything I saw: who actually benefits when AI restructures work? The market is already voting &#8212; punishing the companies AI threatens and rewarding those aligned with it. Agents are moving from novelty to infrastructure, and the governance debates are arriving right on cue. And on the ground, knowledge workers are quietly discovering that the biggest shift isn&#8217;t capability &#8212; it&#8217;s continuity. </p><p><em>Here&#8217;s what I&#8217;ve been reading this week.</em></p><div><hr></div><h2>The Market Is Repricing AI&#8217;s Impact &#8212; and the Signal is What&#8217;s Falling, <em>Not What&#8217;s Rising</em></h2><p>The story everyone was watching this week was the so-called &#8220;<em>SaaSpocalypse</em>&#8221; &#8212; a wave of selling that wiped over a trillion dollars from enterprise software valuations. Salesforce down 26%. ServiceNow down 28%. HubSpot down 39%. The hot take was that AI is overhyped and the bubble is cracking. A much better read, from Dave Brear on LinkedIn, makes the opposite case: <strong>markets aren&#8217;t losing faith in AI &#8212; they&#8217;re losing faith in the companies AI is threatening to replace.</strong> (<a href="https://www.linkedin.com/pulse/ai-sceptics-right-wrong-thing-dave-brear-i47xe/">The AI Skeptics Are Right About the Wrong Thing</a>)</p><p>That reframe matters. Deutsche Bank ran a remarkable experiment this week &#8212; they asked their own AI tool, dbLumina, which industries it planned to disrupt. The answer was uncomfortably direct: <strong>information technology and software are the most exposed</strong>, followed by wealth management (80% of retail investor interactions handled by AI by 2027) and customer service (75% automated by 2026). The machine identified its own targets. (<a href="https://finance.yahoo.com/news/deutsche-bank-asked-ai-planning-205002276.html">Deutsche Bank asked AI how it was planning to destroy jobs</a>)</p><p>Figma is a useful case study in what it looks like to be on the right side of that line. The stock is down 80% since IPO &#8212; and yet investors cheered its Q4 earnings, which showed 40% year-over-year revenue growth and the highest net dollar retention in ten quarters. In a CNBC interview, CEO Dylan Field ranked Claude at the top of the LLM stack, said 75% of their larger customers are already consuming AI credits weekly, and pushed back hard on the &#8220;software is dead&#8221; narrative. <strong>Figma&#8217;s partnerships with Anthropic and OpenAI aren&#8217;t decoration &#8212; they&#8217;re the thesis.</strong> (<a href="https://archive.is/20260218214855/https://fortune.com/2026/02/18/fig-stock-q4-earnings/">Figma Q4 earnings</a> | <a href="https://www.youtube.com/watch?v=MhP9nOunyFo">Figma CEO interview</a>) Citi&#8217;s equity analyst Tyler Radke added useful color: he thinks we&#8217;re nearing a bottom for software stocks, but his advice is to be selective &#8212; <strong>favor companies with real AI stories and avoid legacy SaaS businesses that aren&#8217;t showing growth or margin improvement.</strong> (<a href="https://www.youtube.com/watch?v=xU2r_iSQD6M">Citi sell-off analysis</a>)</p><div><hr></div><h2>Agents Are Becoming Infrastructure &#8212; and the Governance Questions Just Arrived</h2><p>Underneath the market noise, something more structurally significant is happening with AI agents. A team at UC Santa Barbara published a new framework called Group-Evolving Agents (GEA), which allows AI agents to evolve collectively &#8212; sharing innovations across the group rather than siloing improvements in individual branches. <strong>GEA matched the performance of human-engineered agent systems on real-world coding benchmarks, at zero additional deployment cost.</strong> It&#8217;s an early signal that the &#8220;design the agent&#8221; phase of enterprise AI may give way to &#8220;let agents design themselves.&#8221; (<a href="https://venturebeat.com/orchestration/new-agent-framework-matches-human-engineered-ai-systems-and-adds-zero">GEA framework</a>)</p><p>At the same time, the agent ecosystem got its first real governance friction. Anthropic quietly updated its documentation to clarify that routing Claude requests through personal Pro or Max subscriptions &#8212; which is exactly how popular personal agents like OpenClaw and NanoClaw work &#8212; violates their terms. The community reacted, Anthropic walked it back as a documentation clarification rather than a policy change, but <strong>the underlying tension is real: where is the line between personal experimentation and building a business on subsidized platform access?</strong> (<a href="https://thenewstack.io/anthropic-agent-sdk-confusion/">Anthropic/OpenClaw ToS</a>) The denouement was almost too on-the-nose: the creator of OpenClaw, <strong>Peter Steinberger, promptly announced he was joining OpenAI</strong> to work on &#8220;the next generation of personal agents.&#8221; (<a href="https://www.pcmag.com/news/creator-of-viral-ai-tool-openclaw-joins-openai">OpenClaw creator joins OpenAI</a>)</p><p>The governance theme ran deeper than platform ToS disputes. Anthropic CEO Dario Amodei resurfaced in a Fortune piece with a quote that deserves more attention than it got: &#8220;I&#8217;m deeply uncomfortable with these decisions being made by a few companies, by a few people.&#8221; He wasn&#8217;t talking about competitors &#8212; he was talking about himself. <strong>No one elected him. No federal AI regulations exist.</strong> Thirty-eight states have adopted some form of AI legislation, but there&#8217;s no coherent national framework. <strong>Amodei is advocating for more regulation while acknowledging the commercial pressures working against it from inside his own company.</strong> It&#8217;s an honest and uncomfortable position &#8212; which is probably why it&#8217;s worth taking seriously. (<a href="https://fortune.com/article/why-is-anthropic-ceo-dario-amodei-deeply-uncomfortable-companies-in-charge-ai-regulating-themselves/">Anthropic CEO on governance</a>)</p><div><hr></div><h2>What AI Actually Feels Like From Inside the Work</h2><p>Three more grounded reads this week, all relevant to anyone doing knowledge work.</p><p>The Claude Code diary by Dave Brear is worth reading slowly. He spent a week connecting Claude Code to his personal knowledge vault and discovered something easy to miss in the broader AI hype: <strong>the value isn&#8217;t in any individual response &#8212; it&#8217;s in continuity.</strong> <strong>When AI knows your files, your conventions, your half-formed ideas, the collaboration changes character entirely</strong>. You stop prompting and start conversing. He hit rate limits, felt the absence like a colleague stepping out for lunch, and ended the week staring at the Claude Max pricing page. The piece captures something true about what persistent AI context actually feels like. (<a href="https://www.linkedin.com/pulse/diary-cheapskate-knowledge-worker-one-week-claude-code-dave-brear-bqt2e/">Claude Code diary</a>)</p><p><strong>Microsoft open-sourced VibeVoice this week</strong> &#8212; a local audio model that handles text-to-speech, speech-to-text, and voice cloning, all without a cloud subscription or API key. <strong>Capabilities that were expensive and gated six months ago are now free and local</strong> &#8212; and that&#8217;s increasingly the story with open-source AI. (<a href="https://www.youtube.com/watch?v=AyHSSslWeHE">Microsoft VibeVoice</a>)</p><p>Finally, a study worth bookmarking for anyone creating content professionally: an analysis of 1.2 million AI-generated answers found that <strong>44% of ChatGPT citations come from the first 30% of any given piece of content.</strong> AI models weight early framing more heavily and interpret the rest through that lens. The practical implication is simple: <strong>if you want AI to surface your ideas, stop burying your best thinking at the end. Front-load your substance</strong>. (<a href="https://searchengineland.com/chatgpt-citations-content-study-469483">ChatGPT citation study</a>)</p><div><hr></div><h2>On the Bigger Picture</h2><p>Two reads that don&#8217;t fit the workplace frame neatly but are worth noting. IEEE Spectrum published a careful piece arguing that <strong>the US and China aren&#8217;t racing toward the same finish line</strong> &#8212; the US is doubling down on AGI, while China is focused on embedding AI into manufacturing, healthcare, and logistics as a near-term productivity engine. The &#8220;arms race&#8221; framing may be creating a self-fulfilling prophecy that increases risk for everyone. (<a href="https://spectrum.ieee.org/us-china-ai">US vs. China AI futures</a>) And separately, China&#8217;s humanoid robots went from stumbling through folk dances at last year&#8217;s Spring Festival Gala to performing kung fu flips and choreographed gymnastics this year &#8212; <strong>China now accounts for more than 85% of global humanoid robot installations.</strong> The AI model race, one analyst noted, will ultimately matter more than the hardware. But the hardware is catching up faster than most expected. (<a href="https://www.cnbc.com/2026/02/20/china-humanoid-robots-spring-festival-gala-unitree-tesla-ai-race.html">China humanoid robots</a>)</p><div><hr></div><p><em>The thread connecting most of this week&#8217;s reading: AI isn&#8217;t disrupting work in the future tense anymore. The disruption is active &#8212; in valuations, in agent governance debates, in how knowledge workers structure their days and their content. The question for organizations isn&#8217;t whether to engage with this. It&#8217;s whether they&#8217;re paying close enough attention to know which side of the line they&#8217;re on.</em></p>]]></content:encoded></item></channel></rss>