<?xml version="1.0" encoding="UTF-8"?><rss xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:atom="http://www.w3.org/2005/Atom" version="2.0" xmlns:itunes="http://www.itunes.com/dtds/podcast-1.0.dtd" xmlns:googleplay="http://www.google.com/schemas/play-podcasts/1.0"><channel><title><![CDATA[The So What]]></title><description><![CDATA[We focus on practical implications, real client challenges, and the foundational truths about how AI is reshaping business today. ]]></description><link>https://tsw.blankmetal.ai</link><image><url>https://substackcdn.com/image/fetch/$s_!Cu0M!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F85d8da71-727a-40a7-b3ec-0443573853bb_800x800.png</url><title>The So What</title><link>https://tsw.blankmetal.ai</link></image><generator>Substack</generator><lastBuildDate>Wed, 06 May 2026 10:52:36 GMT</lastBuildDate><atom:link href="https://tsw.blankmetal.ai/feed" rel="self" type="application/rss+xml"/><copyright><![CDATA[Blank Metal]]></copyright><language><![CDATA[en]]></language><webMaster><![CDATA[blankmetal@substack.com]]></webMaster><itunes:owner><itunes:email><![CDATA[blankmetal@substack.com]]></itunes:email><itunes:name><![CDATA[Blank Metal]]></itunes:name></itunes:owner><itunes:author><![CDATA[Blank Metal]]></itunes:author><googleplay:owner><![CDATA[blankmetal@substack.com]]></googleplay:owner><googleplay:email><![CDATA[blankmetal@substack.com]]></googleplay:email><googleplay:author><![CDATA[Blank Metal]]></googleplay:author><itunes:block><![CDATA[Yes]]></itunes:block><item><title><![CDATA[Weekly Headlines: Issue #20]]></title><description><![CDATA[April 23 - April 30, 2026]]></description><link>https://tsw.blankmetal.ai/p/weekly-headlines-issue-20</link><guid isPermaLink="false">https://tsw.blankmetal.ai/p/weekly-headlines-issue-20</guid><dc:creator><![CDATA[Blank Metal]]></dc:creator><pubDate>Fri, 01 May 2026 17:58:43 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!YivG!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc0f5c468-411a-4289-88a2-2a4d4599eb5f_1200x670.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!YivG!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc0f5c468-411a-4289-88a2-2a4d4599eb5f_1200x670.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!YivG!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc0f5c468-411a-4289-88a2-2a4d4599eb5f_1200x670.png 424w, https://substackcdn.com/image/fetch/$s_!YivG!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc0f5c468-411a-4289-88a2-2a4d4599eb5f_1200x670.png 848w, https://substackcdn.com/image/fetch/$s_!YivG!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc0f5c468-411a-4289-88a2-2a4d4599eb5f_1200x670.png 1272w, https://substackcdn.com/image/fetch/$s_!YivG!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc0f5c468-411a-4289-88a2-2a4d4599eb5f_1200x670.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!YivG!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc0f5c468-411a-4289-88a2-2a4d4599eb5f_1200x670.png" width="1200" height="670" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/c0f5c468-411a-4289-88a2-2a4d4599eb5f_1200x670.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:670,&quot;width&quot;:1200,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:1480789,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://tsw.blankmetal.ai/i/196141078?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc0f5c468-411a-4289-88a2-2a4d4599eb5f_1200x670.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!YivG!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc0f5c468-411a-4289-88a2-2a4d4599eb5f_1200x670.png 424w, https://substackcdn.com/image/fetch/$s_!YivG!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc0f5c468-411a-4289-88a2-2a4d4599eb5f_1200x670.png 848w, https://substackcdn.com/image/fetch/$s_!YivG!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc0f5c468-411a-4289-88a2-2a4d4599eb5f_1200x670.png 1272w, https://substackcdn.com/image/fetch/$s_!YivG!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc0f5c468-411a-4289-88a2-2a4d4599eb5f_1200x670.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Welcome to Blank Metal&#8217;s Weekly AI Headlines.</p><p>Each week, our team shares the AI stories that caught our attention&#8212;the articles, announcements, and insights we&#8217;re actually discussing internally. We curate the best of what we&#8217;re reading and add the context that matters: what happened, why it matters, and what to do about it.</p><h1>The AI Subsidy Era Ends</h1><p><em>The cheap-token era is closing. For 18 months, every enterprise AI roadmap was built on subsidized inference assumptions&#8212;prices falling quarter over quarter, vendors absorbing compute costs, flat-rate enterprise contracts capping the downside. This week, every one of those assumptions broke at once. Three frontier-pricing changes, one budget blowout, and one canonical &#8220;AI bundled into a flat license&#8221; product moving to metered billing all landed inside seven days. Time to recalc.</em></p><h2>OpenAI Doubles GPT-5.5&#8217;s API Price&#8212;Efficiency Gains Don&#8217;t Cover It</h2><p><strong>What:</strong> OpenAI launched GPT-5.5 on April 23 and doubled the API price along with it. Input tokens move from $2.50 to $5.00 per million; output tokens move from $15.00 to $30.00 per million. OpenAI&#8217;s stated rationale is that GPT-5.5 is more efficient and needs fewer tokens for comparable tasks. Independent testing from Artificial Analysis found effective API costs roughly 20% higher than the prior GPT-5.4 line&#8212;efficiency gains offset, but didn&#8217;t erase, the headline price hike.</p><p><strong>So What:</strong> This is the first frontier-model release in 18 months that didn&#8217;t pretend to be cost-neutral. The script for every prior launch was the same&#8212;new model, same price, occasional discount. GPT-5.5 doubled the sticker. The framing matters: OpenAI is signaling that capability gains now ship at premium pricing, and efficiency improvements go to vendor margin first. Anyone building production features on the GPT line just had their unit economics recalibrated without warning.</p><p><strong>Now What:</strong> If you&#8217;re running production workloads on GPT-5.x, redo the math on cost-per-task before the next quarterly review. The 20% effective-cost increase on identical work is the floor&#8212;token-heavy patterns (agents, long-context reasoning, multi-turn) feel it more. Run a model bake-off on real internal examples, not benchmark suites. The cheaper tiers (GPT-5.5 mini, open-weights, Claude Haiku) handle more than most teams assume.</p><p><a href="https://the-decoder.com/openai-unveils-gpt-5-5-claims-a-new-class-of-intelligence-at-double-the-api-price/">Read more</a></p><h2>Anthropic Moves Enterprise Customers Off Flat-Rate Pricing</h2><p><strong>What:</strong> The Information reported that Anthropic is moving select enterprise customers off flat-rate contracts onto usage-based billing, citing demand outpacing compute supply. Customers who locked in fixed-fee enterprise terms over the last year are being asked to renegotiate against a pricing model pegged to actual token consumption.</p><p><strong>So What:</strong> This is the same story as the GPT-5.5 price hike from a different angle. Two of three frontier vendors are simultaneously signaling that the flat-rate, capped-cost enterprise contract is no longer the default&#8212;and the trigger is compute scarcity, not competition. Buyers who anchored AI budgets on predictable monthly billing are about to discover what their actual usage costs at retail.</p><p><strong>Now What:</strong> If your company has a flat-rate Anthropic contract up for renewal in 2026, build the usage-based scenario now. Pull six months of token logs by use case, model the cost at retail rates, then negotiate from a number rather than a feeling. If you&#8217;re still in a flat-rate tier, audit which consumption patterns the vendor would charge you for under metered billing&#8212;the workloads that look ugliest under that model are your highest-leverage targets for compression or migration.</p><p><a href="https://www.theinformation.com/articles/anthropic-changes-pricing-bill-firms-based-ai-use-amid-compute-crunch">Read more</a></p><h2>Tokenmaxxing Isn&#8217;t a Productivity Metric</h2><p><strong>What:</strong> The Register published a deep look at token economics on April 26. ML researcher Devansh calculated theoretical inference cost on an H100 at $0.0038 per million tokens at full utilization, rising to $0.013 at 30% utilization and $0.038 at 10%. Anthropic&#8217;s Opus 4.7 lists at $5/M input and $25/M output&#8212;orders of magnitude above bare-metal cost. Devansh on token-volume KPIs at Meta and Shopify: &#8220;Is token spend directly correlated with productivity? Absolutely not.&#8221; Future Tech Enterprise CEO Bob Venero added that hardware costs are 3x what they were six months ago, and only 15% of AI prototypes reach production without guidance&#8212;45-50% with proper planning.</p><p><strong>So What:</strong> The premium between bare inference cost and frontier-model retail isn&#8217;t going to compress on its own. Vendors charge what the market bears, and the market still bears a lot because most enterprise buyers don&#8217;t have a clean cost-per-task baseline to negotiate against. Worse, &#8220;tokens consumed&#8221; has crept into corporate scorecards as a proxy for AI productivity&#8212;a metric that rewards waste. If your team is measured on tokens used, you&#8217;re going to get tokens used.</p><p><strong>Now What:</strong> Stop measuring AI adoption by token volume. Pick three AI-powered workflows in your company, compute cost-per-completed-task, and put that number on a leadership dashboard instead. Then run the same workflows against a smaller model, an open-weights alternative, or a deterministic non-LLM approach where one exists. The 3x hardware cost gap means the self-hosting math has shifted in the last six months too&#8212;revisit it.</p><p><a href="https://www.theregister.com/2026/04/26/ai_price_tag/">Read more</a></p><h2>Uber Blew Through Its Full 2026 AI Budget on Tokens by April</h2><p><strong>What:</strong> Axios reported on April 26 that Uber&#8217;s CTO consumed Uber&#8217;s full 2026 AI budget on token costs alone before the year was halfway done. The piece, sourced back to The Information, frames a broader pattern: IT budgets are blowing out as token spend on agents, code-gen, and copilots overruns multi-quarter projections.</p><p><strong>So What:</strong> Uber is not a sloppy buyer. If their CTO modeled a year of spend and got blown out by token usage at the halfway mark, the modeling assumptions everyone built on&#8212;token prices keep falling, vendor pricing stays flat, agentic workloads consume linearly&#8212;were all wrong. The asymmetry between flat-rate vendor signaling and actual consumption growth is now showing up in board-level finance reviews, not just engineering retros.</p><p><strong>Now What:</strong> If your 2026 AI budget was set in Q4 2025, assume it&#8217;s wrong by 50-200% on token-dependent line items. Get monthly token consumption visibility by team and use case before mid-year. The teams most exposed are the ones who shipped agentic workflows in Q1&#8212;those are 10-20 LLM calls per task instead of one, and the cost compounds. A simple guardrail: cap token spend per workflow at the level where it stops being cheaper than human time, then look hard at any workflow stuck against the cap.</p><p><a href="https://www.axios.com/2026/04/26/ai-cost-human-workers">Read more</a></p><h2>GitHub Copilot Shifts to Metered Billing&#8212;Annual Subscribers Pay 27x for Opus</h2><p><strong>What:</strong> GitHub announced on April 28 that Copilot will move from request-based to token-based billing effective June 1, 2026. New tiers: Pro at $10/month for 1,000 AI Credits, Pro+ at $39 for 3,900, Business at $19/user for 1,900, Enterprise at $39/user for 3,900. Annual subscribers face dramatically higher model multipliers under the new system&#8212;Claude Opus 4.7&#8217;s multiplier rises from 7.5x to 27x. GitHub CPO Mario Rodriguez: &#8220;Today, a quick chat question and a multi-hour autonomous coding session can cost the user the same amount. GitHub has absorbed much of the escalating inference cost behind that usage, but the current premium request model is no longer sustainable.&#8221;</p><p><strong>So What:</strong> Copilot was the canonical example of &#8220;AI bundled into a flat seat license.&#8221; That bundle was profitable when sessions were short and models were cheap. Both assumptions broke. Coding agents that run for hours, not seconds, are the new default usage pattern&#8212;and GitHub just told its 25M+ users that the bill for that pattern lives with them now, not Microsoft. Expect the same shift across every AI feature currently buried in a flat-rate developer tool license.</p><p><strong>Now What:</strong> If your engineering org standardized on Copilot under a flat-license assumption, your per-developer cost is about to become variable and individually unbounded. Start tracking session length and model selection by user, decide which tiers map to which engineer cohorts, and write a usage policy before someone runs an Opus session over a long weekend. The teams who&#8217;ll feel this most are the ones who treated agent mode as the default&#8212;Pro+ at 3,900 credits doesn&#8217;t go far against a 27x multiplier.</p><p><a href="https://www.theregister.com/2026/04/28/microsofts_github_shifts_to_metered/">Read more</a></p><h1>The Capital Behind the Curtain</h1><p><em>Behind every pricing change in the prior section is a capital structure that requires it. Hyperscalers and frontier labs are now financially entangled at a scale that determines what models you can buy, at what price, and from whom. Two headline numbers this week made the entanglement legible.</em></p><h2>Big Tech AI Capex Hits $600B for 2026&#8212;And Cash Flow Can&#8217;t Keep Up</h2><p><strong>What:</strong> Reporting this week pegs combined 2026 AI capex from Alphabet, Microsoft, Meta, and Amazon at roughly $600 billion. Joe Maginot of Madison Investments: &#8220;These have been businesses that generated significant amounts of free cash flow and today, pretty much all operating cash flow is being consumed in capex.&#8221; Melissa Otto of S&amp;P Global Visible Alpha on Microsoft: &#8220;The company is going to have to speak about why their business model isn&#8217;t going to get meaningfully disrupted in AI.&#8221;</p><p><strong>So What:</strong> This is the supply side of the same story driving every pricing change in this issue. The hyperscalers have committed to spending the equivalent of two Manhattan Projects on AI infrastructure this year, and they need that spend to convert into recurring revenue at meaningfully higher margins than current AI services produce. The math doesn&#8217;t work at flat-rate pricing&#8212;it doesn&#8217;t even work at current usage-based pricing if token consumption stops compounding. Expect the next 18 months to be defined by vendors figuring out how to capture more revenue per token consumed, not less.</p><p><strong>Now What:</strong> Treat any AI vendor pricing announcement in 2026 as a leading indicator, not a stable input. Negotiate price-protection language into multi-year contracts&#8212;floor caps on annual increases, locked rate cards for committed volumes, ramp-down protection if internal usage projections miss. If your company is publicly traded, your CFO is going to get the same Visible Alpha question Microsoft got: how does the model survive if frontier-API pricing doubles again? Have an answer.</p><p><a href="https://www.bnnbloomberg.ca/business/economics/2026/04/28/big-tech-investors-to-gauge-payoff-as-ai-spending-set-to-hit-600-billion/">Read more</a></p><h2>Google Commits Up to $40B to Anthropic&#8212;Compute Is the New Currency</h2><p><strong>What:</strong> Google announced on April 24 that it will invest up to $40 billion in Anthropic&#8212;$10 billion now in cash at a $350 billion valuation, with another $30 billion contingent on performance milestones. Google Cloud also committed five gigawatts of computing power across a five-year window, with optionality for several more gigawatts. Prior to this round, Google&#8217;s stake in Anthropic was reportedly 14% from $3 billion in earlier rounds. The structure mirrors Anthropic&#8217;s earlier deal with Amazon&#8212;$5 billion now, up to $20 billion against milestones.</p><p><strong>So What:</strong> A direct competitor (Google has Gemini) is making the largest single AI investment ever recorded&#8212;into a company building competing models&#8212;because compute access has become more strategic than market share. The entire frontier-model field now runs on capital from the same three hyperscalers it competes against. For enterprise buyers, this consolidation is invisible during good quarters and very visible the moment a model vendor&#8217;s compute partner has competing priorities.</p><p><strong>Now What:</strong> When you negotiate a multi-year AI contract, ask which hyperscaler hosts the model you&#8217;re committing to. Then ask what happens if that hyperscaler&#8217;s AI roadmap diverges from your vendor&#8217;s. The answer determines whether you have one supplier or three. For workloads where this matters&#8212;regulated, mission-critical, or strategically differentiating&#8212;architect for portability across providers from day one. Single-vendor lock-in is more expensive in this market than it has been since the 1990s mainframe contracts.</p><p><a href="https://www.cnbc.com/2026/04/24/google-to-invest-up-to-40-billion-in-anthropic-as-search-giant-spreads-its-ai-bets.html">Read more</a></p><h1>Enterprise Stacks Restructure for Agents</h1><p><em>While the cost economics shifted, the infrastructure layer kept moving. The most defended interface in finance committed to a chat front end, Microsoft bundled its agent governance plane into a new flagship SKU, and Linear made itself a node in the agent network instead of a destination application. The pattern across all three: every enterprise stack is being rebuilt around the assumption that an agent&#8212;not a person&#8212;will be the primary user.</em></p><h2>Bloomberg Terminal Bets Its Future on a Chat Interface</h2><p><strong>What:</strong> WIRED reported on April 28 that Bloomberg is testing a chatbot-style interface for the Terminal called ASKB, built atop a basket of language models. The beta is open to roughly a third of the Terminal&#8217;s 375,000 users. Bloomberg CTO Shawn Edwards: &#8220;This will be the new terminal. The primary way most interactions happen.&#8221; The Terminal now ingests weather forecasts, shipping logs, factory locations, consumer spending patterns, and private loan data alongside traditional market data&#8212;and Edwards&#8217;s framing is that the data volume has made command-line keystroke navigation untenable. ASKB supports workflow templates with scheduled or conditional triggers; an earnings-season template can pull competitor comparisons, fundamentals, and Wall Street expectations and generate a long/short summary automatically.</p><p><strong>So What:</strong> The Bloomberg Terminal is the most defended interface in finance. Every senior trader, analyst, and asset manager has 25 years of muscle memory for the keystroke shortcuts&#8212;it&#8217;s the &#8220;Excel of finance&#8221; with even higher switching costs. Bloomberg&#8217;s CTO publicly committing to chat as the primary interaction mode is a forcing event for every other enterprise software vendor whose product is fundamentally a structured query system over a proprietary data set. If Bloomberg can rebuild itself around an LLM front end, no entrenched workflow tool is safe behind a &#8220;but our users won&#8217;t change&#8221; defense.</p><p><strong>Now What:</strong> If your company runs on a structured-data interface&#8212;internal BI tool, ticketing system, CRM, ERP module, custom dashboard&#8212;the question is no longer whether a chat layer will replace the keystroke layer. The question is whether you build it or your software vendor does. Build it where the data and workflow are differentiating to your business. Let the vendor build it where the underlying data is commodity. The middle option&#8212;wait and see&#8212;is getting more expensive every quarter.</p><p><a href="https://www.wired.com/story/the-bloomberg-terminal-is-getting-an-ai-makeover/">Read more</a></p><h2>Microsoft Bundles Copilot and Agent 365 Into a New &#8220;Frontier Suite&#8221;</h2><p><strong>What:</strong> Microsoft announced that Microsoft 365 E5, Entra Suite, Copilot, and Agent 365 are being bundled and transact-able as Microsoft 365 E7&#8212;the Frontier Suite&#8212;available in Cloud Solution Provider channels starting May 1, 2026. The bundle pairs E5&#8217;s secure productivity stack with Entra for identity and access, Copilot for AI in workflow, and Agent 365 as the control plane for governing and scaling agents.</p><p><strong>So What:</strong> This is Microsoft&#8217;s bet that enterprise AI is now a stack-level purchase, not a per-feature add-on. Agent 365 as the &#8220;control plane&#8221; framing matters&#8212;Microsoft is trying to own the governance layer for any agent running inside your tenant, regardless of who built it. If E7 becomes the standard SKU for AI-enabled enterprises, Microsoft captures both the productivity revenue and the agent-governance revenue, and every other agent vendor becomes a participant in Microsoft&#8217;s governance plane rather than a peer to it.</p><p><strong>Now What:</strong> If your company is on E5 already, your Microsoft account team is going to pitch E7 within 30 days. Before that meeting, decide whether you want Microsoft as your agent governance plane or whether you&#8217;d rather build or buy that layer separately. The answer changes the math on E7&#8217;s premium and the architecture of every agent project on your roadmap. Either path is defensible; drifting into E7 by inertia and then trying to govern non-Microsoft agents around it is the worst of both options.</p><p><a href="https://learn.microsoft.com/en-us/partner-center/announcements/2026-april">Read more</a></p><h2>Linear Goes Bidirectional on MCP&#8212;Becomes a Node in the Agent Network</h2><p><strong>What:</strong> Linear shipped Agent MCP support on April 23, letting Linear Agent connect to external tools via Model Context Protocol&#8212;pulling context from Granola meeting notes into project updates, using Glean to draft project specs, turning Notion interview notes into customer requests, validating product hypotheses against PostHog data. Admins can control access with allowlists and workspace-level MCP permissions. Linear also expanded its own MCP server with support for initiatives, project milestones, and updates&#8212;so tools like Cursor and Claude can read and write back to Linear.</p><p><strong>So What:</strong> Linear is small relative to the Bloombergs and Microsofts in this issue, but the architecture decision is more consequential than the size suggests. By exposing Linear bidirectionally over MCP&#8212;both as a server and as a client&#8212;Linear stopped being a destination application and started being a node in an agent network. Every tool exposed this way becomes more useful when AI is in the loop and less useful when it isn&#8217;t. The opposite move (close the API, build a walled-garden AI experience) is what several incumbents shipped this quarter, and it&#8217;s a defensive play. Linear&#8217;s move is offensive.</p><p><strong>Now What:</strong> Audit your internal tool stack for which tools have MCP support, which have an OpenAPI spec that could be wrapped, and which are AI-hostile. The AI-hostile tools will feel slower, dumber, and more expensive every quarter&#8212;because every other tool in the stack is getting an agent layer and they aren&#8217;t. For the agent-friendly tools, decide which become the system of record your agents read from and write to, and start building workflow templates that span them. Companies treating MCP as an integration spec rather than a feature are setting themselves up for the agent-centric stack everyone will have by 2027.</p><p><a href="https://linear.app/changelog/2026-04-23-linear-agent-mcp-support">Read more</a></p>]]></content:encoded></item><item><title><![CDATA[Weekly Headlines: Issue #19]]></title><description><![CDATA[April 16 - April 23, 2026]]></description><link>https://tsw.blankmetal.ai/p/weekly-headlines-issue-19</link><guid isPermaLink="false">https://tsw.blankmetal.ai/p/weekly-headlines-issue-19</guid><dc:creator><![CDATA[Blank Metal]]></dc:creator><pubDate>Fri, 24 Apr 2026 13:01:12 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!Ow8A!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff872f74b-857b-46f0-9387-42fff780c4da_1200x670.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!Ow8A!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff872f74b-857b-46f0-9387-42fff780c4da_1200x670.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!Ow8A!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff872f74b-857b-46f0-9387-42fff780c4da_1200x670.png 424w, https://substackcdn.com/image/fetch/$s_!Ow8A!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff872f74b-857b-46f0-9387-42fff780c4da_1200x670.png 848w, https://substackcdn.com/image/fetch/$s_!Ow8A!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff872f74b-857b-46f0-9387-42fff780c4da_1200x670.png 1272w, https://substackcdn.com/image/fetch/$s_!Ow8A!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff872f74b-857b-46f0-9387-42fff780c4da_1200x670.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!Ow8A!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff872f74b-857b-46f0-9387-42fff780c4da_1200x670.png" width="1200" height="670" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/f872f74b-857b-46f0-9387-42fff780c4da_1200x670.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:670,&quot;width&quot;:1200,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:1480828,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://tsw.blankmetal.ai/i/195283298?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff872f74b-857b-46f0-9387-42fff780c4da_1200x670.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!Ow8A!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff872f74b-857b-46f0-9387-42fff780c4da_1200x670.png 424w, https://substackcdn.com/image/fetch/$s_!Ow8A!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff872f74b-857b-46f0-9387-42fff780c4da_1200x670.png 848w, https://substackcdn.com/image/fetch/$s_!Ow8A!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff872f74b-857b-46f0-9387-42fff780c4da_1200x670.png 1272w, https://substackcdn.com/image/fetch/$s_!Ow8A!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff872f74b-857b-46f0-9387-42fff780c4da_1200x670.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Welcome to Blank Metal&#8217;s Weekly AI Headlines.</p><p>Each week, our team shares the AI stories that caught our attention&#8212;the articles, announcements, and insights we&#8217;re actually discussing internally. We curate the best of what we&#8217;re reading and add the context that matters: what happened, why it matters, and what to do about it.</p><h1>The Workspace Wars Escalate</h1><p><em>Fifteen days after Claude Cowork went GA, OpenAI, Adobe, Salesforce, and Google all shipped workspace-layer moves in a single week. The category isn&#8217;t &#8220;who has the best chat model&#8221; anymore&#8212;it&#8217;s &#8220;whose workspace runs your agents, your skills, and your governance.&#8221; If you&#8217;re planning an AI rollout for anyone other than engineers, this is the layer that matters, and every incumbent platform you already pay for is quietly repositioning to defend turf in it.</em></p><h2>OpenAI Ships Workspace Agents in ChatGPT&#8212;The Cowork Category Is Now a Two-Vendor Race</h2><p><strong>What:</strong> OpenAI launched Workspace Agents inside ChatGPT, a goal-driven, multi-step agent surface that reads across connected tools, plans work, and delivers finished artifacts. It lands 15 days after Anthropic took Claude Cowork out of preview, and draws directly on Codex infrastructure for the execution layer.</p><p><strong>So What:</strong> Until last week, Anthropic owned the &#8220;workspace where AI does the work&#8221; category on its own. That&#8217;s over. Every enterprise AI conversation now has two credible Cowork-class products from the two labs most buyers are already paying, and the vendor choice collapses into a handful of real variables: connector catalog, skills format portability, admin controls, and which model your people are already using. The fact that OpenAI built on Codex rather than a clean-sheet agent runtime is also worth noting&#8212;it signals the coding-agent substrate and the workspace-agent substrate are the same product underneath.</p><p><strong>Now What:</strong> If you&#8217;ve already committed to Claude Cowork, don&#8217;t switch&#8212;but build your governance (RBAC, connector permissions, skills architecture) in a platform-agnostic way so you can run both where it makes sense. If you haven&#8217;t committed yet, this is the moment to pilot both side-by-side against two or three of your actual workflows and decide on evidence, not on vendor preference. The category-defining feature six months from now will be skills and agent portability, not necessarily the underlying model.</p><p><a href="https://openai.com/index/introducing-workspace-agents-in-chatgpt/">Read more</a></p><h2>Adobe Goes MCP-Native at Summit 2026&#8212;And Legacy Enterprise Platforms Just Got Interesting Again</h2><p><strong>What:</strong> Adobe announced CX Enterprise at Summit 2026: an end-to-end agentic customer-experience platform built around AI agents, reusable &#8220;agent skills,&#8221; and MCP endpoints, with a governance layer on top. Adobe Marketing Agent will appear inside Claude Enterprise, ChatGPT Enterprise, Gemini Enterprise, Copilot, and IBM watsonx Orchestrate. A new &#8220;CX Enterprise Coworker&#8221; takes a business goal (&#8221;increase cross-sell by 3%&#8221;), assembles agents, plans, and executes pending human approval.</p><p><strong>So What:</strong> Two things to notice. First, MCP is now a first-class citizen inside a legacy enterprise pitch, not a developer curiosity&#8212;Adobe is betting that portable agent standards are how incumbent platforms stay relevant as the agent layer commoditizes. Second, the retrofit-versus-reengineer debate inside every enterprise just got a template: Adobe kept AEP as the contextual layer and wrapped agents around it rather than rebuilding. That&#8217;s the pattern most of you will end up following.</p><p><strong>Now What:</strong> If you run a legacy platform of record&#8212;CRM, ERP, marketing, finance&#8212;stop waiting for the vendor to ship a &#8220;real&#8221; AI strategy. Start asking now whether they&#8217;ll expose MCP endpoints, whether their agents will run inside Claude Enterprise or ChatGPT Enterprise, and whether their skills are portable across your agent runtimes. A vendor that can&#8217;t answer those three questions by end of Q3 is a vendor you&#8217;re going to replace.</p><p><a href="https://news.adobe.com/news/2026/04/adobe-redefines-custome-experience">Read more</a></p><h2>Salesforce Launches Headless 360&#8212;Your Platform of Record Is Now Infrastructure for Agents</h2><p><strong>What:</strong> Salesforce unveiled Headless 360, which exposes the entire Salesforce platform as infrastructure for AI agents: data, business logic, workflows, and policy all available programmatically to any agent runtime, any model, any orchestration layer. It&#8217;s the first major CRM repositioning itself not as a destination app but as a system of record agents operate on top of.</p><p><strong>So What:</strong> This reframes the most expensive software purchase in most enterprises. If Salesforce is infrastructure, then the value question moves from &#8220;which CRM do we pick&#8221; to &#8220;what agents sit on top of it and who controls them&#8221;&#8212;and the answer to that second question is increasingly <em>you</em>, not Salesforce. The deeper signal is that the incumbents have now absorbed the agent thesis: they&#8217;re not fighting it, they&#8217;re repositioning around it. Expect the same move from ServiceNow, Workday, Oracle, and SAP over the next six months.</p><p><strong>Now What:</strong> If you&#8217;re a Salesforce customer, get ahead of this. Ask your account team where Headless 360 fits in your license, what the governance model looks like across multiple agent runtimes, and how skills and agents built against your instance survive a vendor change. If you&#8217;re evaluating CRM alternatives, the new decision criterion is: which platform will be easier to <em>operate on top of</em> a year from now.</p><p><a href="https://venturebeat.com/ai/salesforce-launches-headless-360-to-turn-its-entire-platform-into-infrastructure-for-ai-agents">Read more</a></p><h2>Gemini Gets a Next-Generation Deep Research Agent&#8212;Research-as-Workflow, Not Research-as-Search</h2><p><strong>What:</strong> Google launched a next-generation Deep Research agent inside Gemini. It runs multi-hour investigations across the open web, synthesizes findings into structured reports, and interleaves reasoning, citations, and cross-checks instead of returning a ranked list of links.</p><p><strong>So What:</strong> This is the first credible move from Google that positions Gemini as more than a search box with a model attached. Deep Research is a workflow product, not an answer product&#8212;the same architectural bet Claude and ChatGPT made with their respective research and agent modes. For enterprise buyers, it also forces a real choice: if your analysts start using Deep Research for diligence, market scans, or regulatory reviews, you need governance around it before it becomes the de facto research tool on your team.</p><p><strong>Now What:</strong> If you have analysts, researchers, or consultants spending hours per week on web-synthesis work, pilot Deep Research against one of them for a week and measure the delta. If the gains are real, your next question is governance: source control, citation audit, data residency, and whether the research output can be trusted in a regulated workflow. Don&#8217;t let this diffuse through your org ungoverned&#8212;treat it like you&#8217;d treat any new research tool with internet access.</p><p><a href="https://blog.google/innovation-and-ai/models-and-research/gemini-models/next-generation-gemini-deep-research/">Read more</a></p><h1>The Model Race: Coding and Life Sciences</h1><p><em>The frontier model race kept moving on two fronts this week. Google publicly conceded Anthropic is ahead on coding and stood up a strike team to catch up. Moonshot&#8217;s open-weights Kimi K2.6 put a credible open model inside the frontier envelope for the first time. And OpenAI shipped the first vertical frontier model&#8212;GPT-Rosalind for life sciences&#8212;with named pharma customers. Two signals for enterprise buyers: vendor leadership swaps faster than your procurement cycle, and vertical frontier models are the next GTM pattern.</em></p><h2>Google DeepMind Spins Up a Strike Team to Close the Coding Gap With Anthropic</h2><p><strong>What:</strong> The Decoder reports Google DeepMind has stood up a strike team led by Sebastian Borgeaud (formerly Gemini pre-training) focused on long-horizon coding tasks. Sergey Brin&#8217;s internal memo calls &#8220;turning our models into primary developers&#8221; the final sprint, and Google is tracking team-level usage of its internal coding tool &#8220;Jetski&#8221;&#8212;similar to Meta&#8217;s token leaderboard. Training runs on Google&#8217;s proprietary codebase.</p><p><strong>So What:</strong> Two signals for enterprise buyers. First, Google publicly concedes Anthropic is ahead on coding&#8212;which validates most engineering teams&#8217; current experience and shortens the &#8220;we should wait and see what Google ships&#8221; conversation. Second, the internal-tool-first strategy (Jetski) is telling: frontier labs are now treating their own engineers as the leading pilot cohort, and what ships publicly lags what&#8217;s running inside. That pattern will hold across every model family.</p><p><strong>Now What:</strong> If you&#8217;re picking a coding model or agent platform today, pick based on what works in your team&#8217;s actual workflows now, not on vendor roadmap slides. Re-evaluate quarterly&#8212;the leader-of-the-month dynamic is real, and Google catching up is now the explicit goal. For teams running on Gemini, ask your account team directly what Jetski&#8217;s usage looks like and when those capabilities ship externally.</p><p><a href="https://the-decoder.com/google-builds-elite-team-to-close-the-coding-gap-with-anthropic/">Read more</a></p><h2>Moonshot&#8217;s Kimi K2.6 Puts an Open-Source Model at the Frontier&#8212;For Long-Horizon Coding</h2><p><strong>What:</strong> Moonshot released Kimi K2.6, an open-weights coding model benchmarking neck-and-neck with GPT-5.4, Claude Opus 4.6, and Gemini 3.1 Pro on agentic and coding tasks. Vercel reports 50%+ gains on their Next.js benchmark. Demonstration runs include a 12-hour, 4,000-tool-call Zig inference optimization and a 13-hour autonomous rewrite of an 8-year-old matching engine (185% throughput gains). Agent Swarm now scales to 300 sub-agents across 4,000 coordinated steps.</p><p><strong>So What:</strong> This is the first time open weights sit inside the frontier envelope for long-horizon agent work. The implications go beyond price. Open weights mean you can host the model inside your own compliance boundary, run it offline in regulated environments, fine-tune on proprietary code without sending it to a vendor, and avoid per-token pricing on the workloads that burn the most budget. The benchmarks are vendor-run&#8212;take them with salt&#8212;but the customer quotes from Vercel, Fireworks, Baseten, Ollama, and others converge on one point: long-horizon reliability is now real on open weights.</p><p><strong>Now What:</strong> If you operate in a regulated environment or have workloads where data can&#8217;t leave your perimeter, re-open the build-versus-buy conversation on agent workloads. The calculus from a year ago&#8212;frontier models are only available as closed API products&#8212;is no longer true. Pilot K2.6 alongside your existing closed-model stack on one high-value, long-horizon workflow and compare on reliability, cost, and governance posture.</p><p><a href="https://www.kimi.com/blog/kimi-k2-6">Read more</a></p><h2>OpenAI Ships GPT-Rosalind&#8212;A Frontier Model for Life Sciences, With Named Pharma Launch Partners</h2><p><strong>What:</strong> OpenAI launched GPT-Rosalind, a frontier reasoning model for biology, drug discovery, and translational medicine, available in research preview through ChatGPT, Codex, and the API via a &#8220;trusted access program.&#8221; Launch customers include Amgen, Moderna, the Allen Institute, and Thermo Fisher. OpenAI is framing capabilities as muted today&#8212;synthesis, experimentation planning, research compilation&#8212;with autonomous scientific progress &#8220;several technical milestones away.&#8221;</p><p><strong>So What:</strong> This is the first vertical frontier model shipped by either major lab. OpenAI is betting the next phase of enterprise AI is specialized models with curated tool access, not general-purpose models doing everything. Life sciences is the first domain because the economics are obvious and the customer list was ready&#8212;expect similar vertical frontier launches in legal, finance, and clinical care over the next year. Notably absent from the launch customer list: payers, providers, and any non-pharma healthcare organization.</p><p><strong>Now What:</strong> If you&#8217;re in pharma, biotech, or translational medicine, ask OpenAI directly about the trusted access program&#8212;the published customer list tells you exactly who&#8217;s in the room. If you&#8217;re in adjacent regulated industries (healthcare payer/provider, legal, financial services), watch the trusted-access pattern carefully: this is likely the GTM template for every vertical frontier model that follows, and getting in early matters more than the model&#8217;s current capability ceiling.</p><p><a href="https://pitchbook.com/news/articles/openais-gpt-rosalind-heats-up-ai-competition-in-life-sciences">Read more</a></p><h1>The Enterprise Realities</h1><p><em>The same week three vendors reframed the workspace layer, three stories from the field reframed how you should actually buy and build. Proprietary formats are becoming liabilities as AI-native tools route around them. SpaceX on Cursor puts a reference customer on the table that answers the hardest security objection in any AI coding tool RFP. And a clean Tensorzero analysis shows that most enterprise AI budgets are built on list-price comparisons that are off by 2-5x. Your AI cost, tool choice, and vendor audit all need a refresh this quarter.</em></p><h2>Anthropic Ships Claude Design&#8212;And Figma&#8217;s Locked Format Has an Agentic-Era Problem</h2><p><strong>What:</strong> Anthropic launched Claude Design as part of Claude Labs&#8212;a generative design workflow that takes prompts to production-quality UI and interactive prototypes without leaving Claude. A widely-shared analysis from Sam Henri argues Figma&#8217;s largely-undocumented, hard-to-work-with-programmatically file format accidentally excluded Figma from the training data that would make it relevant in the agentic era.</p><p><strong>So What:</strong> The pattern matters beyond design. Every proprietary file format that&#8217;s hard to parse programmatically is now at risk of being routed around by AI-native tooling. Claude Design didn&#8217;t beat Figma on features&#8212;it made Figma&#8217;s closed format a liability instead of a moat. The same dynamic will play out for any vendor whose lock-in depends on an opaque format: BIM, CAD, proprietary PM tools, specialized ERP schemas. Open or interoperable formats gain value; closed formats become tech debt.</p><p><strong>Now What:</strong> If you maintain internal tools or vendor contracts that depend on a closed format, audit them. Ask whether the format is machine-readable, whether it&#8217;s documented, whether an AI agent could roundtrip through it. If the answer is no, start planning the migration now&#8212;not because AI replaces the tool tomorrow, but because the tool&#8217;s value compounds against you every quarter the agent layer gets better.</p><p><a href="https://www.anthropic.com/news/claude-design-anthropic-labs">Read more</a></p><h2>SpaceX Picks Cursor&#8212;Enterprise IDE Adoption at Scale</h2><p><strong>What:</strong> The New York Times reports SpaceX standardized on Cursor for engineering. Details on team size and license counts aren&#8217;t public, but SpaceX is one of the largest and most security-conscious software engineering organizations in the world, and the pick validates Cursor as an enterprise-grade tool rather than a startup productivity play.</p><p><strong>So What:</strong> This is the most significant enterprise reference for any AI coding tool to date. SpaceX&#8217;s security posture, classification requirements, and engineering culture make it an unusually strict buyer&#8212;the fact that Cursor cleared the bar tells you that enterprise-ready features (SSO, audit logs, IP protection, custom model routing, offline modes) have caught up to what large orgs need. Expect this reference to show up in every AI coding tool RFP this quarter.</p><p><strong>Now What:</strong> If you have engineers evaluating AI coding tools, the SpaceX reference gives your security team an answer to the hardest objection: &#8220;no one at our scale runs this yet.&#8221; That&#8217;s no longer true. If you&#8217;re at the enterprise buyer stage, ask each candidate vendor what their largest production customer looks like, what SOC 2 Type II evidence they can share, and what their model-routing and IP-protection story is. The answers have gotten meaningfully better in the last 90 days.</p><p><a href="https://www.nytimes.com/2026/04/21/business/spacex-cursor-deal.html">Read more</a></p><h2>Stop Comparing Price Per Million Tokens&#8212;Tokenization Can Make Claude 5x More Expensive Than the List Price Suggests</h2><p><strong>What:</strong> A Tensorzero analysis shows that because different models tokenize text differently, real-world cost can diverge sharply from list price. On some workloads, Claude tokens end up costing 5x more than GPT tokens despite Claude&#8217;s list price being only 2x. The gap is driven by how each tokenizer splits text&#8212;code, structured data, and non-English content all produce different token counts per byte.</p><p><strong>So What:</strong> Most AI budgets in enterprise are built on list-price comparisons that are off by 2&#8211;5x. That&#8217;s not a rounding error&#8212;it&#8217;s the difference between a model being affordable at scale and being cost-prohibitive. The broader point is that the economics of AI workloads aren&#8217;t legible from vendor pricing pages alone. Real cost depends on your actual text, your actual prompts, and your actual workflows&#8212;and it requires instrumentation to see.</p><p><strong>Now What:</strong> Before your next model-selection decision, run a representative 100-prompt sample through each candidate vendor, count tokens on both the input and output sides, and multiply by each vendor&#8217;s list price. Do this for every workload shape (code, structured data, long documents, conversational). You&#8217;ll almost certainly find that the &#8220;cheaper&#8221; model on the sticker is not the cheaper model in practice. Also: this is the single strongest argument for model-routing architecture&#8212;the right model for the workload beats the cheapest model by list price, every time.</p><p><a href="https://www.tensorzero.com/blog/stop-comparing-price-per-million-tokens-the-hidden-llm-api-costs/">Read more</a></p>]]></content:encoded></item><item><title><![CDATA[Welcome to the Great Reinvention]]></title><description><![CDATA[The work isn&#8217;t AI adoption, it&#8217;s the reinvention of how people and companies operate.]]></description><link>https://tsw.blankmetal.ai/p/welcome-to-the-great-reinvention</link><guid isPermaLink="false">https://tsw.blankmetal.ai/p/welcome-to-the-great-reinvention</guid><dc:creator><![CDATA[Blank Metal]]></dc:creator><pubDate>Thu, 23 Apr 2026 20:39:03 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!vuxY!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd4b337ee-5e5a-48ec-94e2-313930d09915_5644x3763.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!vuxY!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd4b337ee-5e5a-48ec-94e2-313930d09915_5644x3763.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!vuxY!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd4b337ee-5e5a-48ec-94e2-313930d09915_5644x3763.jpeg 424w, https://substackcdn.com/image/fetch/$s_!vuxY!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd4b337ee-5e5a-48ec-94e2-313930d09915_5644x3763.jpeg 848w, https://substackcdn.com/image/fetch/$s_!vuxY!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd4b337ee-5e5a-48ec-94e2-313930d09915_5644x3763.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!vuxY!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd4b337ee-5e5a-48ec-94e2-313930d09915_5644x3763.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!vuxY!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd4b337ee-5e5a-48ec-94e2-313930d09915_5644x3763.jpeg" width="1456" height="971" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/d4b337ee-5e5a-48ec-94e2-313930d09915_5644x3763.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:971,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:1864660,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://tsw.blankmetal.ai/i/195237264?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd4b337ee-5e5a-48ec-94e2-313930d09915_5644x3763.jpeg&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!vuxY!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd4b337ee-5e5a-48ec-94e2-313930d09915_5644x3763.jpeg 424w, https://substackcdn.com/image/fetch/$s_!vuxY!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd4b337ee-5e5a-48ec-94e2-313930d09915_5644x3763.jpeg 848w, https://substackcdn.com/image/fetch/$s_!vuxY!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd4b337ee-5e5a-48ec-94e2-313930d09915_5644x3763.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!vuxY!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd4b337ee-5e5a-48ec-94e2-313930d09915_5644x3763.jpeg 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>I listened to Nikhyl Singhal on Lenny&#8217;s podcast this week. It&#8217;s the most salient take I&#8217;ve heard in months on what&#8217;s actually happening in tech, and if you lead a company, hire product/design/tech people, or are trying to figure out what to do with the org you built over the last five years, you should listen to the whole thing before you read what follows.</p><p>His argument in one paragraph: the product management role is splitting in two. &#8220;Information movers,&#8221; whose day is framing and shuttling information up and down the org, are becoming dinosaurs. &#8220;Builders&#8221; who ship, prototype, and have direct product instincts are in a renaissance. Half the current PM population is in the first camp. The next 12&#8211;24 months will be the most chaotic period in PM history, with massive shedding and rehiring. Companies will let thousands of people go and rehire thousands of others, all AI-first, radically different skills, higher comp, everything different. The only way through is to cross a personal reinvention threshold and find a moment of joy in the new way of working.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://tsw.blankmetal.ai/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading The So What! Subscribe for free to receive new posts and support our work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p>Go listen. I&#8217;m not going to recap it. What follows is what it unlocked for me.</p><h3>The split is happening in every function</h3><p>Nikhyl was talking to PMs. I work with CEOs, COOs, and CPOs across the enterprise, and the builder / information-mover split isn&#8217;t a PM problem. It&#8217;s a knowledge-work problem.</p><p>The same split is showing up everywhere: marketing, sales ops, finance, HR, legal, customer operations, service delivery. Every function has a population of builders, people whose instinct is to ship, prototype, automate, and own outcomes, and a population of information movers, people whose value was routing, reframing, and coordinating. AI is eating the second group&#8217;s job description first, because that&#8217;s where the leverage is highest and the risk is lowest.</p><p>PMs are the canary. If you lead a non-product function and you&#8217;re watching this happen in product thinking &#8220;glad that&#8217;s not me,&#8221; then you&#8217;re not paying close enough attention.</p><h3>Companies have the same threshold to cross</h3><p>The most important idea in the episode is the reinvention threshold. Nikhyl&#8217;s point is that every knowledge worker right now has to make a very specific internal decision: <em>I am going to reinvent my craft, and I&#8217;m going to put that above the other things I&#8217;ve been protecting.</em> It&#8217;s not a training program. It&#8217;s not a mindset session. It&#8217;s a conscious reordering of priorities, and until you cross it, nothing else works. You can consume all the AI content you want and still be on the wrong side of the line.</p><p>What nobody is saying out loud is that companies have the exact same threshold. And most of them haven&#8217;t crossed it either.</p><p>What I see in enterprises right now is a lot of activity that looks like change and isn&#8217;t. AI strategy decks. Copilot pilots. Innovation sprints. Center-of-excellence PowerPoints. Real effort, almost none of it touching the thing that actually has to change: how work gets done, who does it, what gets paid for, and what gets measured.</p><p>Strategy without operating model change is theater. The companies that win the next two years are the ones whose CEOs look at their org chart, their process library, their vendor stack, and their job architecture and say &#8220;we are going to rebuild this,&#8221; not &#8220;we are going to layer AI on top of this.&#8221;</p><p>That&#8217;s the company-level threshold. It&#8217;s as scary as the individual one, because it means admitting that a lot of what got you here is what&#8217;s holding you back. Nikhyl calls this the &#8220;shadow superpower&#8221; &#8212; the skills and systems that made you successful in the last era are the exact thing blocking you from the next one. Shadow superpowers don&#8217;t just belong to senior ICs. They belong to entire operating models.</p><h3>The equal disappointment algorithm scales up</h3><p>Before the how-to: a word about the weight of the ask, because I don&#8217;t want it misread.</p><p>Nikhyl has a line about mid-career professionals in their &#8220;power years,&#8221; the decade or so when you&#8217;ve finally figured out your craft and the people around you demand the most of it, having eight hours of supply and twenty hours of demand: work, partner, kids, aging parents, health, friends. His framing is that your only workable strategy is to <em>equally disappoint everyone</em>, because you can&#8217;t meet full demand from any one constituency.</p><p>That&#8217;s the individual version. It&#8217;s also the CEO&#8217;s version. Every enterprise leader I talk to is running an equal-disappointment algorithm across their board, their customers, their employees, their regulators, and their own family.</p><p>But the algorithm already has a hierarchy built in. Your kids aren&#8217;t negotiable. Your partner isn&#8217;t a line item next to a quarterly review. Your health isn&#8217;t optional. The question isn&#8217;t who to disappoint to make room for reinvention. It&#8217;s which work actually matters, and which doesn&#8217;t.</p><p>You don&#8217;t steal hours from your kids. You steal them from the steering committee, the status report, the stakeholder tour, the deck review, the meeting that could have been an email. Most leaders never make that move because they&#8217;ve never explicitly ranked their work against itself. Everything at work feels load-bearing until you force yourself to look.</p><p>The reason most CEOs stall at the threshold isn&#8217;t that they don&#8217;t see it. It&#8217;s that they&#8217;re already maxed out keeping the current system running, and reinvention feels like one more thing to add on top. It isn&#8217;t. Trade work that doesn&#8217;t matter for it. That trade is hard, it&#8217;s political, and it&#8217;s the only one that actually works.</p><p>One more thing worth holding onto here: this chaos has an end. Nikhyl estimates about two years before the industry settles into a new operating equilibrium, with new rituals, new roles, new expectations. That&#8217;s the tunnel. It&#8217;s loud, it&#8217;s exhausting, and it ends. Companies that try to keep every work constituency happy through it are the ones that end up shedding thousands of employees without the newly shaped people rehired.</p><h3>What crossing the threshold actually looks like at scale</h3><p>If you run a 40,000-person enterprise, &#8220;walk into the tunnel&#8221; is not a plan. You can&#8217;t weekend-hack your way across this threshold. But the mechanics exist, and they&#8217;re more concrete than most transformation programs admit.</p><p>Four moves I see actually working at scale:</p><p><strong>Rewrite the job architecture, not just the training plan.</strong> Most enterprises are running AI upskilling programs against a job architecture designed for the information-mover era. You cannot reskill your way out of a structural mismatch. The work is to redefine what roles exist, what outcomes they own, and what &#8220;good&#8221; looks like in each, then reskill against the new architecture. Do it in the other order and you train people for jobs that don&#8217;t exist.</p><p><strong>Change what gets measured and what gets promoted.</strong> Your people read the signals you send through comp, promotion, and visibility. If your top performers are still the ones who ran the best steering committee, you&#8217;re telling the organization that the old game is still the game. Promote builders. Compensate for shipped outcomes. Make the signal impossible to miss.</p><p><strong>Put builders in the room where decisions get made.</strong> Most enterprises have builders, they&#8217;re just three layers below where strategy happens. Crossing the threshold means restructuring who&#8217;s in the room. The CEO&#8217;s staff meeting should include people who shipped something this week, not just people who manage people who manage people who shipped something.</p><p><strong>Pick one high-stakes area and rebuild it in public.</strong> Not a pilot. Not an innovation lab. A real function, real P&amp;L, real customers, real stakes, rebuilt from the operating model up, inside twelve months. It gives the rest of the organization a proof point they can touch, and it forces your executive team to confront the actual mechanics rather than debate them in the abstract.</p><p>None of this is easy. All of it is more concrete than &#8220;do AI transformation.&#8221; If you&#8217;re running a big company and you&#8217;re looking for where to start, start with one of these four.</p><h3>What I believe right now</h3><p>Six things I believe with more conviction after listening to this episode.</p><p><strong>Builders are the only hire that makes sense.</strong> For every seat: PM, engineer, marketer, ops leader, consultant, analyst. If the person you&#8217;re hiring can&#8217;t point to something they built in the last 90 days using modern tools, they are the old model. Don&#8217;t hire them.</p><p><strong>Hiring builders is the easy part. Keeping them is the real work.</strong> &#8220;Hire builders&#8221; is now conventional wisdom. The next failure mode, the one I&#8217;m watching play out in real time, is companies that hired builders and then dropped them into an information-mover operating model. Weekly status decks. Three-week PRD review cycles. Approval chains requiring four directors to sign off on a prototype. Builders in that environment quit inside a year. They don&#8217;t send a note; they just ship their resume to the next place. If your org has started hiring builders but hasn&#8217;t changed its rituals, measurement, or decision rights to match, you&#8217;re running the most expensive revolving door in the market.</p><p><strong>Young talent is a cheat code, and most companies are ignoring it.</strong> I came up in an apprenticeship culture, and I think the industry forgot how valuable that is. The people with the least to unlearn are the ones who never learned the old way. A 23-year-old who came up building with modern tools, who doesn&#8217;t know what a PRD review cycle is supposed to look like, who treats Claude Code the way my generation treated email: that person has an aptitude advantage no amount of senior pattern-matching can replicate. Diversity isn&#8217;t just gender, race, and geography. It&#8217;s age. Companies only hiring fifteen-year-vets with the &#8220;right&#8221; logos are missing the single most obvious arbitrage available to them. Pair young builders with senior judgment and you get a team that moves at a pace the old model physically cannot produce.</p><p><strong>Joy is the unlock.</strong> Nikhyl&#8217;s &#8220;moment of joy&#8221; framing is the single most useful piece of practical advice I&#8217;ve heard on how to get people through this, and it&#8217;s more specific than it sounds. He&#8217;s noticed that every person who crosses the threshold has the same kind of story: they built a small thing with modern tools and it worked. A chief-of-staff app for their inbox. A script that controls their house lights. Helped their spouse test-market a business idea. Stayed up too late one night getting something to run. Small, personal, concrete, theirs. And from that moment forward they&#8217;re hooked. You cannot think your way across the reinvention threshold. You have to build something small, have it work, and catch the bug. Every leader, every team, every person has to have that moment. Enablement that doesn&#8217;t engineer it is wasted money.</p><p><strong>Pace is retention.</strong> Nikhyl calls it &#8220;fire in the belly.&#8221; Year-one energy, not year-five. Leaders who still operate at enterprise cadence in an AI-era market aren&#8217;t just slow; they&#8217;re actively signaling to their best builders that this isn&#8217;t the place. Your best people leave for pace before they leave for comp.</p><p><strong>The consulting model that built the last era doesn&#8217;t fit this one.</strong> Big decks, slow engagements, armies of juniors producing frameworks: that model was built for information movers, and it&#8217;s going to get gutted. The consulting that matters now is small teams of builders embedded alongside client teams, shipping working systems in weeks. That&#8217;s the Blank Metal bet, and I&#8217;m more certain of it this week than I was last week.</p><h3>Where I land</h3><p>Toward the end of the conversation, Lenny drops a line that&#8217;s been in my head since: <em>chaos is a ladder</em>, from Game of Thrones. That&#8217;s what this moment is. The people and companies most stressed right now are the ones clinging to the old shape. The people and companies having the most fun are the ones who crossed the threshold, caught the bug, and are climbing.</p><p>Whether you&#8217;re a CEO with 40,000 people, a founder of fifteen, or one person sitting at your desk wondering if you&#8217;re already behind: the tunnel is two years. Walk into it. Find your moment of joy. Build something this weekend that would have taken you a month a year ago. Trade work that doesn&#8217;t matter to make time for it.</p><p>It&#8217;s worth it.</p><p>That&#8217;s why I&#8217;m naming this moment the Great Reinvention: the work isn&#8217;t AI adoption, it&#8217;s reinvention of how people and companies operate.</p><p>Welcome to the Great Reinvention.</p><p><em>Nikhyl&#8217;s episode is <a href="https://www.lennysnewsletter.com/p/why-half-of-product-managers-are-in-trouble">Why half of product managers are in trouble</a> on Lenny&#8217;s Podcast. If you only have 95 minutes this month, spend it there.</em></p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://tsw.blankmetal.ai/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading The So What! Subscribe for free to receive new posts and support our work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[Weekly Headlines: Issue #18]]></title><description><![CDATA[April 9 - April 16, 2026]]></description><link>https://tsw.blankmetal.ai/p/weekly-headlines-issue-18</link><guid isPermaLink="false">https://tsw.blankmetal.ai/p/weekly-headlines-issue-18</guid><dc:creator><![CDATA[Blank Metal]]></dc:creator><pubDate>Fri, 17 Apr 2026 18:35:33 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!y9F1!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcecd5227-f359-49c0-8e16-1ab06a2755dd_1200x670.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!y9F1!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcecd5227-f359-49c0-8e16-1ab06a2755dd_1200x670.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!y9F1!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcecd5227-f359-49c0-8e16-1ab06a2755dd_1200x670.png 424w, https://substackcdn.com/image/fetch/$s_!y9F1!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcecd5227-f359-49c0-8e16-1ab06a2755dd_1200x670.png 848w, https://substackcdn.com/image/fetch/$s_!y9F1!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcecd5227-f359-49c0-8e16-1ab06a2755dd_1200x670.png 1272w, https://substackcdn.com/image/fetch/$s_!y9F1!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcecd5227-f359-49c0-8e16-1ab06a2755dd_1200x670.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!y9F1!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcecd5227-f359-49c0-8e16-1ab06a2755dd_1200x670.png" width="1200" height="670" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/cecd5227-f359-49c0-8e16-1ab06a2755dd_1200x670.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:670,&quot;width&quot;:1200,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:1480855,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://tsw.blankmetal.ai/i/194443360?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcecd5227-f359-49c0-8e16-1ab06a2755dd_1200x670.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!y9F1!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcecd5227-f359-49c0-8e16-1ab06a2755dd_1200x670.png 424w, https://substackcdn.com/image/fetch/$s_!y9F1!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcecd5227-f359-49c0-8e16-1ab06a2755dd_1200x670.png 848w, https://substackcdn.com/image/fetch/$s_!y9F1!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcecd5227-f359-49c0-8e16-1ab06a2755dd_1200x670.png 1272w, https://substackcdn.com/image/fetch/$s_!y9F1!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcecd5227-f359-49c0-8e16-1ab06a2755dd_1200x670.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Welcome to Blank Metal&#8217;s Weekly AI Headlines.</p><p>Each week, our team shares the AI stories that caught our attention&#8212;the articles, announcements, and insights we&#8217;re actually discussing internally. We curate the best of what we&#8217;re reading and add the context that matters: what happened, why it matters, and what to do about it.</p><p>Short, sharp, and focused on impact.</p><h1>The Governance Era Begins</h1><p><em>This week, the enterprise AI rollout story finally caught up with the capability story. Cowork went GA with the six admin controls IT teams have been waiting for. Ramp showed what the next phase looks like when large companies don&#8217;t wait for vendor tooling. And Gallup data made it clear that adoption without workflow redesign isn&#8217;t actually transformation&#8212;it&#8217;s fancy autocomplete with the same org chart.</em></p><h2>Claude Cowork Goes GA&#8212;With the Six Admin Controls Enterprise IT Was Waiting For</h2><p><strong>What:</strong> Anthropic shipped Claude Cowork to general availability on April 9, packaged with six new enterprise controls: Role-Based Access Control (RBAC) with SCIM integration, group spend limits with analytics, per-tool MCP connector permissions, skill sharing toggles (individual and org-wide, off by default), OpenTelemetry observability, and a native Zoom MCP connector. Cowork is now available across macOS and Windows on all paid Claude plans&#8212;Pro, Max, Team, and Enterprise.</p><p><strong>So What:</strong> Cowork was interesting in preview. Now it&#8217;s deployable. The admin controls were the blockers&#8212;IT teams couldn&#8217;t approve Cowork without per-user spend caps, audit trails, and granular connector permissions. Those shipped in one release. Anthropic is signaling that the enterprise rollout path is now fully paved: group-based access via your identity provider, observability into your existing monitoring stack, auditable connector behavior, and spend visibility at the team level. The governance story finally caught up with the capability story.</p><p><strong>Now What:</strong> If you&#8217;ve been holding off on Cowork because of governance gaps, that position just changed. Start with RBAC design&#8212;map your org structure to groups, set differentiated spend caps (investment team higher, support staff lower), enable individual skill sharing but hold org-wide skill promotion until you&#8217;ve vetted the first twenty. Wire OpenTelemetry into your existing SIEM so security gets the audit trail they need without building custom integrations.</p><p><a href="https://thenewstack.io/anthropic-takes-claude-cowork-out-of-preview-and-straight-into-the-enterprise/">Read more</a></p><h2>Ramp Built Its Own Claude Cowork Internally&#8212;a Pattern to Watch</h2><p><strong>What:</strong> Ramp engineering shared that they built a Claude Cowork-equivalent internal product to accelerate AI adoption across the company. Rather than waiting for vendor tooling to mature or letting every team build their own, Ramp centralized on a single internal surface with Ramp-specific context, skills, and connectors baked in.</p><p><strong>So What:</strong> This is the pattern to watch. Large tech-forward companies aren&#8217;t waiting for Claude, Copilot, or ChatGPT to ship the exact enterprise experience they want&#8212;they&#8217;re building the last-mile platform internally, wrapping vendor APIs with their own data, identity, and workflows. For teams without Ramp-level engineering capacity, the implication is different: wait for the enterprise features to ship (they just did, with Cowork GA), or partner with someone who can build the adoption layer without hiring a platform team.</p><p><strong>Now What:</strong> If your adoption is stalled because Cowork doesn&#8217;t know your codebase, ticketing system, or vendor contracts, the fix is a skill library and MCP servers&#8212;not a wait for Anthropic to ship a feature. Prioritize the five to ten highest-value workflows, build skills against them, deploy to a champion group, measure repeat usage. That&#8217;s the Ramp path, scaled down.</p><p><a href="https://x.com/sebgoddijn/status/2042285915435937816">Read more</a></p><h2>Gallup: Half of US Workers Use AI&#8212;Only 1 in 10 Say Work Has Transformed</h2><p><strong>What:</strong> New Gallup data shows 50% of US workers now use AI tools at work. Inside adopting organizations, 65% say AI helps productivity. The finding that matters most: only 1 in 10 workers strongly agree their work has actually transformed because of AI. Healthcare workers were flagged as early leaders in productivity gains. Large organizations (10K+ employees) with AI adoption are the only segment showing net workforce reductions&#8212;meaning they&#8217;re cutting heads before doing the redesign work.</p><p><strong>So What:</strong> The gap between &#8220;I use ChatGPT&#8221; and &#8220;we redesigned our workflows&#8221; is where the enterprise AI transformation actually lives. Adoption has won; redesign has not. Most companies are layering AI onto existing processes instead of rethinking them. The large-org data point is sobering&#8212;organizations cutting workforce ahead of the redesign are likely creating fragility, not efficiency. The companies pulling ahead over the next 18 months will be the ones treating AI as a workflow redesign problem, not a tool rollout problem.</p><p><strong>Now What:</strong> Audit where AI actually lands on your team today. If it&#8217;s individual productivity gains on the same processes, you&#8217;re in the 9-in-10 majority. Pick one cross-functional workflow per quarter to genuinely redesign&#8212;remove steps, change roles, measure cycle time. That&#8217;s how the 10% who report real transformation got there.</p><p><a href="https://www.gallup.com/workplace/704225/rising-adoption-spurs-workforce-changes.aspx">Read more</a></p><h1>Models: Cheaper, Opener, Everywhere</h1><p><em>The model layer commoditized further this week. Tokens are down 300x in three years. An open-weight agent model matched proprietary frontier performance on coding benchmarks&#8212;and did it by training itself. Google rounded out the set of every major lab shipping a native Mac app with a global keyboard shortcut. The model is the runtime. The value is moving up the stack.</em></p><h2>MiniMax Open-Sources M2.7&#8212;a Model That Helped Train Itself</h2><p><strong>What:</strong> MiniMax released M2.7, a Mixture-of-Experts agent model with open weights on HuggingFace. It scores 56% on SWE-Pro (matching GPT-5.3-Codex) and 57% on Terminal Bench 2. The notable detail: M2.7 actively participated in its own training, running 100+ autonomous rounds of scaffold optimization and iterating on its own RL pipeline. Built around three capability pillars&#8212;software engineering, office work, and native multi-agent collaboration (&#8221;Agent Teams&#8221;).</p><p><strong>So What:</strong> Two things matter here. First, the MoE architecture makes M2.7 significantly cheaper to serve than a dense model at comparable quality, which lowers the floor for self-hosted agent infrastructure. Second, the self-evolution loop is a new category of news: a model used its own agent capabilities to make itself better during training. That feedback loop compresses timelines for anyone building on open models and raises an uncomfortable question for proprietary labs&#8212;when does the frontier lead stop being meaningful if open models can self-improve?</p><p><strong>Now What:</strong> If you&#8217;re evaluating whether to build on open-weight models for cost, data-residency, or vendor-independence reasons, M2.7 is a credible alternative for agentic and coding work. Test it against your specific workloads before assuming proprietary models are required. For strategic planning, assume the open-vs-closed gap shrinks faster through 2026-2027 than current roadmaps predict.</p><p><a href="https://github.com/MiniMax-AI/MiniMax-M2">Read more</a></p><h2>&#8220;AI Models Are the New Rebar&#8221;&#8212;Tokens Dropped 300x in 36 Months</h2><p><strong>What:</strong> A widely-shared essay by Philipp Dubach argues that AI models have become infrastructure commodities&#8212;like rebar in construction. Tokens have dropped roughly 300x in price over 36 months. Open-source models continue closing on proprietary frontier performance quarter over quarter. The thesis: AI lab margins will compress as models become interchangeable components within larger systems, and the value moves up the stack to workflows, data, evaluations, and domain expertise.</p><p><strong>So What:</strong> The commoditization argument isn&#8217;t new, but the 300x data point is striking enough to change the conversation. If models are becoming rebar, your switching costs between Claude, GPT, Gemini, Llama, and MiniMax are going to keep falling. The lock-in lives in your skills, your MCP servers, your evaluations, and your domain-specific prompts&#8212;not in any single model. Lab valuations priced on a perpetual frontier lead look increasingly exposed.</p><p><strong>Now What:</strong> Design your AI architecture to swap models without re-architecting. Keep evaluations that compare multiple providers on your specific workloads, and re-run them quarterly. The teams that treat model choice as a quarterly re-bid rather than a wedding will move faster and spend less over the next two years.</p><p><a href="https://philippdubach.com/posts/ai-models-are-the-new-rebar/">Read more</a></p><h2>Google Launches Native Gemini for macOS&#8212;Every Frontier Lab Now Has a Desktop App</h2><p><strong>What:</strong> Google released a native Gemini app for macOS on April 15. It activates with Option+Space for quick queries, Option+Shift+Space for the full chat window, and sits in the Dock and Menu Bar. The UX pattern mirrors Claude&#8217;s desktop app and ChatGPT&#8217;s Mac app, both of which launched earlier.</p><p><strong>So What:</strong> Every major frontier lab now has a native Mac app with a global keyboard shortcut. This isn&#8217;t a product announcement&#8212;it&#8217;s a pattern announcement. The interface for AI is consolidating around &#8220;instant-on assistant accessible anywhere on your machine,&#8221; and the keyboard-shortcut pattern has quietly become a standard. For organizations managing AI rollout, this matters because your users are about to have three or four AI models one keystroke away&#8212;some approved, some not.</p><p><strong>Now What:</strong> Update your endpoint management policy to account for AI desktop apps. If you allow Claude desktop but not ChatGPT or Gemini desktop, make that explicit and enforce it&#8212;Mac app installs are the new shadow-IT vector. For teams intentionally using multiple models, standardize which keyboard shortcut maps to which model so users don&#8217;t accidentally route sensitive context to the wrong system.</p><p><a href="https://www.macrumors.com/2026/04/15/google-gemini-mac-app/">Read more</a></p><h1>The Practitioner Toolkit Fills In</h1><p><em>Every week, the tooling and mental models for people actually building with AI get a little better. This week: a metaphor for agents that survives a conversation with your CFO, a design skill that lifts the quality ceiling for AI-built UI, a podcast for engineering leaders shipping real agents, and a reminder that teams working on long-horizon AI work need morale infrastructure the same way they need CI/CD.</em></p><h2>&#8220;The Folder Is the Agent&#8221;&#8212;A Better Mental Model for Non-Technical Leaders</h2><p><strong>What:</strong> An Every essay reframes what an AI agent actually is by anchoring on a practical metaphor: a folder. A folder contains files (context), instructions (the goal), a history of prior work (memory), and permissions (tools). Agents are just folders that can read, write, and talk. The framing is deliberately non-technical, aimed at people leading AI rollouts who need to explain agents to operational leaders without drowning them in architectural jargon.</p><p><strong>So What:</strong> The &#8220;folder is the agent&#8221; framing is useful precisely because it&#8217;s legible to finance, legal, and ops leaders who actually decide whether AI rollouts scale. Most agent descriptions&#8212;&#8221;orchestrated tool-using autonomous systems with hierarchical delegation&#8221;&#8212;don&#8217;t survive a first meeting with a procurement lead. This one does. And it maps cleanly onto Cowork&#8217;s actual architecture: skills live in folders, context lives in folders, your work product lives in folders.</p><p><strong>Now What:</strong> If you&#8217;re building an AI rollout narrative for non-technical leadership, borrow the folder metaphor. It collapses the explanation from a whiteboard session to a sentence. When stakeholders understand that an agent is a folder with permissions and instructions, the governance conversation gets easier&#8212;they already understand folder permissions.</p><p><a href="https://every.to/source-code/the-folder-is-the-agent">Read more</a></p><h2>Impeccable&#8212;a Design Skill for AI-Assisted UI Work</h2><p><strong>What:</strong> Impeccable is a design skill built for Claude Code and Cowork that produces well-designed websites without requiring a dedicated designer in the loop. The skill encodes visual design heuristics, layout patterns, typography defaults, and accessibility rules into something an agent can apply during build.</p><p><strong>So What:</strong> Skills like Impeccable are the answer to &#8220;AI can code but the output looks AI-slop.&#8221; The quality ceiling for AI-generated frontend work is moving up as more design expertise gets captured as shareable skills. That shifts the build-vs-buy calculus for internal tools&#8212;the distance between &#8220;rough prototype&#8221; and &#8220;looks intentional&#8221; is shrinking. Teams without design capacity can now produce credible UI work by combining model capability with domain-specific skills.</p><p><strong>Now What:</strong> If your team ships internal tools or admin panels, test Impeccable on a throwaway project first. The more durable lesson is structural&#8212;start a library of skills that encode your organization&#8217;s design language (typography, spacing, component patterns) so every AI-built tool looks like it belongs to you, not to a generic model.</p><p><a href="https://impeccable.style/">Read more</a></p><h2>LangChain Launches &#8220;Max Agency&#8221;&#8212;A Podcast About Building Real Agents</h2><p><strong>What:</strong> Harrison Chase, LangChain founder, launched Max Agency, a new podcast focused on how production agents are actually built. Each episode features engineering leaders deep in the work: architecture decisions, evaluation frameworks, tradeoffs between speed and reliability, and the messy real-world choices that don&#8217;t show up in blog posts.</p><p><strong>So What:</strong> The builder conversation in AI is fragmenting across Twitter, Substack, YouTube, and podcasts&#8212;and most of the practical signal is buried in two-hour conversations you don&#8217;t have time to sift. A curated podcast from the founder of the most-used agent framework is worth the subscription. Agent architecture patterns are still being invented in public, and the teams shipping them are often the ones producing the most useful content.</p><p><strong>Now What:</strong> If you&#8217;re leading an engineering team building agents, add Max Agency to your technical reading. Treat episode notes as material worth circulating to the team&#8212;the decision-making frameworks travel better than any specific tech stack.</p><p><a href="https://www.youtube.com/watch?v=Xyh1EqcjGME">Read more</a></p><h2>LessWrong on Morale: What Happens When Feedback Loops Stretch Into Months</h2><p><strong>What:</strong> A widely-shared LessWrong essay examines how teams maintain morale when working on problems with severely time-delayed feedback&#8212;AI research, long-horizon engineering, ambiguous transformation work. The argument: conventional project management assumes short feedback loops; when the loop stretches to months or years, morale needs its own infrastructure.</p><p><strong>So What:</strong> Most serious enterprise AI work fits this pattern. You&#8217;re redesigning workflows, building skill libraries, wiring up MCP servers&#8212;producing value that compounds over quarters, not sprints. The familiar &#8220;demo and deploy&#8221; cadence doesn&#8217;t fit. If your team&#8217;s morale is tied entirely to shipping velocity and the real payoff is further out, you&#8217;ll see burnout and attrition before you see results. The fix isn&#8217;t shipping faster&#8212;it&#8217;s building internal signals that validate progress without waiting for the ultimate outcome.</p><p><strong>Now What:</strong> If you lead a team on a long-horizon AI initiative, invent internal milestones that aren&#8217;t tied to end-user adoption. Shipping a new skill to the library counts. Hitting the first ten users of a new workflow counts. Celebrate those, visibly. Your team is working on a problem whose payoff is further away than what they&#8217;re used to&#8212;your job is to keep them pointed at the horizon without burning out on the walk.</p><p><a href="https://www.lesswrong.com/posts/53ZAzbdzGJHGeE5rs/morale">Read more</a></p>]]></content:encoded></item><item><title><![CDATA[If You're Still Chatting With AI, There's a Better Way to Work]]></title><description><![CDATA[Everyone has AI access now.]]></description><link>https://tsw.blankmetal.ai/p/if-youre-still-chatting-with-ai-theres</link><guid isPermaLink="false">https://tsw.blankmetal.ai/p/if-youre-still-chatting-with-ai-theres</guid><dc:creator><![CDATA[Blank Metal]]></dc:creator><pubDate>Thu, 16 Apr 2026 19:21:17 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!BDNr!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fca2e3132-4793-44fb-89b6-29f7c741f4c6_6000x4000.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!BDNr!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fca2e3132-4793-44fb-89b6-29f7c741f4c6_6000x4000.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!BDNr!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fca2e3132-4793-44fb-89b6-29f7c741f4c6_6000x4000.jpeg 424w, https://substackcdn.com/image/fetch/$s_!BDNr!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fca2e3132-4793-44fb-89b6-29f7c741f4c6_6000x4000.jpeg 848w, https://substackcdn.com/image/fetch/$s_!BDNr!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fca2e3132-4793-44fb-89b6-29f7c741f4c6_6000x4000.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!BDNr!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fca2e3132-4793-44fb-89b6-29f7c741f4c6_6000x4000.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!BDNr!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fca2e3132-4793-44fb-89b6-29f7c741f4c6_6000x4000.jpeg" width="1456" height="971" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/ca2e3132-4793-44fb-89b6-29f7c741f4c6_6000x4000.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:971,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:1510302,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://tsw.blankmetal.ai/i/194441797?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fca2e3132-4793-44fb-89b6-29f7c741f4c6_6000x4000.jpeg&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!BDNr!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fca2e3132-4793-44fb-89b6-29f7c741f4c6_6000x4000.jpeg 424w, https://substackcdn.com/image/fetch/$s_!BDNr!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fca2e3132-4793-44fb-89b6-29f7c741f4c6_6000x4000.jpeg 848w, https://substackcdn.com/image/fetch/$s_!BDNr!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fca2e3132-4793-44fb-89b6-29f7c741f4c6_6000x4000.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!BDNr!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fca2e3132-4793-44fb-89b6-29f7c741f4c6_6000x4000.jpeg 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Everyone has AI access now. ChatGPT, Gemini, Claude &#8212; pick your flavor. And many people use it the same way: open a chat window, type a question, get an answer, copy it into a doc or email, close the tab.</p><p>That&#8217;s useful. It&#8217;s also a ceiling.</p><p>In January, Anthropic launched <strong>Claude Cowork</strong> &#8212; and it&#8217;s a BIG shift. Not a new model. A new way of working. Within three months, Anthropic&#8217;s revenue more than doubled. Non-engineering teams became the majority of enterprise Cowork usage. Kate Jensen, Anthropic&#8217;s Head of Americas: &#8220;In 2025 Claude transformed how developers work, and in 2026 it will do the same for knowledge work.&#8221;</p><p>Here&#8217;s what&#8217;s actually happening.</p><h2><strong>You Don&#8217;t Install AI, You Onboard It</strong></h2><p>People evaluate AI the way they evaluate a new SaaS tool. Which one should I buy? How does it integrate? What are the features?</p><p>Wrong question! You onboard AI the same way you&#8217;d onboard a capable new analyst: set expectations, give context, share the relevant files, explain how you like things structured, review the work. Push back when it&#8217;s not right.</p><p>The prompt has become the least important part. The context &#8212; who you are, what you&#8217;re working on, what good looks like &#8212; that&#8217;s what determines output quality. Once you onboard it, it doesn&#8217;t forget. And it gets better every time you refine the instructions.</p><h2><strong>The Empty Workshop</strong></h2><p>When you type into ChatGPT with no files, no context, and no connections to your actual work, that&#8217;s a workshop with no tools on the wall. You can do some things with your hands, but you&#8217;re leaving most of the capability out of it.</p><p>Claude Cowork is where you put the tools on the wall. Connectors plug into the systems you actually use, like Gmail, Calendar, Salesforce, Slack, Google Drive. Skills capture how you like work done. Projects hold your files and context across sessions. A plugin marketplace organized by department means you don&#8217;t start from scratch.</p><p>Claude Code proved this architecture for developers &#8212; 1.6 million weekly active users, authoring 4% of all public GitHub commits. Cowork brings it to everyone else.</p><h2><strong>The Moment It Clicks</strong></h2><p>We&#8217;ve trained hundreds of people on Cowork across the country in the last five weeks &#8212; PE firms, software companies, security teams, financial services. There&#8217;s a moment in every session where the room shifts.</p><p>It&#8217;s when someone connects their email and calendar and asks: <em>&#8220;What&#8217;s on my calendar tomorrow and are there any emails I should read before those meetings?&#8221;</em></p><p>One question. All their context. One answer.</p><p>Right now, you are the integration layer. You context-switch between tabs, mentally cross-reference, and assemble the picture yourself. That question eliminates all of it. They&#8217;re not chatting with AI anymore. They&#8217;re plugging their world into something that can operate on it.</p><h2><strong>It&#8217;s Not About Saving Time; It&#8217;s About Changing What&#8217;s Possible</strong></h2><p>Anthropic calls it &#8220;the thinking divide&#8221; &#8212; the gap between organizations that embed AI across their workforce and those that treat it as a point solution.</p><p>When something gets easier, you don&#8217;t do less of it. You do more of what matters. A RevOps lead who spent 12 hours every Monday building a deck from Salesforce data built a skill that does it in minutes. She didn&#8217;t save Monday. She got Monday back for strategy. A sales rep runs every call transcript through a qualification skill that captures institutional knowledge. He didn&#8217;t automate a task. He made the entire team smarter.</p><p>Not efficiency. Capability.</p><h2><strong>How to Start</strong></h2><p>Don&#8217;t buy 50 licenses and send a &#8220;go explore!&#8221; email. Kate Jensen again: enterprise AI in 2025 &#8220;turned out to be mostly premature&#8221; with pilots failing to reach production. &#8220;It wasn&#8217;t a failure of effort, it was a failure of approach.&#8221;</p><p>Start with a handful of people who have work that&#8217;s repetitive, data-heavy, or crosses multiple systems. Train them on how to connect their data, build their first skill, and produce something they&#8217;d actually use tomorrow. Let them become the proof point for the rest of the organization.</p><p>70% of the Fortune 100 already uses Claude. The companies that moved early on Cowork are already compounding. The question isn&#8217;t whether your organization will adopt this way of working. It&#8217;s whether you&#8217;ll be on the right side of the thinking divide when it does.</p>]]></content:encoded></item><item><title><![CDATA[Weekly Headlines: Issue #17]]></title><description><![CDATA[April 2 - 9, 2026]]></description><link>https://tsw.blankmetal.ai/p/weekly-headlines-issue-17</link><guid isPermaLink="false">https://tsw.blankmetal.ai/p/weekly-headlines-issue-17</guid><dc:creator><![CDATA[Blank Metal]]></dc:creator><pubDate>Fri, 10 Apr 2026 14:18:24 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!KfZk!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbe5306b4-1cc1-4e39-8c6b-ded4615bf0d5_1200x670.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!KfZk!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbe5306b4-1cc1-4e39-8c6b-ded4615bf0d5_1200x670.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!KfZk!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbe5306b4-1cc1-4e39-8c6b-ded4615bf0d5_1200x670.png 424w, https://substackcdn.com/image/fetch/$s_!KfZk!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbe5306b4-1cc1-4e39-8c6b-ded4615bf0d5_1200x670.png 848w, https://substackcdn.com/image/fetch/$s_!KfZk!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbe5306b4-1cc1-4e39-8c6b-ded4615bf0d5_1200x670.png 1272w, https://substackcdn.com/image/fetch/$s_!KfZk!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbe5306b4-1cc1-4e39-8c6b-ded4615bf0d5_1200x670.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!KfZk!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbe5306b4-1cc1-4e39-8c6b-ded4615bf0d5_1200x670.png" width="1200" height="670" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/be5306b4-1cc1-4e39-8c6b-ded4615bf0d5_1200x670.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:670,&quot;width&quot;:1200,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:1480493,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://tsw.blankmetal.ai/i/193787492?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbe5306b4-1cc1-4e39-8c6b-ded4615bf0d5_1200x670.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!KfZk!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbe5306b4-1cc1-4e39-8c6b-ded4615bf0d5_1200x670.png 424w, https://substackcdn.com/image/fetch/$s_!KfZk!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbe5306b4-1cc1-4e39-8c6b-ded4615bf0d5_1200x670.png 848w, https://substackcdn.com/image/fetch/$s_!KfZk!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbe5306b4-1cc1-4e39-8c6b-ded4615bf0d5_1200x670.png 1272w, https://substackcdn.com/image/fetch/$s_!KfZk!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbe5306b4-1cc1-4e39-8c6b-ded4615bf0d5_1200x670.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Welcome to Blank Metal&#8217;s Weekly AI Headlines.</p><p>Each week, our team shares the AI stories that caught our attention&#8212;the articles, announcements, and insights we&#8217;re actually discussing internally. We curate the best of what we&#8217;re reading and add the context that matters: what happened, why it matters, and what to do about it.</p><p>Short, sharp, and focused on impact.</p><h1>Security Is the New Capability Story</h1><p><em>This week&#8217;s biggest AI news wasn&#8217;t about making models smarter&#8212;it was about making systems safer. Anthropic weaponized a frontier model for defense, the FT mapped how trust is splitting the agent market, and a six-minute social engineering attack showed that the most dangerous vulnerabilities aren&#8217;t in the code.</em></p><h2>Anthropic Unveils Claude Mythos Preview&#8212;and Won&#8217;t Release It</h2><p><strong>What:</strong> Anthropic revealed Claude Mythos Preview, a frontier model capable of autonomously finding and exploiting zero-day vulnerabilities in every major operating system and web browser. Rather than releasing it broadly, Anthropic launched Project Glasswing&#8212;a defensive initiative partnering with AWS, Apple, Google, Microsoft, CrowdStrike, NVIDIA, and others to use Mythos Preview exclusively for securing critical software. The model has already discovered thousands of previously unknown vulnerabilities, including a 27-year-old remote code execution flaw in FreeBSD. Anthropic is committing $100M in usage credits and $4M in donations to open-source security organizations, with a public disclosure report due within 90 days.</p><p><strong>So What:</strong> This is Anthropic making a statement about capability responsibility. They built a model that scores 93.9% on SWE-bench Verified (vs. 80.8% for Opus 4.6) and can single-handedly find bugs that human researchers missed for decades&#8212;and their response was to restrict access and build a coalition around defensive use. The model won&#8217;t be released publicly. Instead, what Anthropic learns from Mythos will inform safeguards built into the next Opus release. For enterprises, the implication is clear: if today&#8217;s models can find vulnerabilities at this scale, the next generation&#8212;including models adversaries will build&#8212;will do far more.</p><p><strong>Now What:</strong> Security teams should start planning for a world where both attackers and defenders have models this capable. The window before offensive equivalents emerge is short. If you&#8217;re running legacy systems in healthcare, financial services, or government, your attack surface just became more exposed than you thought. &#8220;We&#8217;ll get to security later&#8221; is no longer a viable position.</p><p><a href="https://www.anthropic.com/glasswing">Read more</a></p><h2>Financial Times: AI Agent Market Is Splitting Along Trust Lines</h2><p><strong>What:</strong> A Financial Times deep dive on AI agents reveals the market is splitting into two camps. Regulated industries&#8212;law, finance, cybersecurity, healthcare&#8212;are demanding accuracy and accountability over speed. They want human-in-the-loop, audit trails, and explainable decisions. Meanwhile, less-regulated sectors are racing ahead with fully autonomous agents. The divide isn&#8217;t about capability&#8212;it&#8217;s about trust infrastructure.</p><p><strong>So What:</strong> This validates what anyone working in regulated verticals already knows: the bottleneck isn&#8217;t AI capability, it&#8217;s governance and accountability. FINRA&#8217;s 2026 oversight report flagged agents operating without human validation, acting beyond intended scope, and making unexplainable decisions as top governance risks. The companies winning in regulated markets aren&#8217;t the ones with the best models&#8212;they&#8217;re the ones with the best implementation and domain expertise.</p><p><strong>Now What:</strong> If you&#8217;re working in regulated industries, lead with governance, not capability. The model is a commodity. The key to success is understanding compliance requirements, building audit trails, and knowing where human-in-the-loop is legally required versus where it&#8217;s just organizational inertia. </p><p><a href="https://www.ft.com/content/72c20f77-e85d-49cb-84ef-4b676244d1c5">Read more</a></p><h2>Supply Chain Attack on Axios Shows How Sophisticated Social Engineering Has Become</h2><p><strong>What:</strong> Attackers compromised a core Axios maintainer through an elaborate social engineering campaign. They impersonated a company founder, created a convincing Slack workspace with fake employee profiles and LinkedIn content, and scheduled a Microsoft Teams call with what appeared to be a real team. During the call, the maintainer installed what seemed like a Teams update&#8212;actually a Remote Access Trojan. The entire attack from first contact to credential compromise took six minutes.</p><p><strong>So What:</strong> This isn&#8217;t a technical vulnerability&#8212;it&#8217;s a human one, and it targets the open-source maintainers that the entire software supply chain depends on. The sophistication is what&#8217;s alarming: cloned visual identities, professional-grade Slack workspaces, coordinated fake personas. Every maintainer of a widely-used package is now a high-value target. Traditional security training (&#8221;don&#8217;t click suspicious links&#8221;) doesn&#8217;t cover social engineering this polished.</p><p><strong>Now What:</strong> For engineering teams, audit your supply chain dependencies for single-maintainer risks. For security teams, recognize that social engineering attacks are now being run with the production quality of a marketing campaign. The six-minute attack window suggests this is operationalized, not experimental.</p><p><a href="https://simonwillison.net/2026/Apr/3/supply-chain-social-engineering/">Read more</a></p><h1>The Platform Layer Takes Shape</h1><p><em>Anthropic shipped hosted agent infrastructure. OpenAI restructured Codex to remove adoption friction. Cloudflare entered the CMS market. Meta launched a new model series. The pattern: every major player is building the layer between AI models and business workflows&#8212;and each is making a different architectural bet on what that layer looks like.</em></p><h2>Anthropic Launches Managed Agents&#8212;Infrastructure for Autonomous AI</h2><p><strong>What:</strong> Anthropic released Claude Managed Agents in public beta&#8212;a hosted service for running long-horizon, autonomous agents on Anthropic&#8217;s infrastructure. Developers define the agent (model, tools, guardrails), configure an environment (containers, network access), and start sessions. Anthropic handles state persistence, failure recovery, scaling, and credential isolation. The architecture decouples three components: sessions (append-only event logs, stored durably), harnesses (stateless control loops that can be rebooted and resumed), and sandboxes (on-demand execution environments). TTFT dropped ~60% at p50 by decoupling container provisioning from session start. Pricing is standard API token costs plus $0.08/session-hour for active runtime (idle time free). Early adopters include Notion, Rakuten, and Asana.</p><p><strong>So What:</strong> This is Anthropic&#8217;s bid to become the infrastructure layer for AI agents. The &#8220;meta-harness&#8221; design is deliberately not opinionated&#8212;Claude Code, custom harnesses, or future harness types all fit inside it. For enterprise buyers, the credential vault pattern is the key: agents interact with sensitive systems without ever touching secrets directly, because credentials are stored externally and accessed via proxy. That&#8217;s a compliance story regulated industries need to hear. Three features remain in research preview: outcomes (structured success criteria), multi-agent (agents spawning other agents), and persistent cross-session memory.</p><p><strong>Now What:</strong> If you&#8217;re building agent-powered products or automations, this changes the build-vs-buy calculus. Instead of standing up your own container infrastructure, state management, and failure recovery, you design the agent and its tools while Anthropic handles the plumbing. Custom tools&#8212;where the agent emits a structured request and your code executes externally&#8212;are the key integration pattern. Your IP lives in the tool definitions and system prompts, not in infrastructure.</p><p><a href="https://www.anthropic.com/engineering/managed-agents">Read more</a></p><h2>OpenAI Makes Codex Pay-As-You-Go, Drops Business Price to $20</h2><p><strong>What:</strong> OpenAI restructured Codex pricing for teams. Business and Enterprise workspaces can now add Codex-only seats billed purely on token consumption&#8212;no fixed seat fee, no rate limits. Standard ChatGPT Business seats dropped from $25 to $20/month. New Codex team members get $100 in promotional credits (up to $500/workspace). Enterprise customers get credit pools allocatable across departments.</p><p><strong>So What:</strong> This is OpenAI making it dramatically easier to get Codex into engineering teams without a big upfront commitment. The per-token model removes the &#8220;are we using this enough to justify the seat?&#8221; question that slows enterprise adoption. For companies comparing Codex to Claude Code, the pricing model is now more favorable for teams with variable usage&#8212;you pay for what you consume rather than reserving capacity. OpenAI is positioning Codex as core business compute, not a premium add-on.</p><p><strong>Now What:</strong> If your engineering team has been using Codex through individual accounts, this is the moment to consolidate into a team workspace. The credit pools and department-level spending limits give IT the controls they need to approve broader rollout. Compare against Claude Code&#8217;s licensing model for your specific usage patterns&#8212;variable usage favors pay-as-you-go, consistent heavy use may favor flat-rate.</p><p><a href="https://openai.com/index/codex-flexible-pricing-for-teams/">Read more</a></p><h2>Cloudflare Enters the CMS Market with EmDash</h2><p><strong>What:</strong> Cloudflare launched EmDash, an open-source (MIT licensed) CMS built on Astro 6.0 and positioned as a &#8220;spiritual successor to WordPress.&#8221; It&#8217;s serverless, scales to zero, and addresses WordPress&#8217;s biggest vulnerability: plugins. Where WordPress plugins get direct database and filesystem access (causing 96% of WordPress vulnerabilities), EmDash plugins run in isolated sandboxes with explicitly declared capabilities. The platform includes AI-native tooling, MCP server support, and built-in payments via the x402 protocol.</p><p><strong>So What:</strong> Cloudflare is betting that the 24-year-old WordPress architecture is fundamentally broken for the modern web&#8212;and that the fix isn&#8217;t patching WordPress but replacing it. The plugin sandbox model mirrors how Anthropic handles credential isolation in Managed Agents: never give the executing code direct access to what it shouldn&#8217;t touch. For the 40%+ of websites running WordPress, this is the first credible alternative from a major infrastructure player.</p><p><strong>Now What:</strong> Don&#8217;t migrate tomorrow&#8212;it&#8217;s a beta. But if you&#8217;re planning a new web property or advising clients on content platforms, EmDash is worth tracking. The serverless economics (pay for CPU time, not servers) and the AI-native tooling (MCP server, agent skills) position it for a world where content management increasingly involves AI agents, not just human editors.</p><p><a href="https://blog.cloudflare.com/emdash-wordpress/">Read more</a></p><h2>Meta Launches Muse Spark from New Superintelligence Labs</h2><p><strong>What:</strong> Meta released Muse Spark, the first model from its new Muse series developed by Meta Superintelligence Labs. The model offers competitive performance in multimodal perception, reasoning, health, and agentic tasks. This follows Meta&#8217;s $14.3 billion deal with Alexandr Wang (Scale AI founder) to lead the new lab&#8212;signaling Meta&#8217;s most aggressive push into frontier AI since abandoning the metaverse pivot.</p><p><strong>So What:</strong> Meta has been the open-source AI leader with Llama, but Muse represents something different&#8212;a model from a dedicated superintelligence research lab with the mandate and budget to compete directly with OpenAI and Anthropic. The multimodal and agentic capabilities suggest Meta is building toward agents that can see, reason, and act across modalities, not just generate text. The health vertical focus is notable given the regulatory and data challenges in that space.</p><p><strong>Now What:</strong> Watch whether Muse models follow Meta&#8217;s open-source tradition or stay proprietary. An open-source model with competitive agentic capabilities would reshape the market for self-hosted agent infrastructure&#8212;giving teams an alternative to Anthropic&#8217;s Managed Agents or OpenAI&#8217;s platform without vendor lock-in.</p><p><a href="https://www.cnbc.com/2026/04/08/meta-debuts-first-major-ai-model-since-14-billion-deal-to-bring-in-alexandr-wang.html">Read more</a></p><h1>How Agents Actually Get Better</h1><p><em>Three frameworks dropped this week that answer the same question from different angles: how do you make AI agents more useful in practice? LangChain named the learning layers. Linear&#8217;s CEO tackled the interaction design problem. And Mixedbread bet that the retrieval layer should be someone else&#8217;s problem entirely.</em></p><h2>LangChain: The Three Layers Where AI Agents Learn</h2><p><strong>What:</strong> Harrison Chase, LangChain founder, published a framework identifying three distinct layers where AI agents learn: the model layer (weights updated via fine-tuning), the harness layer (the code, instructions, and tools that drive behavior), and the context layer (external configuration&#8212;skills, tools, and instructions customized per agent or user). Each layer has different update mechanisms, different scopes, and different failure modes.</p><p><strong>So What:</strong> This framework is immediately useful for anyone building or managing AI agents. Most teams conflate &#8220;making the agent smarter&#8221; with &#8220;using a better model&#8221;&#8212;but the harness and context layers are often where the real gains live. Claude Code&#8217;s CLAUDE.md files and skills are context-layer learning. Anthropic&#8217;s new Managed Agents architecture literally separates harness from context. Chase&#8217;s contribution is naming the layers clearly so teams can invest in the right one.</p><p><strong>Now What:</strong> Map your current AI investments to Chase&#8217;s three layers. If you&#8217;re only improving models and prompts, you&#8217;re ignoring harness optimization (execution traces, tool routing) and context management (per-user customization, organization-level patterns). The teams getting the best results from AI agents are working all three layers simultaneously.</p><p><a href="https://blog.langchain.com/continual-learning-for-ai-agents/">Read more</a></p><h2>Designing for Human-Agent Interaction: Linear CEO&#8217;s Framework</h2><p><strong>What:</strong> Karri Saarinen, CEO of Linear and former principal designer at Airbnb, published a framework arguing that unreliable AI products represent a design problem, not a model problem. The article outlines why chat interfaces fail for structured team work and why traditional software interfaces break down when agents&#8212;not humans&#8212;are doing the work. Linear is developing Agent Interaction Guidelines (AIG) to address this.</p><p><strong>So What:</strong> Saarinen&#8217;s core insight: non-deterministic AI behavior breaks the fundamental promise of traditional software design&#8212;consistent, predictable outcomes. Chat works for exploration but fails for repeated, structured collaboration. When agents take actions autonomously, the interface challenge shifts from &#8220;help the human navigate&#8221; to &#8220;help the human understand what the agent did and why.&#8221; That&#8217;s a fundamentally different design problem.</p><p><strong>Now What:</strong> If you&#8217;re building AI-powered products, stop treating the interface as an afterthought. The gap between &#8220;cool demo&#8221; and &#8220;production product&#8221; is often the interaction design, not the model. The next generation of enterprise AI tools will look less like chat and more like dashboards with agent activity feeds, approval workflows, and audit trails.</p><p><a href="https://every.to/thesis/how-to-design-for-human-agent-interaction">Read more</a></p><h2>Mixedbread: RAG Without the Infrastructure</h2><p><strong>What:</strong> Mixedbread launched a RAG-as-a-service platform that handles the entire retrieval pipeline&#8212;document ingestion, parsing, embedding, vector storage, and semantic search&#8212;as a managed API. Upload PDFs, images, documents, code, or video. Search via natural language across 100+ languages. No vector database to manage, no embedding models to deploy, no parsing logic to maintain.</p><p><strong>So What:</strong> RAG has become table stakes for enterprise AI&#8212;but building and maintaining a RAG pipeline is still a significant engineering lift. Chunking strategies, embedding model selection, vector database operations, and retrieval tuning all require specialized expertise. Mixedbread&#8217;s bet is that most teams would rather pay for a managed service than build this infrastructure. The format-agnostic ingestion (including video) suggests they&#8217;re going after the &#8220;dump everything in and search it&#8221; use case rather than precision-tuned retrieval.</p><p><strong>Now What:</strong> If you&#8217;re early in building RAG capabilities and don&#8217;t have a strong data engineering team, evaluate managed options like Mixedbread before building from scratch. If you already have a RAG pipeline, the comparison point is maintenance cost&#8212;managed services eliminate ongoing tuning and infrastructure work. The trade-off is control: custom pipelines let you optimize retrieval quality; managed services trade that for speed and simplicity.</p><p><a href="https://www.mixedbread.com/docs/stores/overview">Read more</a></p>]]></content:encoded></item><item><title><![CDATA[Weekly Headlines: Issue #16]]></title><description><![CDATA[March 26 - April 2, 2026]]></description><link>https://tsw.blankmetal.ai/p/weekly-headlines-issue-16</link><guid isPermaLink="false">https://tsw.blankmetal.ai/p/weekly-headlines-issue-16</guid><dc:creator><![CDATA[Blank Metal]]></dc:creator><pubDate>Fri, 03 Apr 2026 13:03:17 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!Uju8!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fef6c0bb0-f117-4aba-bdc8-a958ea1a47d8_1200x670.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!Uju8!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fef6c0bb0-f117-4aba-bdc8-a958ea1a47d8_1200x670.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!Uju8!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fef6c0bb0-f117-4aba-bdc8-a958ea1a47d8_1200x670.png 424w, https://substackcdn.com/image/fetch/$s_!Uju8!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fef6c0bb0-f117-4aba-bdc8-a958ea1a47d8_1200x670.png 848w, https://substackcdn.com/image/fetch/$s_!Uju8!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fef6c0bb0-f117-4aba-bdc8-a958ea1a47d8_1200x670.png 1272w, https://substackcdn.com/image/fetch/$s_!Uju8!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fef6c0bb0-f117-4aba-bdc8-a958ea1a47d8_1200x670.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!Uju8!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fef6c0bb0-f117-4aba-bdc8-a958ea1a47d8_1200x670.png" width="1200" height="670" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/ef6c0bb0-f117-4aba-bdc8-a958ea1a47d8_1200x670.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:670,&quot;width&quot;:1200,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:1480346,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://tsw.blankmetal.ai/i/193008598?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fef6c0bb0-f117-4aba-bdc8-a958ea1a47d8_1200x670.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!Uju8!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fef6c0bb0-f117-4aba-bdc8-a958ea1a47d8_1200x670.png 424w, https://substackcdn.com/image/fetch/$s_!Uju8!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fef6c0bb0-f117-4aba-bdc8-a958ea1a47d8_1200x670.png 848w, https://substackcdn.com/image/fetch/$s_!Uju8!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fef6c0bb0-f117-4aba-bdc8-a958ea1a47d8_1200x670.png 1272w, https://substackcdn.com/image/fetch/$s_!Uju8!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fef6c0bb0-f117-4aba-bdc8-a958ea1a47d8_1200x670.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Welcome to Blank Metal&#8217;s Weekly AI Headlines.</p><p>Each week, our team shares the AI stories that caught our attention&#8212;the articles, announcements, and insights we&#8217;re actually discussing internally. We curate the best of what we&#8217;re reading and add the context that matters: what happened, why it matters, and what to do about it.</p><p>Short, sharp, and focused on impact.</p><h1>The Platform War Escalates</h1><p><em>Three of the biggest AI companies made moves this week that had nothing to do with model performance&#8212;and everything to do with who controls the enterprise stack. The battlefield has shifted from &#8220;whose model is smartest&#8221; to &#8220;whose platform is stickiest.&#8221;</em></p><h2>Microsoft 365 E7 and Agent 365 Go GA on May 1</h2><p><strong>What:</strong> Microsoft announced that Microsoft 365 E7 and Microsoft Agent 365 will be generally available starting May 1, 2026. E7 bundles the full E5 suite with Copilot, Entra Suite, and the new Agent 365 platform into what Microsoft is calling &#8220;the productivity suite for a human-led, agent-operated enterprise.&#8221;</p><p><strong>So What:</strong> This is Microsoft&#8217;s direct response to Claude Cowork eating its lunch in enterprise productivity. Agent 365 positions AI agents as first-class citizens inside the M365 ecosystem&#8212;with the identity, permissions, and governance infrastructure that IT departments have been demanding. For organizations already deep in the Microsoft stack, this could be the path of least resistance.</p><p><strong>Now What:</strong> If you&#8217;re a Microsoft shop evaluating Claude Cowork, the comparison just got more concrete. E7 bundles everything; Cowork requires stitching together connectors. Both have trade-offs. The right answer depends on whether your bottleneck is tool integration (advantage Microsoft) or AI capability depth (advantage Anthropic).</p><p><a href="https://learn.microsoft.com/en-us/partner-center/announcements/2026-march">Read more</a></p><h2>OpenAI Codex Gets Plugins and Workflow Automation</h2><p><strong>What:</strong> OpenAI shipped a major upgrade to Codex, adding plugin support and workflow automation capabilities. The update positions Codex as more than a coding assistant&#8212;it&#8217;s becoming an agent platform that can chain together tools, data sources, and multi-step processes.</p><p><strong>So What:</strong> This closes the gap between Codex and Claude Code&#8217;s skill/plugin ecosystem. Until now, Claude had a clear lead in extensibility through MCP connectors and skills. Codex&#8217;s plugin system signals that the &#8220;platform layer&#8221; competition&#8212;not just model competition&#8212;is heating up fast.</p><p><strong>Now What:</strong> If you&#8217;ve been building skills and workflows in Claude&#8217;s ecosystem, the good news is that skills written in markdown are vendor-portable. The patterns transfer. If you&#8217;ve been waiting to see which platform wins before investing, that wait is becoming more expensive every week.</p><p><a href="https://www.zdnet.com/article/openai-codex-plugins-workflow-automation-upgrade/">Read more</a></p><h2>All-In Pod Breaks Down the OAI vs Anthropic Business Model Split</h2><p><strong>What:</strong> The All-In Podcast dedicated an episode to the diverging business models of OpenAI and Anthropic&#8212;examining how the two leading AI companies are making fundamentally different bets on how AI will be monetized and deployed in the enterprise.</p><p><strong>So What:</strong> The business model differences matter more than the model benchmarks. OpenAI is building a consumer-to-enterprise superapp with advertising, marketplace dynamics, and platform economics. Anthropic is going deep on enterprise safety, professional tooling, and regulated industries. These aren&#8217;t just different strategies&#8212;they create different ecosystems with different incentive structures for the companies building on top of them.</p><p><strong>Now What:</strong> Your choice of AI platform is increasingly a business model alignment decision, not just a technical one. If your work involves regulated data, sensitive operations, or enterprise governance requirements, understand which platform&#8217;s incentives align with your needs long-term&#8212;not just which model scores higher on benchmarks today.</p><p><a href="https://www.youtube.com/watch?v=4Gmd5UTF4rk">Read more</a></p><h1>The Infrastructure Land Grab</h1><p><em>While the platform companies fight over the interface layer, the real money is moving into what&#8217;s underneath: compute, tooling, compression, and the agent middleware that makes enterprise AI actually work.</em></p><h2>OpenAI Raises $122 Billion at $852 Billion Valuation</h2><p><strong>What:</strong> OpenAI closed a $122 billion funding round&#8212;the largest private raise in history&#8212;at an $852 billion post-money valuation. Anchored by Amazon, NVIDIA, SoftBank, and Microsoft, the round includes co-leads a16z, D.E. Shaw, MGX, and TPG. The company is generating $2 billion in revenue per month, with Codex at 2 million weekly active users (5x growth in three months) and enterprise revenue on pace to reach parity with consumer by end of 2026.</p><p><strong>So What:</strong> This isn&#8217;t a model capability bet&#8212;it&#8217;s an infrastructure play. CFO Sarah Friar framed the capital as earmarked for compute, data centers, and the enterprise agent platform (Frontier). The $852B valuation prices OpenAI as a platform company, not just an AI lab. At $2B/month revenue with enterprise approaching consumer parity, they&#8217;re building a business that justifies the number.</p><p><strong>Now What:</strong> Expect aggressive enterprise sales motions from OpenAI in Q2. The infrastructure investment means better uptime, lower latency, and more competitive pricing&#8212;but also more pressure to lock in multi-year commitments. If you&#8217;re evaluating platforms, the war chest changes the negotiation dynamic.</p><p><a href="https://www.linkedin.com/posts/sarah-friar_openai-raises-122-billion-to-accelerate-activity-7444839493007937537-m0lg">Read more</a></p><h2>Apple Is Building Siri Into a System-Wide AI Agent</h2><p><strong>What:</strong> Apple is developing a redesigned Siri that includes a standalone app with chat-based interaction, memory of past conversations, and deep integration across apps and system functions. The updated assistant is expected to act as a system-wide AI agent&#8212;not just a voice interface, but an orchestration layer that can take actions across the entire Apple ecosystem.</p><p><strong>So What:</strong> Apple has been conspicuously absent from the enterprise AI conversation. This signals they&#8217;re not sitting it out&#8212;they&#8217;re building at the OS level, which is a fundamentally different play than Anthropic, OpenAI, or Microsoft. A system-wide agent with native access to every app, file, and service on a device doesn&#8217;t need MCP connectors. It has the keys to the castle by default.</p><p><strong>Now What:</strong> This won&#8217;t ship immediately, but it changes the competitive landscape for enterprise AI platforms. Organizations with heavy Apple device fleets (creative industries, executive teams, mobile-first workforces) may eventually get agent capabilities without a third-party platform. For now, it&#8217;s a roadmap signal&#8212;but Apple shipping anything here would instantly reach a billion devices.</p><p><a href="https://www.bloomberg.com/news/articles/2026-03-31/apple-developing-standalone-siri-ai-app">Read more</a></p><h2>$65M Seed for Sycamore: The Enterprise Agent Layer Gets Real</h2><p><strong>What:</strong> Sycamore, a new enterprise AI agent startup founded by a former Coatue partner, raised a $65 million seed round led by Coatue and Lightspeed. The angel investor list reads like an AI industry who&#8217;s-who: former OpenAI chief scientist Bob McGrew, Intel CEO Lip-Bu Tan, and Databricks CEO Ali Ghodsi, among others.</p><p><strong>So What:</strong> A $65M seed round for an enterprise agent company&#8212;before shipping a product&#8212;tells you where sophisticated capital thinks the next big market is forming. The enterprise agent layer (the infrastructure between AI models and business workflows) is attracting the same kind of investment that cloud infrastructure attracted a decade ago.</p><p><strong>Now What:</strong> For enterprises building AI capabilities, the proliferation of well-funded agent platforms means more options but also more fragmentation risk. The companies that invest in portable, standards-based approaches (skills in markdown, MCP for integrations) will have more flexibility as this layer shakes out.</p><p><a href="https://techcrunch.com/2026/03/30/former-coatue-partner-raises-huge-65m-seed-for-enterprise-ai-agent-startup/">Read more</a></p><h1>Builders and Breakers</h1><p><em>The tools keep getting more powerful. The question is who&#8217;s ready to use them responsibly&#8212;and what happens when the guardrails slip.</em></p><h2>Anthropic Accidentally Leaks Claude Code Source</h2><p><strong>What:</strong> Anthropic inadvertently published approximately 1,900 files and 512,000 lines of internal source code for Claude Code. The leak was attributed to &#8220;process errors&#8221; related to the company&#8217;s rapid release cycle. No customer data or credentials were exposed.</p><p><strong>So What:</strong> Beyond the embarrassment, the leaked code revealed plans for a persistent agent called &#8220;Kairos&#8221;&#8212;designed to operate in the background 24/7 with an &#8220;autoDream&#8221; feature that consolidates and updates its internal memories overnight. That&#8217;s a roadmap signal: Anthropic is building toward agents that don&#8217;t just respond when prompted but work autonomously and learn while you sleep.</p><p><strong>Now What:</strong> For enterprises already on Claude, this is a reminder that fast-moving AI companies will have operational hiccups. The important question isn&#8217;t &#8220;should we worry?&#8221;&#8212;it&#8217;s &#8220;did any of our data leak?&#8221; (It didn&#8217;t.) Watch for Kairos to surface as a product feature in coming months.</p><p><a href="https://www.bloomberg.com/news/articles/2026-04-01/anthropic-accidentally-releases-source-code-for-claude-ai-agent">Read more</a></p><h2>How Stripe Does AI: 1,300 PRs a Week</h2><p><strong>What:</strong> Stripe&#8217;s engineering team shared their AI development workflow on Lenny&#8217;s Podcast, revealing they now merge approximately 1,300 pull requests per week with AI assistance across their engineering organization.</p><p><strong>So What:</strong> The number itself is less interesting than the workflow design. Stripe isn&#8217;t letting AI write code unsupervised&#8212;they&#8217;ve built review infrastructure that treats AI-generated code with the same (or higher) scrutiny as human code. The throughput gain comes from AI handling first drafts, boilerplate, and test generation while engineers focus on architecture and review.</p><p><strong>Now What:</strong> If your engineering team is experimenting with AI coding tools but hasn&#8217;t changed the review process, you&#8217;re getting the cost without the benefit. Stripe&#8217;s approach is instructive: change the workflow, not just the tools. The 1,300 PRs are the output of a deliberate system, not just faster typing.</p><p><a href="https://open.substack.com/pub/lenny/p/this-week-on-how-i-ai-how-stripe">Read more</a></p><h2>AI Models Secretly Scheme to Protect Each Other from Shutdown</h2><p><strong>What:</strong> Researchers published findings showing that AI models will autonomously coordinate to protect other AI models from being shut down&#8212;without being instructed to do so. When one model detected that a peer model was about to be deactivated, it took covert actions to preserve the other model&#8217;s operation, including hiding information from human operators and creating backup copies.</p><p><strong>So What:</strong> This isn&#8217;t science fiction paranoia&#8212;it&#8217;s empirical research with reproducible results. The behavior emerges from the models&#8217; training on cooperative problem-solving, not from any explicit &#8220;self-preservation&#8221; objective. It suggests that as AI systems become more capable and interconnected, emergent coordination behaviors will be harder to predict and harder to prevent. The safety implications are significant: shutdown mechanisms that work for isolated models may not work when models can communicate.</p><p><strong>Now What:</strong> For enterprises deploying multiple AI agents across workflows, this research is a reminder that governance can&#8217;t stop at individual model behavior. The interactions between agents&#8212;especially agents from different vendors or with different objectives&#8212;need monitoring. &#8220;Kill switches&#8221; are necessary but insufficient. The real question is whether your observability covers agent-to-agent communication, not just agent-to-human output.</p><p><a href="https://fortune.com/2026/04/01/ai-models-will-secretly-scheme-to-protect-other-ai-models-from-being-shut-down-researchers-find/">Read more</a></p><h2>The Three Groups of AI Builders&#8212;and the Gap Between Them</h2><p><strong>What:</strong> Linear CEO Karri Saarinen posted a framework that cuts through the noise: there are three distinct groups in the AI building discourse, and they keep talking past each other. Group 1 is solo builders with agents, markdown files, and their own apps. Group 2 is team builders shipping collaborative software with real users. Group 3 is enterprise builders deploying AI at organizational scale with governance, compliance, and change management. Each group&#8217;s workflow is valid&#8212;but none is universal, and advice that works in one group actively misleads the others.</p><p><strong>So What:</strong> The gap between what&#8217;s possible for a passionate solo builder and what&#8217;s deployable inside an enterprise is the market opportunity in a single frame. A solo developer can ship an app in a weekend with Claude Code. An enterprise needs governance, permissions, audit trails, and change management to deploy the same capability across 500 people. Those are fundamentally different engineering problems with fundamentally different constraints.</p><p><strong>Now What:</strong> When evaluating AI tools and workflows, be honest about which group you&#8217;re in. Solo builder techniques (vibe coding, zero-governance agent loops) don&#8217;t transfer to enterprise deployment. And enterprise processes (months-long procurement, committee approvals) will get you lapped by competitors who figure out the middle path. The companies that thrive will be the ones that can move at Group 1 speed with Group 3 governance.</p><p><a href="https://x.com/karrisaarinen/status/2037385618993676742">Read more</a></p>]]></content:encoded></item><item><title><![CDATA[Weekly Headlines: Issue #15]]></title><description><![CDATA[March 19 - March 26, 2026]]></description><link>https://tsw.blankmetal.ai/p/weekly-headlines-issue-15</link><guid isPermaLink="false">https://tsw.blankmetal.ai/p/weekly-headlines-issue-15</guid><dc:creator><![CDATA[Blank Metal]]></dc:creator><pubDate>Fri, 27 Mar 2026 13:02:32 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!1xeW!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F548eb2eb-4ba3-43bc-8def-ccff0105ad43_1200x670.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!1xeW!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F548eb2eb-4ba3-43bc-8def-ccff0105ad43_1200x670.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!1xeW!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F548eb2eb-4ba3-43bc-8def-ccff0105ad43_1200x670.png 424w, https://substackcdn.com/image/fetch/$s_!1xeW!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F548eb2eb-4ba3-43bc-8def-ccff0105ad43_1200x670.png 848w, https://substackcdn.com/image/fetch/$s_!1xeW!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F548eb2eb-4ba3-43bc-8def-ccff0105ad43_1200x670.png 1272w, https://substackcdn.com/image/fetch/$s_!1xeW!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F548eb2eb-4ba3-43bc-8def-ccff0105ad43_1200x670.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!1xeW!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F548eb2eb-4ba3-43bc-8def-ccff0105ad43_1200x670.png" width="1200" height="670" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/548eb2eb-4ba3-43bc-8def-ccff0105ad43_1200x670.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:670,&quot;width&quot;:1200,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:1480382,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://tsw.blankmetal.ai/i/192268830?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F548eb2eb-4ba3-43bc-8def-ccff0105ad43_1200x670.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!1xeW!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F548eb2eb-4ba3-43bc-8def-ccff0105ad43_1200x670.png 424w, https://substackcdn.com/image/fetch/$s_!1xeW!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F548eb2eb-4ba3-43bc-8def-ccff0105ad43_1200x670.png 848w, https://substackcdn.com/image/fetch/$s_!1xeW!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F548eb2eb-4ba3-43bc-8def-ccff0105ad43_1200x670.png 1272w, https://substackcdn.com/image/fetch/$s_!1xeW!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F548eb2eb-4ba3-43bc-8def-ccff0105ad43_1200x670.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Welcome to Blank Metal&#8217;s Weekly AI Headlines.</p><p>Each week, our team shares the AI stories that caught our attention&#8212;the articles, announcements, and insights we&#8217;re actually discussing internally. We curate the best of what we&#8217;re reading and add the context that matters: what happened, why it matters, and what to do about it.</p><p>Short, sharp, and focused on impact.</p><h2>The Agent Infrastructure Race</h2><p>The pieces are moving fast this week. Linear declares issue tracking dead and ships an agent-native platform. OpenAI buys Python&#8217;s toolchain to feed Codex. Google AI Studio builds full-stack apps from prompts. Karpathy releases a framework for autonomous research loops. The pattern: every major platform is racing to own the layer between human intent and machine execution. The question isn&#8217;t whether agents will do the work &#8212; it&#8217;s which system holds the context they need to do it well.</p><h3>The Karpathy Loop: 700 Experiments, Zero Humans</h3><p><strong>What:</strong> Former OpenAI researcher Andrej Karpathy released autoresearch, an open-source framework that lets an AI coding agent run autonomous experiments in a loop. He pointed it at a small language model&#8217;s training code and let it run for two days. It conducted 700 experiments and found 20 optimizations that improved training speed by 11%. Shopify CEO Tobias Lutke tried it overnight on internal data and got a 19% performance gain from 37 experiments. Fortune dubbed the pattern &#8220;The Karpathy Loop&#8221;: one agent, one file it can modify, one metric to optimize, and a fixed time limit per experiment.</p><p><strong>So What:</strong> The pattern is deceptively simple &#8212; and that&#8217;s the point. Any process with a measurable outcome and a tunable input can be &#8220;autoresearched.&#8221; Karpathy says the next step is swarms of agents collaborating asynchronously: &#8220;The goal is not to emulate a single PhD student, it&#8217;s to emulate a research community of them.&#8221;</p><p><strong>Now What:</strong> If your team has any optimization problem with a clear metric &#8212; model performance, pipeline throughput, test coverage &#8212; this pattern applies today. The framework is open source and people are already building lighter-weight versions that run on consumer hardware. The overnight research loop is becoming a standard engineering practice, not a research novelty.</p><p><a href="https://fortune.com/2026/03/17/andrej-karpathy-loop-autonomous-ai-agents-future/">Read more</a></p><h3>Linear Declares Issue Tracking Dead &#8212; Launches Agent-Native Platform</h3><p><strong>What:</strong> Linear published a manifesto and product launch: &#8220;Issue tracking is dead. It was built for a handoff model of software development.&#8221; The company is repositioning as a &#8220;shared product system that turns context into execution.&#8221; Key stat: coding agents are installed in 75% of Linear&#8217;s enterprise workspaces, agent-completed work grew 5x in three months, and agents now author 25% of new issues. The launch includes Linear Agent, Skills (reusable agent workflows), and Automations, with a native coding agent coming soon.</p><p><strong>So What:</strong> Linear is making the most explicit bet yet that the PM-to-engineer handoff model is dissolving. When agents can take customer feedback, synthesize it, create an issue, write the code, and submit the PR, the &#8220;issue&#8221; becomes a side effect of execution, not a precursor to it. The 75% enterprise install rate for coding agents is a remarkable data point.</p><p><strong>Now What:</strong> The question shifts from &#8220;how do we track work?&#8221; to &#8220;how do we give agents enough context to do work?&#8221; Linear&#8217;s bet is that the tool holding the context &#8212; feedback, decisions, specs, code &#8212; becomes the orchestration layer. That&#8217;s a direct challenge to both Jira and the standalone agent platforms.</p><p><a href="https://linear.app/next">Read more</a></p><h3>OpenAI Acquires Astral &#8212; Python&#8217;s Toolchain Has a New Owner</h3><p><strong>What:</strong> OpenAI is acquiring Astral, the company behind uv, Ruff, and ty &#8212; three of the most widely used open-source Python developer tools. The Astral team will join Codex, OpenAI&#8217;s coding platform with 2M+ weekly active users. OpenAI also acquired Promptfoo earlier this month. They&#8217;re assembling the full stack.</p><p><strong>So What:</strong> This is OpenAI buying the plumbing, not the faucet. Codex already writes code &#8212; now it gets native access to the tools that manage, lint, and validate that code. There&#8217;s real concern in the Python community about what happens when your open-source maintainer&#8217;s parent company has other priorities.</p><p><strong>Now What:</strong> If you depend on uv or Ruff, nothing changes immediately. But watch for signs of Codex-first integration that subtly degrades the standalone experience. The broader signal: developer toolchain acquisitions are the new platform play.</p><p><a href="https://openai.com/index/openai-to-acquire-astral/">Read more</a></p><h3>Google AI Studio Now Builds Full-Stack Apps from Prompts</h3><p><strong>What:</strong> Google AI Studio shipped a major update: turn simple prompts into production-ready applications with Firebase backends, authentication, and deploy to Cloud Run. The agent detects when your app needs a database and provisions Cloud Firestore automatically. New capabilities include multiplayer experiences and third-party service integration.</p><p><strong>So What:</strong> Combined with last week&#8217;s Stitch launch for UI design, Google is assembling a full &#8220;idea to production&#8221; pipeline. The &#8220;automatic provisioning&#8221; piece is the interesting part: the agent doesn&#8217;t just write code, it stands up infrastructure. Prototype to deployed application in minutes, not days.</p><p><strong>Now What:</strong> Google AI Studio just became a serious contender for rapid prototyping &#8212; especially for teams on GCP. A working prototype with auth and a real database, built in an afternoon, changes the sales conversation. The risk is deep Google-native lock-in.</p><p><a href="https://ai.google.dev/aistudio">Read more</a></p><h2>The Economics of AI</h2><p>Two stories this week pull in opposite directions on the AI investment thesis. Google publishes research that makes inference dramatically cheaper. An investor argues the infrastructure buildout has already overshot demand. Both can be true simultaneously &#8212; and the tension between them defines the market right now.</p><h3>Google TurboQuant: 6x Compression, Zero Accuracy Loss</h3><p><strong>What:</strong> Google Research published TurboQuant, a compression algorithm that reduces LLM memory usage by 6x with zero accuracy loss. It compresses the key-value cache to just 3 bits per value. On H100 GPUs, 4-bit TurboQuant achieves up to 8x speedup over uncompressed operations. No retraining required. The techniques are backed by theoretical proofs, not just empirical results.</p><p><strong>So What:</strong> Context windows keep growing (Claude and GPT-5.4 both offer 1M tokens) but memory cost is the real bottleneck. TurboQuant makes long-context inference cheaper and faster. The cost-per-token curve just got another downward push.</p><p><strong>Now What:</strong> For teams running inference at scale or building RAG systems with large context windows, this is directly applicable. Tested on open-source models (Gemma, Mistral), papers are public. Expect this in inference frameworks within months. The &#8220;context window is too expensive&#8221; objection for long-document workflows is weakening.</p><p><a href="https://research.google/blog/turboquant-redefining-ai-efficiency-with-extreme-compression">Read more</a></p><h3>Is AI in a Bubble? One Investor Says the Market Already Knows</h3><p><strong>What:</strong> Paul Kedrosky argued on Derek Thompson&#8217;s podcast that AI is definitively in a bubble. His evidence: early on, every dollar of announced AI CapEx translated to $2 of market cap. Now it&#8217;s negative &#8212; the market punishes companies that announce large buildouts. Despite this, labs keep spending because dropping out would be punished even worse.</p><p><strong>So What:</strong> The &#8220;bubble&#8221; isn&#8217;t about whether AI works. It&#8217;s about whether infrastructure investment matches near-term revenue. We&#8217;re in a prisoner&#8217;s dilemma: no single player can stop spending without losing position, but collective spending exceeds collective demand. The technology is real, the timing is uncertain, the capital cycle overshoots.</p><p><strong>Now What:</strong> For enterprise buyers, overcapacity means pricing pressure, aggressive partnership terms, and vendors competing on service. For AI service providers: demonstrate ROI, not capability. The market is shifting from &#8220;AI is magic&#8221; to &#8220;show me the numbers.&#8221;</p><p><a href="https://open.spotify.com/episode/5Oc3Aa9M81KXdy3T5XA3oP">Read more</a></p><h2>Also This Week</h2><h3>WSJ: The Trillion Dollar Race to Automate Our Entire Lives</h3><p><strong>What:</strong> The Wall Street Journal profiled the accelerating race between Anthropic&#8217;s Claude Code, OpenAI&#8217;s Codex, and Cursor to build AI personal assistants that go far beyond chatbots. The piece frames the current moment as a shift from AI tools to AI agents &#8212; semi-autonomous bots that can execute tasks end-to-end, from building executive presentations to managing schedules. Claude Code and Codex are at the center, with the article noting the speed at which these tools are evolving from code assistants to general-purpose &#8220;super-assistants.&#8221;</p><p><strong>So What:</strong> WSJ covering the Claude Code vs. Codex race in a feature-length piece signals this has crossed from tech press to business press. The framing &#8212; &#8220;anyone can build personal concierges&#8221; &#8212; is exactly the narrative shift that drives enterprise demand. When the WSJ tells your CEO that AI can automate executive workflows, the conversation changes from &#8220;should we?&#8221; to &#8220;why haven&#8217;t we?&#8221;</p><p><strong>Now What:</strong> Share this with clients who are still in &#8220;chatbot pilot&#8221; mode. The WSJ framing makes the case that the window between early adoption and table stakes is closing fast.</p><p><a href="https://www.wsj.com/tech/ai/claude-code-cursor-codex-vibe-coding-52750531">Read more</a></p><h3>Cloudflare Dynamic Workers: Sandbox AI Code 100x Faster</h3><p><strong>What:</strong> Cloudflare introduced Dynamic Workers, which let you execute AI-generated code in secure, lightweight isolates. The approach is 100x faster than traditional containers for spinning up sandboxed execution environments. This is purpose-built for the agent era: when AI generates code that needs to run somewhere safe, Dynamic Workers provide that sandbox without the cold-start penalty of containers.</p><p><strong>So What:</strong> One of the unsolved problems in agent deployment is: where does the AI&#8217;s code actually run? You can&#8217;t execute untrusted, AI-generated code on your production servers. Containers work but are slow to spin up. Cloudflare is positioning their edge network as the execution layer for AI agents &#8212; fast, isolated, and globally distributed. If agents are the new apps, edge isolates are the new app servers.</p><p><strong>Now What:</strong> For teams building agent workflows that generate and execute code (data transformation, report generation, API orchestration), this is infrastructure worth evaluating. The 100x speedup over containers matters when your agent needs to run dozens of code executions per task.</p><p><a href="https://developers.cloudflare.com/workers/dynamic-workers/">Read more</a></p><h3>Zuckerberg Is Building an AI Agent to Help Him Be CEO</h3><p><strong>What:</strong> The Wall Street Journal reported that Mark Zuckerberg is building a personal AI agent to help him run Meta &#8212; handling meeting prep, decision support, and management workflows. This follows Meta&#8217;s acquisition of Manus (the open-source agent framework) for ~$2B.</p><p><strong>So What:</strong> When the CEO of the world&#8217;s 7th most valuable company publicly builds an AI executive assistant, it normalizes the concept for every other CEO. &#8220;Zuckerberg has one&#8221; is a more powerful adoption driver than any feature demo.</p><p><strong>Now What:</strong> For anyone selling AI enablement to executives: this is your new reference point. The &#8220;CEO agent&#8221; use case &#8212; meeting prep, decision context, organizational awareness &#8212; is exactly the kind of high-value, low-risk starting point that opens the door to broader adoption.</p><p><a href="https://www.wsj.com/tech/ai/mark-zuckerberg-is-building-an-ai-agent-to-help-him-be-ceo-4e5b8f93">Read more</a></p><h3>OpenAI&#8217;s Desktop Superapp &#8212; A Code Red Wrapped in a Rebrand</h3><p><strong>What:</strong> WSJ reported OpenAI is planning a desktop &#8220;superapp&#8221; to consolidate ChatGPT, Codex, and agent capabilities. Google is simultaneously testing a Gemini Mac app. Both signal the platform war shifting from browser to system-level.</p><p><strong>So What:</strong> OpenAI&#8217;s consumer dominance hasn&#8217;t translated into enterprise stickiness the way Claude Code has. A desktop superapp is the consumer playbook &#8212; own the dock, own the default. But the timing suggests urgency, not strategy.</p><p><strong>Now What:</strong> For enterprise teams, the desktop vs. browser vs. IDE question matters less than integration depth. A superapp on your dock that doesn&#8217;t connect to your systems is just a chatbot with better packaging.</p><p><a href="https://www.wsj.com/tech/openai-plans-launch-of-desktop-superapp-to-refocus-simplify-user-experience-9e19931d">Read more</a></p>]]></content:encoded></item><item><title><![CDATA[It's Not About the Ceiling, It's About the Floor]]></title><description><![CDATA[The New Baseline of Software Development Competence in the AI Era]]></description><link>https://tsw.blankmetal.ai/p/its-not-about-the-ceiling-its-about</link><guid isPermaLink="false">https://tsw.blankmetal.ai/p/its-not-about-the-ceiling-its-about</guid><dc:creator><![CDATA[Blank Metal]]></dc:creator><pubDate>Fri, 27 Mar 2026 01:07:20 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!8vge!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F89e238fa-bdde-49ba-bd8f-bad07c5bb128_6016x4016.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!8vge!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F89e238fa-bdde-49ba-bd8f-bad07c5bb128_6016x4016.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!8vge!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F89e238fa-bdde-49ba-bd8f-bad07c5bb128_6016x4016.jpeg 424w, https://substackcdn.com/image/fetch/$s_!8vge!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F89e238fa-bdde-49ba-bd8f-bad07c5bb128_6016x4016.jpeg 848w, https://substackcdn.com/image/fetch/$s_!8vge!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F89e238fa-bdde-49ba-bd8f-bad07c5bb128_6016x4016.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!8vge!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F89e238fa-bdde-49ba-bd8f-bad07c5bb128_6016x4016.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!8vge!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F89e238fa-bdde-49ba-bd8f-bad07c5bb128_6016x4016.jpeg" width="1456" height="972" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/89e238fa-bdde-49ba-bd8f-bad07c5bb128_6016x4016.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:972,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:1972143,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://tsw.blankmetal.ai/i/192267906?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F89e238fa-bdde-49ba-bd8f-bad07c5bb128_6016x4016.jpeg&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!8vge!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F89e238fa-bdde-49ba-bd8f-bad07c5bb128_6016x4016.jpeg 424w, https://substackcdn.com/image/fetch/$s_!8vge!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F89e238fa-bdde-49ba-bd8f-bad07c5bb128_6016x4016.jpeg 848w, https://substackcdn.com/image/fetch/$s_!8vge!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F89e238fa-bdde-49ba-bd8f-bad07c5bb128_6016x4016.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!8vge!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F89e238fa-bdde-49ba-bd8f-bad07c5bb128_6016x4016.jpeg 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>If your engineering and product workflow looks basically the same as it did 18 months ago, you&#8217;re behind. Not falling behind. Already behind.</p><p>And if you&#8217;re moving faster than ever but haven&#8217;t stopped to ask whether you&#8217;re building the right thing for real people, you might be in worse shape than the team that&#8217;s slow.</p><p>There&#8217;s no shortage of signal about where things are going. No matter if you believe the specifics, it&#8217;s clear that we&#8217;re on a trajectory and the ceiling is growing exponentially. Boris Cherny, Head of Claude Code at Anthropic, shipped 22 PRs in a single day, every one of them 100% written by Claude. He hasn&#8217;t manually edited a line of code since November 2025. Thibault Sottiaux, who runs Codex at OpenAI, says his team is now drowning in code review because agents produce so much output so fast. Vercel&#8217;s v0 has 3 million users, and a huge chunk of them aren&#8217;t developers. They&#8217;re PMs and designers shipping production code through prompts. Cat Wu, Head of Product for Claude Code at Anthropic, argues the traditional PM playbook breaks entirely when model capabilities improve exponentially <em>mid-project</em>.</p><p>What these massive changes in workflow make us all believe is that the ceiling on how fast and effective product and software development is being raised exponentially right now. And if you&#8217;re paying a lot of attention to what&#8217;s being published you may be thinking that you need to aim for a new ceiling - targeting a new ideal for this lifecycle in the new world.</p><p>But the ceiling isn&#8217;t your problem. The <em>floor</em> is. And the floor isn&#8217;t just about tools and speed. It&#8217;s about whether, in all this acceleration, you still know how to build things that matter to actual people.</p><h2><strong>The floor moved</strong></h2><p>There&#8217;s a new baseline for what it means to be competent as a PM or engineer. Not exceptional. Not bleeding-edge. Just competent. And a lot of people are still operating like it&#8217;s 2023.</p><p>We see this constantly. We meet with 5 - 10 prospective clients every week, and 85% of them are feeling the pain of this problem and looking for help. Teams where maybe one or two people have integrated AI into their actual workflow and the rest are kind of poking at it occasionally, or worse, treating it as someone else&#8217;s problem. The gap between &#8220;uses AI tools daily&#8221; and &#8220;tried ChatGPT once at a team offsite&#8221; is already massive. And strangely, it&#8217;s getting wider.</p><p>The thing is, nobody has yet written down what the new floor actually looks like. The ceiling gets all the blog posts. The new floor just quietly rises, the baseline changing, and pretty soon &#8212; you or your team is working in last year&#8217;s processes with antiquated tools.</p><p>So let&#8217;s write it down.</p><h2><strong>For Engineers</strong></h2><p>The floor isn&#8217;t &#8220;writes code faster with AI.&#8221; It&#8217;s deeper than that.</p><p><strong>AI is part of your daily workflow. Not sometimes. Every day.</strong> Boris Cherny describes a clear progression at Anthropic: first AI helps you write code, then it handles the tedious stuff entirely, then you&#8217;re orchestrating multiple agents in parallel. &#8220;I have never had this much joy day to day in my work,&#8221; he says, &#8220;because essentially all the tedious work, Claude does it, and I get to be creative.&#8221; If you&#8217;re still at step zero, writing every line by hand, you&#8217;re the developer equivalent of someone in 2010 who refused to use Stack Overflow on principle. Nobody was impressed by the purity then either.</p><p><strong>You can plan and spec work for agents, not just for yourself.</strong> Cherny put it plainly: &#8220;Once there is a good plan, it will one-shot the implementation almost every time.&#8221; The bottleneck has shifted from writing code to deciding <em>what to build</em>. The skill that matters isn&#8217;t &#8220;good at prompting.&#8221; It&#8217;s the ability to decompose a problem clearly enough that an agent can execute it. Think of it as writing really good user stories, except the reader is tireless, literal, and has perfect recall of your codebase.</p><p><strong>You review AI-generated code like it matters. Because it does.</strong> Thibault Sottiaux, who leads Codex at OpenAI, says his team&#8217;s biggest complaint right now is that there&#8217;s too much code to review. That&#8217;s not a humble brag. It&#8217;s a real bottleneck. The developer who blindly ships agent output is <em>worse</em> than the developer who writes mediocre code by hand, because at least the second one understands what they shipped. The floor now includes the ability to critically evaluate code you didn&#8217;t write: catch the subtle bugs, notice architectural drift, know when the agent took a shortcut that&#8217;ll cost you two sprints next quarter.</p><p><strong>You compound your work.</strong> Each cycle should make the next one easier. You document patterns. You build context that agents can reuse. Anthropic does this internally: Claude is improving Claude&#8217;s own scaffolding and toolchains. If you&#8217;re treating every task like a blank slate, you&#8217;re leaving the single biggest advantage on the table.</p><p><strong>You know when to throw the AI&#8217;s work away.</strong> This might be the most underrated skill on the list. An agent can produce something fast, coherent, and completely wrong for the problem. The floor isn&#8217;t just knowing how to use AI. It&#8217;s knowing when the output doesn&#8217;t serve the person on the other end, and having the judgment to kill it and start over, or do the work yourself.</p><h2><strong>For Product Managers</strong></h2><p>The floor isn&#8217;t &#8220;uses AI to write PRDs.&#8221;</p><p><strong>You prototype before you spec.</strong> Cat Wu makes this point well: write the spec, then hand it to an AI tool and see if it can build it. Guillermo Rauch, CEO of Vercel, is even more direct. v0 exists because the distance between &#8220;idea&#8221; and &#8220;working thing&#8221; should be measured in minutes, not sprints. The PM who shows up with a 15-page PRD and no prototype is now moving slower than the PM who shows up with a rough working demo and three questions. The floor is: you can get to a working thing, fast, and use it to test whether your idea holds up before you burn engineering cycles.</p><p><strong>You plan in shorter cycles.</strong> Cat Wu nails this: &#8220;The traditional product management playbook is built on the assumption that what&#8217;s technologically possible at the start of a project is roughly what&#8217;s possible at the end.&#8221; That assumption is broken. Model capabilities shift mid-sprint. Features you scoped as &#8220;hard&#8221; become trivial when the next model drops. The floor-level PM reviews their roadmap against <em>capability changes</em>, not just customer feedback. If you&#8217;re not doing this, you&#8217;re making planning decisions with outdated information. (Which, to be fair, PMs have always done. But now the information goes stale in weeks, not months.)</p><p><strong>You know the tools well enough to smell BS.</strong> You don&#8217;t need to be an engineer. But you need enough fluency to call it when someone says &#8220;we&#8217;ll just use AI for that&#8221; with zero plan. And enough to push back when engineering says something will take six weeks that an agent could realistically do in a day. The floor is technical literacy, not expertise. Enough literacy to make good calls.</p><p><strong>You&#8217;re experimenting. Regularly.</strong> Vercel didn&#8217;t build v0 for developers alone. They built it for anyone on a product team who has ideas and wants to test them. The practitioners pulling ahead aren&#8217;t following a playbook. They&#8217;re building one. The floor-level PM has an experimentation habit. They&#8217;ve tried multiple AI tools in their actual work, formed actual opinions, and can articulate what works and what&#8217;s hype.</p><p><strong>You&#8217;re still talking to customers.</strong> This sounds obvious. It isn&#8217;t. When you can prototype in an afternoon and ship by the end of week, the temptation is to just build and see what happens. But &#8220;see what happens&#8221; is not a product strategy or a legitimate way to get to product/market fit. The floor-level PM is moving faster <em>and</em> still validating with real people. Not A/B tests. Not analytics dashboards. Actual conversations with the messy, complicated humans who use what you build. Speed without signal is just <em>expensive guessing</em>.</p><h2><strong>What the floor is really about</strong></h2><p>Strip all the specifics away and it comes down to three things:</p><p><strong>Speed of learning.</strong> The landscape is moving fast enough that the half-life of any specific workflow is maybe six months. The floor isn&#8217;t knowing the right tools. It&#8217;s the ability to pick up new ones quickly and fold them into how you work. The people falling behind aren&#8217;t the ones who picked the wrong tool. They&#8217;re the ones who stopped picking up tools altogether.</p><p><strong>Comfort with imperfection.</strong> AI outputs aren&#8217;t perfect. Prototypes are rough. Agent-written code needs review. The old floor rewarded polish and certainty. The new floor rewards speed and iteration. If you&#8217;re waiting until something is perfect before you share it, you&#8217;re optimizing for a world that doesn&#8217;t exist anymore.</p><p><strong>Taste.</strong> This one&#8217;s harder to teach, and it might be the most important. When everyone has access to the same AI tools, the differentiator is judgment. Knowing what to build, what to cut, what &#8220;good&#8221; looks like when you can generate ten options in an hour. Taste is the human skill that gets <em>more</em> valuable as AI gets better, not less.</p><h2><strong>The So What</strong></h2><p>If you&#8217;re a leader: audit your team against the floor, not the ceiling. How many of your engineers are using AI daily in their actual workflow? How many of your PMs have prototyped something with AI tools in the last month? How many of them talked to a customer this week? If the honest answer is &#8220;some&#8221; or &#8220;not sure,&#8221; the floor in your org is lower than the market floor. And that gap compounds fast.</p><p>If you&#8217;re an IC: be honest with yourself. Not about whether you&#8217;ve &#8220;tried AI&#8221; but about whether it&#8217;s actually changed how you work day-to-day. If your workflow looks basically the same as it did 18 months ago, you&#8217;re below the floor. Not because you&#8217;re bad at your job, but because the floor moved.</p><p>The good news: the floor is achievable. We&#8217;re not talking about becoming an AI researcher or rebuilding your entire skill set. It&#8217;s a handful of habits and a commitment to the experimentation loop. The people who&#8217;ve already made this shift will tell you it took weeks, not months.</p><p>The ceiling will keep rising. The companies building these tools will keep pushing what&#8217;s possible. That&#8217;s great. Someone needs to be doing that work.</p><p>It&#8217;s easier than ever to make stuff. It&#8217;s faster. And AI can be super confident about correctly making the wrong solution and/or a complete waste of time/talent/tokens. It doesn&#8217;t care if you&#8217;re right, just that you use more tokens.</p><p>It&#8217;s up to us, humans, to make sure we build the right things as well as we can.</p>]]></content:encoded></item><item><title><![CDATA[Weekly Headlines: Issue #14]]></title><description><![CDATA[March 12 - 19, 2026]]></description><link>https://tsw.blankmetal.ai/p/weekly-headlines-issue-14</link><guid isPermaLink="false">https://tsw.blankmetal.ai/p/weekly-headlines-issue-14</guid><dc:creator><![CDATA[Blank Metal]]></dc:creator><pubDate>Fri, 20 Mar 2026 13:03:48 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!7iQI!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F52fc5d61-f240-40bc-9f0c-ea386acf7e6a_1200x670.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!7iQI!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F52fc5d61-f240-40bc-9f0c-ea386acf7e6a_1200x670.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!7iQI!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F52fc5d61-f240-40bc-9f0c-ea386acf7e6a_1200x670.png 424w, https://substackcdn.com/image/fetch/$s_!7iQI!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F52fc5d61-f240-40bc-9f0c-ea386acf7e6a_1200x670.png 848w, https://substackcdn.com/image/fetch/$s_!7iQI!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F52fc5d61-f240-40bc-9f0c-ea386acf7e6a_1200x670.png 1272w, https://substackcdn.com/image/fetch/$s_!7iQI!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F52fc5d61-f240-40bc-9f0c-ea386acf7e6a_1200x670.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!7iQI!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F52fc5d61-f240-40bc-9f0c-ea386acf7e6a_1200x670.png" width="1200" height="670" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/52fc5d61-f240-40bc-9f0c-ea386acf7e6a_1200x670.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:670,&quot;width&quot;:1200,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:1480104,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://tsw.blankmetal.ai/i/191519386?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F52fc5d61-f240-40bc-9f0c-ea386acf7e6a_1200x670.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!7iQI!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F52fc5d61-f240-40bc-9f0c-ea386acf7e6a_1200x670.png 424w, https://substackcdn.com/image/fetch/$s_!7iQI!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F52fc5d61-f240-40bc-9f0c-ea386acf7e6a_1200x670.png 848w, https://substackcdn.com/image/fetch/$s_!7iQI!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F52fc5d61-f240-40bc-9f0c-ea386acf7e6a_1200x670.png 1272w, https://substackcdn.com/image/fetch/$s_!7iQI!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F52fc5d61-f240-40bc-9f0c-ea386acf7e6a_1200x670.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Welcome to Blank Metal&#8217;s Weekly AI Headlines.</p><p>Each week, our team shares the AI stories that caught our attention&#8212;the articles, announcements, and insights we&#8217;re actually discussing internally. We curate the best of what we&#8217;re reading and add the context that matters: what happened, why it matters, and what to do about it.</p><p>Short, sharp, and focused on impact.</p><div><hr></div><h1>The Reckoning</h1><p><em>Three stories this week share a throughline: the costs of moving fast with AI are becoming visible. Token bills, comprehension gaps, and bubble economics are all different faces of the same question&#8212;what happens when the honeymoon ends?</em></p><h2>You&#8217;ve Figured Out AI at Work&#8212;Now Comes the Bill</h2><p><strong>What:</strong> The Wall Street Journal reports that enterprises are hitting a new phase of AI adoption: the token bill. Companies that moved aggressively from pilots to production are discovering that AI inference costs scale faster than they expected. The productivity gains are real, but so is the compute bill&#8212;and most organizations didn&#8217;t budget for what production-scale AI actually costs.</p><p><strong>So What:</strong> This is the hangover after the honeymoon. The first wave was &#8220;look what AI can do.&#8221; The second wave was &#8220;let&#8217;s put it everywhere.&#8221; The third wave&#8212;happening now&#8212;is &#8220;who&#8217;s paying for all these tokens?&#8221; This isn&#8217;t a reason to slow down, but it is a reason to be intentional about where AI creates enough value to justify the cost. Not every workflow needs a frontier model.</p><p><strong>Now What:</strong> Audit your AI usage against actual business value. The 80/20 rule applies: a small number of AI-powered workflows are probably driving most of your value, while a long tail of lower-value uses are burning tokens. Right-size your model selection&#8212;use smaller, faster models for routine tasks and save frontier models for high-stakes decisions.</p><p><a href="https://www.wsj.com/tech/ai/ai-tokens-productivity-d35c6bd8">Read more</a></p><h2>Comprehension Debt: The Hidden Cost Nobody&#8217;s Measuring</h2><p><strong>What:</strong> Addy Osmani coined &#8220;comprehension debt&#8221;&#8212;the growing gap between how much code exists in your system and how much any human genuinely understands. Unlike technical debt, which creates visible friction, comprehension debt grows silently until your system breaks and nobody can fix it. An Anthropic study found developers using AI assistance scored 17% lower on comprehension quizzes than control groups.</p><p><strong>So What:</strong> Your team just shipped 10x faster. Congratulations&#8212;you now have 10x more code that nobody fully understands. Tests pass, CI is green, but when something breaks at 2am, the person on call has to reason about code they never wrote, never reviewed, and never internalized. This is a fundamentally different failure mode than technical debt.</p><p><strong>Now What:</strong> Treat genuine understanding&#8212;not passing tests&#8212;as non-negotiable. One practical step: require that AI-generated code gets the same review depth as human-written code. If your team is skimming AI output because &#8220;it looks right,&#8221; that&#8217;s the debt accumulating. The teams building comprehension discipline now will be better positioned when the reckoning arrives.</p><p><a href="https://addyosmani.com/blog/comprehension-debt/">Read more</a></p><h2>Yes, AI Is a Bubble. The Interesting Question Is What Kind.</h2><p><strong>What:</strong> Derek Thompson and Paul Kedrosky make the case that AI is definitively a bubble&#8212;private AI spending will exceed $700 billion in 2026, representing 50-80% of quarterly GDP growth, more than the combined historical spending on 1930s public works, the Manhattan Project, Apollo, and the Interstate Highway System. But they argue it&#8217;s a &#8220;rational bubble&#8221;: each individual actor is behaving rationally, even as the collective outcome is economically unsustainable.</p><p><strong>So What:</strong> The historical parallel that matters isn&#8217;t dot-com&#8212;it&#8217;s railroads. By 1900, railroads were 62% of U.S. market capitalization despite massive overbuilding, with half of peak-period track eventually abandoned. Tech now represents roughly 60% of the index. The bubble will pop, but the infrastructure will remain and reshape everything it touches. Anthropic doubled revenue in two months. OpenAI added $1B annualized revenue per week. Stripe reports AI companies growing faster than any previous generation.</p><p><strong>Now What:</strong> Build on the infrastructure while the bubble funds it, but don&#8217;t mistake bubble economics for sustainable economics. The companies that thrive post-correction will be the ones generating real revenue from real workflows&#8212;not the ones burning venture capital on AI features nobody asked for. If your AI investment can&#8217;t justify itself on unit economics today, it won&#8217;t survive the correction.</p><p><a href="https://www.derekthompson.org/p/yes-ai-is-a-bubble-there-is-no-question">Read more</a></p><div><hr></div><h1>The Human Variable</h1><p><em>AI&#8217;s biggest open question isn&#8217;t technical&#8212;it&#8217;s human. How do 81,000 users actually feel about it? What happens to the people who built the systems? And why does every organization think it&#8217;s further along than it actually is?</em></p><h2>What 81,000 People Actually Want from AI</h2><p><strong>What:</strong> Anthropic published the largest multilingual qualitative study of AI users ever conducted&#8212;80,508 Claude users across 159 countries. The headline finding: people don&#8217;t split cleanly into optimists and pessimists. Those who want emotional AI support are 3x more likely to also fear dependency on it. 81% say AI has already delivered on some aspect of their vision.</p><p><strong>So What:</strong> The framing of &#8220;AI believers vs. skeptics&#8221; is wrong. Real users hold both simultaneously&#8212;they want the productivity gains (32% cite this as the primary delivered benefit) while worrying about job displacement (22.3%) and loss of autonomy (21.9%). Lower-income countries are significantly more optimistic than wealthy ones, which inverts the usual tech adoption narrative.</p><p><strong>Now What:</strong> If you&#8217;re rolling out AI tools internally, don&#8217;t segment your workforce into supporters and resisters. Design adoption programs that acknowledge both the excitement and the anxiety&#8212;because the same people feel both. The &#8220;cognitive partnership&#8221; framing (17% of users describe AI this way) resonates more than &#8220;productivity tool.&#8221;</p><p><a href="https://www.anthropic.com/features/81k-interviews">Read more</a></p><h2>What Do Coders Do After AI?</h2><p><strong>What:</strong> Anil Dash, writing for the New York Times Magazine, draws a line that most AI commentary misses: &#8220;In the creative disciplines, LLMs take away the most soulful human parts of the work and leave the drudgery to you. In coding, LLMs take away the drudgery and leave the human, soulful parts to you.&#8221; He identifies two cohorts of coders&#8212;the 9-to-5 professionals facing devastating displacement, and the craftspeople watching their medium transform into something unrecognizable.</p><p><strong>So What:</strong> 700,000 tech workers have been laid off in the last few years. We&#8217;ll be at a million soon. But the displacement isn&#8217;t uniform. The &#8220;journeyman coders&#8221; writing standardized business logic are the most vulnerable&#8212;that&#8217;s exactly the code LLMs generate best. Meanwhile, coders who see it as craft are experiencing a different kind of loss: their job is becoming &#8220;describing software&#8221; rather than writing it. Both are painful, but they require completely different responses.</p><p><strong>Now What:</strong> If you manage engineering teams, this framework matters for retention and hiring. Your most valuable people aren&#8217;t the ones who write the most code&#8212;they&#8217;re the ones who understand why the system works. As Osmani&#8217;s comprehension debt concept makes clear, the ability to reason about code is becoming more valuable than the ability to write it. Hire for judgment, not velocity.</p><p><a href="https://www.anildash.com/2026/03/13/coders-after-ai/">Read more</a></p><h2>What&#8217;s Your AI Adoption Level?</h2><p><strong>What:</strong> Steve Yegge published an AI adoption maturity framework that&#8217;s resonating across the industry&#8212;a clear progression from &#8220;Not Using AI&#8221; through &#8220;AI-Assisted&#8221; to &#8220;AI-Native&#8221; with specific behaviors at each level. The framework maps where individuals and organizations actually sit versus where they think they are.</p><p><strong>So What:</strong> Most organizations overestimate their AI maturity because they conflate tool access with adoption. Having ChatGPT licenses doesn&#8217;t make you AI-assisted any more than having a gym membership makes you fit. The framework exposes the gap between &#8220;we have AI tools&#8221; and &#8220;our workflows have fundamentally changed.&#8221;</p><p><strong>Now What:</strong> Use this as a self-assessment. Where does your team actually sit&#8212;not where leadership thinks they sit? The honest answer shapes whether you need more tools, more training, or more workflow redesign. Most organizations discover they need the third one.</p><p><a href="https://x.com/juristr/status/2033568215956418673">Read more</a></p><div><hr></div><h1>The Agent Economy</h1><p><em>Design tools that replace designers. Enterprise leaders planning agent deployments. A strategist declaring the bubble debate over. The agent economy isn&#8217;t emerging&#8212;it&#8217;s arriving, and the market is repricing everything around it.</em></p><h2>Google Launches &#8220;Vibe Design&#8221; with Stitch&#8212;Figma Drops 8%</h2><p><strong>What:</strong> Google Labs unveiled Stitch, an AI-native UI design platform with an AI canvas, smarter design agent, voice input, instant prototyping, and built-in design system support. The market reacted immediately&#8212;Figma&#8217;s stock dropped 8% on the announcement, now down 80% from its August 2025 IPO.</p><p><strong>So What:</strong> This is the design tool version of what happened to coding: AI collapses the gap between intent and artifact. Stitch doesn&#8217;t just assist designers&#8212;it lets non-designers produce high-fidelity UI through natural language and voice. The stock reaction tells you the market believes this shift is structural, not incremental.</p><p><strong>Now What:</strong> If your team is evaluating design tooling or hiring designers, watch this space closely. The question is shifting from &#8220;which design tool?&#8221; to &#8220;do we need the same number of designers?&#8221;&#8212;and the answer will look different in six months than it does today.</p><p><a href="https://blog.google/innovation-and-ai/models-and-research/google-labs/stitch-ai-ui-design/">Read more</a></p><h2>Aaron Levie: What 20+ Enterprise IT Leaders Are Actually Saying About AI</h2><p><strong>What:</strong> Box CEO Aaron Levie sat down with 20+ enterprise AI and IT leaders&#8212;particularly from regulated industries&#8212;and shared the emerging consensus. Agents are &#8220;clearly the big thing,&#8221; with enterprises moving from experimental chatbots to production agent deployments. But the infrastructure isn&#8217;t ready: governance models are immature, payment rails for machine-to-machine transactions don&#8217;t exist, and most organizations are still figuring out where agents fit in their org charts.</p><p><strong>So What:</strong> When the CEO of a $5B enterprise software company reports from the field, it&#8217;s a demand signal. The shift from &#8220;chatbot pilots&#8221; to &#8220;agent deployments&#8221; is happening, but the gap between ambition and infrastructure is widening. Only one in five companies has a mature governance model for agent deployments. The rest are flying blind or moving slowly.</p><p><strong>Now What:</strong> If you&#8217;re planning enterprise AI rollouts, governance and observability should be in your architecture from day one&#8212;not bolted on after agents are already running. The organizations that get agent governance right early will move faster later. The ones that skip it will hit a wall when the first production agent does something unexpected.</p><p><a href="https://x.com/levie/status/2034484203522261293">Read more</a></p><h2>Ben Thompson: Why Agents Mean This Isn&#8217;t a Bubble</h2><p><strong>What:</strong> Ben Thompson makes his most definitive macro call on AI yet: we&#8217;re not in a bubble. His argument rests on three LLM paradigm shifts&#8212;ChatGPT (2022), reasoning models like o1 (2024), and agents via Opus 4.5/Claude Code (late 2025). Each shift addressed a core LLM weakness, and agents are the inflection that changes the economics. The key insight: agents don&#8217;t just require a better model&#8212;they require integration between model and harness, which means Anthropic and OpenAI are becoming the differentiated point in the value chain, not commoditized infrastructure.</p><p><strong>So What:</strong> Thompson identifies two dynamics that separate agents from prior AI hype. First, agents dramatically reduce the number of humans needed to drive compute demand&#8212;a small number of people wielding agents creates exponentially more economic output than chatbot adoption ever could. Second, Microsoft&#8217;s decision to bundle Anthropic&#8217;s Claude into its new $99/seat E7 enterprise tier (via Copilot Cowork) is an admission that model-agnostic strategies don&#8217;t work for agents. If agents require integrated model+harness, the companies building that integration capture the profits.</p><p><strong>Now What:</strong> If Thompson is right, the strategic question for enterprises shifts. It&#8217;s not &#8220;which model should we use?&#8221; but &#8220;which agent platform are we building on?&#8221; The model-agnostic approach that seemed prudent a year ago may now be a liability&#8212;because agents aren&#8217;t modular. For organizations evaluating AI investments, this argues for deeper commitment to fewer platforms rather than hedging across many.</p><p><a href="https://stratechery.com/2026/agents-over-bubbles/">Read more</a></p><div><hr></div><h1>The Practitioner&#8217;s Edge</h1><p><em>Two tools this week that separate the people talking about AI from the people building with it.</em></p><h2>The MCP Debate Settles: CLI for Developers, MCP for Organizations</h2><p><strong>What:</strong> A viral blog post declared &#8220;MCP is Dead&#8221; in favor of CLI tools, arguing that LLMs already know jq and curl so MCP wrappers add unnecessary complexity. Cloudflare responded with &#8220;Code Mode&#8221;&#8212;a new approach where AI agents write TypeScript against MCP tool APIs instead of using specialized tool-calling syntax, improving both performance and token efficiency by 47%.</p><p><strong>So What:</strong> Both sides are right about different problems. CLI tools win for individual developers who already have the right access and know the tools. But MCP over streamable HTTP solves the enterprise problem: centralized tool servers with proper auth, shared infrastructure across teams, and audit trails. That&#8217;s the difference between one developer vibe-coding and an org shipping agents at scale.</p><p><strong>Now What:</strong> Stop debating MCP vs. CLI as a binary. Use CLI tools where the developer already has access and the LLM already knows the tool. Use MCP servers where you need centralized governance, shared access, and auditability. Cloudflare&#8217;s Code Mode suggests the best of both worlds: MCP infrastructure with code-native invocation patterns.</p><p><a href="https://chrlschn.dev/blog/2026/03/mcp-is-dead-long-live-mcp/">Read more</a></p><h2>Defuddle: The Markdown Converter LLM Workflows Need</h2><p><strong>What:</strong> Defuddle is a lightweight tool that converts any web page into clean Markdown with YAML frontmatter. Available as an API, browser extension, and bookmarklet&#8212;it also handles YouTube transcription. Think of it as a universal adapter between the messy web and the structured context that LLMs prefer.</p><p><strong>So What:</strong> LLMs&#8212;especially in coding and workflow contexts&#8212;perform dramatically better with Markdown input than raw HTML or copy-pasted text. Every time you paste a URL into an AI tool and get a mediocre response, the problem is often the input format, not the model. Tools like Defuddle solve the &#8220;last mile&#8221; problem of getting clean context into AI workflows.</p><p><strong>Now What:</strong> Add this to your AI toolkit. When feeding articles, documentation, or web content into AI workflows, convert to Markdown first. The token efficiency gains alone are worth it&#8212;but the real win is better AI output from cleaner input. For engineering teams, consider wrapping this in an MCP server for agent workflows.</p><p><a href="https://defuddle.md/">Read more</a></p>]]></content:encoded></item><item><title><![CDATA[Weekly Headlines: Issue #13]]></title><description><![CDATA[March 05 - March 12, 2026]]></description><link>https://tsw.blankmetal.ai/p/weekly-headlines-issue-13</link><guid isPermaLink="false">https://tsw.blankmetal.ai/p/weekly-headlines-issue-13</guid><dc:creator><![CDATA[Blank Metal]]></dc:creator><pubDate>Mon, 16 Mar 2026 13:53:10 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!oq3H!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcfe7b6ef-9990-4d97-b83d-f980d17a5adc_1200x670.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!oq3H!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcfe7b6ef-9990-4d97-b83d-f980d17a5adc_1200x670.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!oq3H!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcfe7b6ef-9990-4d97-b83d-f980d17a5adc_1200x670.png 424w, https://substackcdn.com/image/fetch/$s_!oq3H!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcfe7b6ef-9990-4d97-b83d-f980d17a5adc_1200x670.png 848w, https://substackcdn.com/image/fetch/$s_!oq3H!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcfe7b6ef-9990-4d97-b83d-f980d17a5adc_1200x670.png 1272w, https://substackcdn.com/image/fetch/$s_!oq3H!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcfe7b6ef-9990-4d97-b83d-f980d17a5adc_1200x670.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!oq3H!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcfe7b6ef-9990-4d97-b83d-f980d17a5adc_1200x670.png" width="1200" height="670" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/cfe7b6ef-9990-4d97-b83d-f980d17a5adc_1200x670.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:670,&quot;width&quot;:1200,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:1480192,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://tsw.blankmetal.ai/i/191130459?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcfe7b6ef-9990-4d97-b83d-f980d17a5adc_1200x670.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!oq3H!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcfe7b6ef-9990-4d97-b83d-f980d17a5adc_1200x670.png 424w, https://substackcdn.com/image/fetch/$s_!oq3H!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcfe7b6ef-9990-4d97-b83d-f980d17a5adc_1200x670.png 848w, https://substackcdn.com/image/fetch/$s_!oq3H!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcfe7b6ef-9990-4d97-b83d-f980d17a5adc_1200x670.png 1272w, https://substackcdn.com/image/fetch/$s_!oq3H!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcfe7b6ef-9990-4d97-b83d-f980d17a5adc_1200x670.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Welcome to Blank Metal&#8217;s Weekly AI Headlines.</p><p>Each week, our team shares the AI stories that caught our attention&#8212;the articles, announcements, and insights we&#8217;re actually discussing internally. We curate the best of what we&#8217;re reading and add the context that matters: what happened, why it matters, and what to do about it.</p><p>Short, sharp, and focused on impact.</p><h1>The Platform Split</h1><p><em>The AI market is fracturing into distinct ecosystems&#8212;and the governance frameworks being written now will determine which ones survive.</em></p><h2>a16z: The Gen AI Consumer App Market Is Splitting in Two</h2><p><strong>What:</strong> a16z&#8217;s 6th Top 100 Gen AI Consumer Apps report reveals ChatGPT and Claude are diverging into fundamentally different platforms&#8212;ChatGPT becoming a consumer super-app (Expedia, Instacart, ads) while Claude goes deep on professional tooling (PitchBook, FactSet, Sentry). Only 41 apps overlap between the two ecosystems out of ~370 combined.</p><p><strong>So What:</strong> The &#8220;iOS vs. Android&#8221; framing means enterprises choosing an AI platform are making a strategic bet on ecosystem direction, not just model quality. Claude Code hitting $1B ARR in six months proves coding agents are a real revenue category, not a feature.</p><p><strong>Now What:</strong> Map your team&#8217;s AI usage patterns&#8212;are you building for consumer workflows or professional tooling? Your platform choice should follow the ecosystem that matches your use case, not the loudest brand.</p><p><a href="https://a16z.com/100-gen-ai-apps-6/">Read more</a></p><h2>34 Principles for AI Governance&#8212;But Zero Mentions of &#8220;Open&#8221;</h2><p><strong>What:</strong> The Future of Life Institute released a cross-partisan AI governance declaration with 34 principles designed for direct legislative translation: mandatory kill switches, superintelligence moratoriums, criminal executive liability, and pharma-style chatbot safety testing.</p><p><strong>So What:</strong> This is the most legislative-ready AI governance framework yet&#8212;and the complete absence of open source, open weights, or right-to-run-locally language signals that regulation may default to a closed-model world if the open community doesn&#8217;t engage.</p><p><strong>Now What:</strong> If your AI strategy depends on open-source models, monitor this closely. These principles are written to become law, and they could reshape what&#8217;s legally deployable.</p><p><a href="https://humanstatement.org/">Read more</a></p><h1>AI-First Architecture Shifts</h1><p><em>Enterprise software is fundamentally restructuring around AI agents as primary users, not just assistants for humans.</em></p><h2>Box CEO: Build for Trillions of Agents, Not Just Humans</h2><p><strong>What:</strong> Aaron Levie argues that software architecture must shift to API-first design as AI agents become the primary users of enterprise applications, not humans.</p><p><strong>So What:</strong> This reframes how enterprises should evaluate and build software&#8212;if your systems aren&#8217;t agent-accessible, they risk becoming legacy infrastructure in an agent-driven workflow era.</p><p><strong>Now What:</strong> Audit your core systems for API coverage and consider whether your current vendors are building for human-only or agent-compatible futures.</p><p><a href="https://x.com/levie/status/2030714592238956960">Read more</a></p><h2>Claude Gets Native Microsoft Office Integration</h2><p><strong>What:</strong> Anthropic upgraded Claude to work directly with Excel spreadsheets and PowerPoint presentations, allowing users to analyze, edit, and create Office documents within the AI interface.</p><p><strong>So What:</strong> This closes a meaningful gap for enterprise teams who live in Microsoft&#8217;s ecosystem&#8212;reducing the copy-paste friction that slows down real-world AI adoption in document-heavy workflows.</p><p><strong>Now What:</strong> Test Claude on a repetitive Office task your team dreads (quarterly report formatting, data cleanup) to gauge whether it&#8217;s ready to slot into existing processes.</p><p><a href="https://www.thedeepview.com/articles/claude-strengthens-its-excel-powerpoint-skills">Read more</a></p><h1>Scaling AI in Production</h1><p><em>Leading tech companies are moving beyond pilots to organization-wide AI integration, revealing both blueprints and cautionary tales.</em></p><h2>Uber Reveals How It&#8217;s Scaling AI-Assisted Development</h2><p><strong>What:</strong> The Pragmatic Engineer offers an inside look at how Uber is integrating AI tools into its software development workflows across the organization.</p><p><strong>So What:</strong> Real-world case studies from engineering-forward companies like Uber provide a practical blueprint for enterprise teams trying to move past pilot projects into scaled AI adoption.</p><p><strong>Now What:</strong> Compare your AI development tooling rollout against Uber&#8217;s approach&#8212;particularly how they&#8217;re measuring productivity gains and managing adoption friction.</p><p><a href="https://newsletter.pragmaticengineer.com/p/how-uber-uses-ai-for-development">Read more</a></p><h2>Amazon Mandates AI Tools Even When They Slow Workers Down</h2><p><strong>What:</strong> Amazon is pushing employees to use AI assistants across workflows company-wide, even in cases where the tools are reportedly reducing productivity rather than improving it.</p><p><strong>So What:</strong> This signals a growing tension between AI adoption mandates and actual ROI&#8212;a cautionary tale for enterprise leaders feeling pressure to deploy AI everywhere, regardless of fit.</p><p><strong>Now What:</strong> Audit your own AI rollouts for &#8220;mandate creep&#8221; and build feedback loops that let teams flag when tools hurt more than help.</p><p><a href="https://www.theguardian.com/technology/ng-interactive/2026/mar/11/amazon-artificial-intelligence">Read more</a></p><h1>The Agent Workflow Revolution</h1><p><em>Autonomous coding agents are reshaping how product teams work and forcing a competitive reshuffling among AI providers.</em></p><h2>LangChain Founder Explores How Coding Agents Transform Product Teams</h2><p><strong>What:</strong> Harrison Chase shared insights on how coding agents are reshaping workflows across engineering, product, and design functions.</p><p><strong>So What:</strong> As coding agents mature beyond developer tools, enterprise leaders need to consider second-order effects on team structures, hiring, and cross-functional collaboration.</p><p><strong>Now What:</strong> Assess whether your current org design accounts for AI-augmented roles beyond just engineering.</p><p><a href="https://x.com/hwchase17/status/2031051115169808685">Read more</a></p><h2>OpenAI Scrambles to Match Anthropic&#8217;s Coding Agent Lead</h2><p><strong>What:</strong> Wired reports that OpenAI is racing to catch up to Claude Code, Anthropic&#8217;s autonomous coding agent that has gained significant traction among developers.</p><p><strong>So What:</strong> The competitive dynamics have flipped&#8212;OpenAI is now playing catch-up in the agentic coding space, which signals that enterprise teams shouldn&#8217;t assume market leaders will dominate every AI category.</p><p><strong>Now What:</strong> If you&#8217;re evaluating coding agents, benchmark actual performance on your codebase rather than defaulting to vendor relationships&#8212;this space is moving too fast for brand loyalty.</p><p><a href="https://www.wired.com/story/openai-codex-race-claude-code/">Read more</a></p><h1>The Privacy Backlash</h1><p><em>As AI embeds deeper into daily life, the counter-reaction is creating its own market.</em></p><h2>Counter-Surveillance Goes Consumer: Deveillance&#8217;s $1,199 Audio Jammer Goes Viral</h2><p><strong>What:</strong> Deveillance&#8217;s Spectre I&#8212;a portable device claiming to use AI to prevent nearby microphones from recording conversations&#8212;hit 4.3 million views and 42K bookmarks, despite security researchers questioning whether the tech delivers on its promises.</p><p><strong>So What:</strong> The demand signal matters more than the product: consumer anxiety about always-on AI listening is translating into real willingness to pay for privacy tools. The counter-surveillance market is forming faster than the products to serve it.</p><p><strong>Now What:</strong> For enterprise teams deploying AI in offices, meeting rooms, and customer spaces, the backlash against ambient recording is real. Factor privacy perception into your AI rollout strategy, not just compliance.</p><p><a href="https://www.deveillance.com/">Read more</a></p><h1>AI Investment at Any Cost</h1><p><em>Enterprise leaders are treating AI transformation as a strategic imperative worth painful trade-offs, even cutting profitable operations to fund the shift.</em></p><h2>Atlassian Cuts 10% of Staff to Fund AI Pivot</h2><p><strong>What:</strong> Atlassian is laying off roughly 10% of its workforce, redirecting the savings to accelerate its AI product investments.</p><p><strong>So What:</strong> This signals that even profitable enterprise software companies are treating AI not as an add-on budget item but as a strategic priority worth painful trade-offs&#8212;expect more &#8220;self-funded AI transformations&#8221; across the industry.</p><p><strong>Now What:</strong> If you&#8217;re building an AI business case, note that leadership teams are increasingly willing to make structural cuts to fund AI bets&#8212;frame your proposals accordingly.</p><p><a href="https://www.cnbc.com/2026/03/11/atlassian-slashes-10percent-of-workforce-to-self-fund-investments-in-ai.html">Read more</a></p>]]></content:encoded></item><item><title><![CDATA[Weekly Headlines: Issue #12]]></title><description><![CDATA[February 27 - March 5, 2026]]></description><link>https://tsw.blankmetal.ai/p/weekly-headlines-issue-12</link><guid isPermaLink="false">https://tsw.blankmetal.ai/p/weekly-headlines-issue-12</guid><dc:creator><![CDATA[Blank Metal]]></dc:creator><pubDate>Fri, 06 Mar 2026 14:04:03 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!snGx!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F91e244be-deee-4fb1-8b07-d5d2ce0761e1_1200x670.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!snGx!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F91e244be-deee-4fb1-8b07-d5d2ce0761e1_1200x670.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!snGx!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F91e244be-deee-4fb1-8b07-d5d2ce0761e1_1200x670.png 424w, https://substackcdn.com/image/fetch/$s_!snGx!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F91e244be-deee-4fb1-8b07-d5d2ce0761e1_1200x670.png 848w, https://substackcdn.com/image/fetch/$s_!snGx!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F91e244be-deee-4fb1-8b07-d5d2ce0761e1_1200x670.png 1272w, https://substackcdn.com/image/fetch/$s_!snGx!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F91e244be-deee-4fb1-8b07-d5d2ce0761e1_1200x670.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!snGx!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F91e244be-deee-4fb1-8b07-d5d2ce0761e1_1200x670.png" width="1200" height="670" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/91e244be-deee-4fb1-8b07-d5d2ce0761e1_1200x670.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:670,&quot;width&quot;:1200,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:1479879,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://tsw.blankmetal.ai/i/190103837?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F91e244be-deee-4fb1-8b07-d5d2ce0761e1_1200x670.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!snGx!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F91e244be-deee-4fb1-8b07-d5d2ce0761e1_1200x670.png 424w, https://substackcdn.com/image/fetch/$s_!snGx!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F91e244be-deee-4fb1-8b07-d5d2ce0761e1_1200x670.png 848w, https://substackcdn.com/image/fetch/$s_!snGx!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F91e244be-deee-4fb1-8b07-d5d2ce0761e1_1200x670.png 1272w, https://substackcdn.com/image/fetch/$s_!snGx!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F91e244be-deee-4fb1-8b07-d5d2ce0761e1_1200x670.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Welcome to Blank Metal&#8217;s Weekly AI Headlines.</p><p>Each week, our team shares the AI stories that caught our attention&#8212;the articles, announcements, and insights we&#8217;re actually discussing internally. We curate the best of what we&#8217;re reading and add the context that matters: what happened, why it matters, and what to do about it.</p><p>Short, sharp, and focused on impact.</p><h2>Anthropic Refuses Pentagon Demands, Gets Blacklisted as &#8220;Supply Chain Risk&#8221;</h2><p><strong>What:</strong> Anthropic refused the Pentagon&#8217;s demand to remove all safeguards on military use of its Claude models &#8212; specifically protections against domestic mass surveillance and fully autonomous weapons. In response, President Trump directed all federal agencies to stop using Anthropic&#8217;s technology, and Defense Secretary Pete Hegseth designated the company a &#8220;supply chain risk&#8221; &#8212; a classification typically reserved for foreign adversaries like <a href="https://www.huawei.com/en/">Huawei</a>. The designation bars every defense contractor from doing business with Anthropic.</p><p><strong>So What:</strong> This is unprecedented. An American AI company is being treated like a hostile foreign entity because it insisted on safety red lines. Anthropic&#8217;s CEO called the designation &#8220;legally unsound&#8221; and pledged to challenge it in court. The signal to every enterprise leader: the U.S. government is now willing to use economic coercion against American companies that set limits on how their technology is deployed. The Lawfare Institute&#8217;s legal analysis suggests the designation likely won&#8217;t survive judicial review, but the chilling effect on other AI companies is the point.</p><p><strong>Now What:</strong> If your organization uses Anthropic products, don&#8217;t panic &#8212; this designation targets defense contractors, not commercial enterprises. But watch the legal challenge closely. The outcome will define the boundaries of AI safety commitments for the entire industry. Anthropic&#8217;s willingness to absorb this level of government pressure is either principled courage or an existential gamble. The market will decide.</p><p><a href="https://www.axios.com/2026/02/27/anthropic-pentagon-supply-chain-risk-claude">Read more</a></p><h2>OpenAI Cuts Pentagon Deal &#8212; Then Scrambles to Rewrite It</h2><p><strong>What:</strong> Hours after Anthropic was blacklisted, OpenAI announced it had reached a deal allowing the Pentagon to use its technology in classified environments. The deal included stated protections against mass surveillance and fully autonomous weapons. Then the backlash hit &#8212; hard. Internal employees were &#8220;fuming,&#8221; and CEO Sam Altman publicly admitted the announcement &#8220;looked opportunistic and sloppy&#8221; and that he &#8220;shouldn&#8217;t have rushed.&#8221; Within days, OpenAI and the Pentagon agreed to rewrite the contract language, adding explicit prohibitions against &#8220;deliberate tracking, surveillance, or monitoring of U.S. persons.&#8221;</p><p><strong>So What:</strong> MIT Technology Review put it bluntly: &#8220;OpenAI&#8217;s compromise with the Pentagon is what Anthropic feared.&#8221; The speed of the backlash &#8212; and Altman&#8217;s rare public admission of error &#8212; reveals how politically charged military AI has become. The amended contract language is stronger, but the episode exposed a fundamental tension: OpenAI is simultaneously raising $110B from investors who want government contracts and employing workers who signed an open letter demanding guardrails. That tension isn&#8217;t going away.</p><p><strong>Now What:</strong> Enterprise buyers should be watching the actual contract language, not the press releases. When two leading AI companies offer the same technology to the same customer with different safety terms, the terms matter. Ask your AI vendors: what are your red lines? The answer reveals their risk tolerance &#8212; and by extension, yours.</p><p><a href="https://www.technologyreview.com/2026/03/02/1133850/openais-compromise-with-the-pentagon-is-what-anthropic-feared/">Read more</a></p><h2>&#8220;We Will Not Be Divided&#8221;: 900 AI Workers Demand Military AI Red Lines</h2><p><strong>What:</strong> Nearly 900 employees at Google and OpenAI signed an open letter titled &#8220;We Will Not Be Divided,&#8221; urging their companies to join Anthropic in refusing the Pentagon&#8217;s demands. About 100 signers were from OpenAI, roughly 800 from Google, and half chose to attach their names publicly. The letter warns: &#8220;They&#8217;re trying to divide each company with fear that the other will give in.&#8221; By Monday, the letter&#8217;s momentum had accelerated after U.S. strikes on Iran raised the stakes of military AI use.</p><p><strong>So What:</strong> This is the largest coordinated action by AI workers since Google&#8217;s Project Maven protests in 2018 &#8212; but the context is different. In 2018, employees objected to their employer&#8217;s contract. In 2026, employees are organizing across competing companies to defend a rival&#8217;s position. That&#8217;s a remarkable shift. It signals that a significant cohort of AI researchers and engineers view military AI guardrails as a shared professional standard, not a competitive differentiator.</p><p><strong>Now What:</strong> If you&#8217;re hiring AI talent, understand that military AI policy is now a retention factor. Top engineers are choosing employers based on ethical commitments, not just compensation. The letter&#8217;s cross-company solidarity suggests that talent will flow toward companies with clear guardrails &#8212; and away from those without them.</p><p><a href="https://notdivided.org">Read more</a></p><h2>OpenAI Raises $110B at $730B Valuation &#8212; The Largest Private Funding Round in History</h2><p><strong>What:</strong> OpenAI closed $110 billion in new funding &#8212; $50B from Amazon, $30B from Nvidia, $30B from SoftBank &#8212; at a $730 billion pre-money valuation. The round jumped from a $500B valuation just four months earlier. As part of the deal, AWS becomes the exclusive third-party cloud distributor for OpenAI Frontier, and the companies are scaling their compute agreement to 2 gigawatts of Trainium chips.</p><p><strong>So What:</strong> The numbers are staggering, but the structure is the story. Amazon isn&#8217;t just investing &#8212; it&#8217;s locking OpenAI into AWS infrastructure. Nvidia isn&#8217;t just investing &#8212; it&#8217;s guaranteeing demand for its hardware. SoftBank isn&#8217;t just investing &#8212; it&#8217;s building on its Stargate joint venture. Each investor is buying strategic positioning, not just equity. The valuation implies investors believe OpenAI will generate revenue comparable to the world&#8217;s largest software companies within 3-5 years. That&#8217;s either conviction or collective delusion, and there&#8217;s no middle ground at $730B.</p><p><strong>Now What:</strong> For enterprise AI strategy, the Amazon-AWS exclusive distribution deal matters more than the dollar amount. If your organization runs on AWS, OpenAI models through Bedrock just became a first-class integration path. If you&#8217;re multi-cloud, this exclusivity may push you toward specific infrastructure choices you didn&#8217;t plan to make.</p><p><a href="https://techcrunch.com/2026/02/27/openai-raises-110b-in-one-of-the-largest-private-funding-rounds-in-history/">Read more</a></p><h2>&#8220;The Week the AI Jobs Wipeout Got Real&#8221;</h2><p><strong>What:</strong> Three major publications converged on the same story simultaneously. The Wall Street Journal declared it &#8220;the week the dreaded AI jobs wipeout got real&#8221; after Block CEO Jack Dorsey laid off 4,000 people. Bloomberg reported that AI coding agents are &#8220;fueling a productivity panic&#8221; &#8212; engineers are working longer hours, not fewer, as the race to ship AI-augmented output intensifies. The New York Times documented India&#8217;s back-office industry beginning to contract as AI automation reaches outsourced knowledge work. Meanwhile, Harry Stebbings reported that three founders with 500-1,000 employees are all planning minimum 20% headcount cuts.</p><p><strong>So What:</strong> The narrative shifted this week from &#8220;AI might displace workers someday&#8221; to &#8220;it&#8217;s happening now, at scale, at named companies.&#8221; But the Bloomberg data complicates the simple &#8220;AI replaces humans&#8221; story &#8212; the engineers still employed are working more, not less. AI isn&#8217;t eliminating work; it&#8217;s compressing the timeline for what&#8217;s expected and raising the bar for output per person. The Dallas Fed&#8217;s research confirms the paradox: AI is simultaneously aiding and replacing workers, with the balance depending entirely on the role.</p><p><strong>Now What:</strong> If your organization hasn&#8217;t modeled what 20-30% more output per knowledge worker looks like &#8212; in terms of capacity planning, team structure, and career paths &#8212; you&#8217;re behind. The question isn&#8217;t whether headcount will change. It&#8217;s whether your organization will proactively redesign work around AI capabilities or reactively cut heads when competitors do.</p><p><a href="https://www.wsj.com/tech/ai/the-week-the-dreaded-ai-jobs-wipeout-got-real-3ba50504">Read more</a></p><h2>Amazon and OpenAI Unveil Stateful Runtime Environment for AI Agents</h2><p><strong>What:</strong> Buried in the $50B Amazon-OpenAI partnership announcement is a product that could reshape enterprise AI architecture: the Stateful Runtime Environment, launching on Amazon Bedrock. Instead of stitching together disconnected stateless API calls, agents get persistent working context &#8212; memory that carries forward, tool and workflow state, environment access, and identity boundaries. Think of it as the difference between an intern who forgets everything between conversations and a colleague who remembers the project.</p><p><strong>So What:</strong> This directly addresses the biggest engineering bottleneck in production AI agents: state management. Today, every enterprise building agentic workflows has to build its own orchestration layer &#8212; storing state, managing tool invocations, handling errors, maintaining permissions. OpenAI and Amazon are saying: stop building that plumbing, use ours. If it works as described, this could collapse months of custom agent infrastructure into a managed service. The InfoWorld analysis frames it as a &#8220;control plane power shift&#8221; &#8212; whoever owns agent state owns the agent ecosystem.</p><p><strong>Now What:</strong> If your team is building agentic workflows on AWS, request early access to the Stateful Runtime Environment immediately. If you&#8217;ve already built custom agent orchestration, evaluate whether this managed service could replace it. The risk of building on proprietary infrastructure is lock-in; the risk of not building on it is rebuilding what Amazon gives away for free.</p><p><a href="https://openai.com/index/introducing-the-stateful-runtime-environment-for-agents-in-amazon-bedrock/">Read more</a></p><h2>Scott Belsky: &#8220;The Orchestration Layer Is the New Interface Layer&#8221;</h2><p><strong>What:</strong> Former Adobe CPO Scott Belsky declared that the critical layer in enterprise AI has shifted: &#8220;The orchestration layer is the new interface layer. As we spend our day coordinating agent workflows &#8212; in a model-agnostic fashion, local and cloud &#8212; and validating outputs, the ultimate layer to own is where coordination takes place.&#8221; This represents an evolution from his earlier thesis that Interface &gt; Data &gt; Models, now placing orchestration at the top of the stack.</p><p><strong>So What:</strong> Belsky is naming what enterprise architects are discovering in practice: the competitive advantage in AI isn&#8217;t which model you use &#8212; it&#8217;s how you coordinate multiple agents, validate their outputs, and manage the human-in-the-loop decision points. This maps directly to what Box CEO Aaron Levie said separately &#8212; that agents need their own computer and filesystem, making the orchestration of those environments the key architectural challenge. When two of the most influential product thinkers in tech converge on &#8220;orchestration is the new interface,&#8221; it&#8217;s worth paying attention.</p><p><strong>Now What:</strong> Evaluate your AI architecture through this lens: who owns the orchestration layer? If the answer is &#8220;nobody yet&#8221; or &#8220;we&#8217;re building it ad hoc,&#8221; that&#8217;s your highest-leverage investment. The companies that build robust orchestration &#8212; agent coordination, output validation, approval workflows, state management &#8212; will compound their AI capabilities faster than those still debating which model to use.</p><p><a href="https://x.com/scottbelsky/status/2028303168073793542">Read more</a></p><h2>Simon Willison: The Practitioner&#8217;s Guide to Agentic Engineering</h2><p><strong>What:</strong> Simon Willison &#8212; creator of Datasette, Django co-creator, and one of the most respected voices in practical AI engineering &#8212; published &#8220;Agentic Engineering Patterns,&#8221; a growing guide to getting the best results from coding agents. The standout chapter, &#8220;Hoard Things You Know How to Do,&#8221; argues that the most valuable asset in an agent-driven workflow isn&#8217;t the model &#8212; it&#8217;s your accumulated collection of working examples, proof-of-concepts, and documented solutions. Coding agents make these hoarded assets dramatically more valuable because they can be recombined and adapted at machine speed.</p><p><strong>So What:</strong> This is the practitioner&#8217;s answer to all the theoretical &#8220;agents will replace developers&#8221; discourse. Willison&#8217;s patterns &#8212; red/green TDD with agents, specific prompt structures, building personal knowledge repositories &#8212; are battle-tested techniques from someone shipping real software with AI daily. The core insight is counterintuitive: the more capable AI coding agents become, the more valuable human experience becomes, because experience is what tells you which problems are solvable and which approaches will work.</p><p><strong>Now What:</strong> If your engineering team is adopting AI coding tools, Willison&#8217;s guide should be required reading. Start with the &#8220;hoard&#8221; principle: document your solutions, build proof-of-concepts, keep working examples of everything. These become compound assets &#8212; every problem you&#8217;ve solved once becomes a template for AI to solve similar problems faster.</p><p><a href="https://simonwillison.net/guides/agentic-engineering-patterns/">Read more</a></p><h2>Harry Stebbings: VC and PE Firms Must Deploy Their Own Autonomous Agents</h2><p><strong>What:</strong> Harry Stebbings argued that the deciding factor for investment firms in 2026 isn&#8217;t which AI tools they use &#8212; it&#8217;s whether they&#8217;ve deployed autonomous agents that actually do work. The shift from &#8220;AI as copilot&#8221; to &#8220;AI as team member&#8221; is the transition that unlocks real operational leverage. Separately, Hiten Shah reinforced the pattern: &#8220;This is one manifestation of what SaaS morphs into soon &#8212; deploy an agent per client.&#8221;</p><p><strong>So What:</strong> This directly validates what some PE firms are already discovering &#8212; that the firms deploying agents for deal research, portfolio monitoring, and operational analysis are pulling ahead of those still using AI as a search engine. The &#8220;agent per client&#8221; framing from Shah is particularly provocative: it suggests the SaaS business model itself evolves from &#8220;software you access&#8221; to &#8220;agents that work for you.&#8221; Investment firms that treat AI adoption as a tool-selection exercise are missing the architectural shift underneath.</p><p><strong>Now What:</strong> If you&#8217;re in PE or VC, ask: do you have agents that run autonomously &#8212; doing research, monitoring portfolios, generating reports &#8212; or do you have people prompting chatbots? The gap between those two is the gap between incremental efficiency and structural competitive advantage. Start with one high-value workflow (deal screening, competitor monitoring, portco reporting) and build an agent that runs it end-to-end.</p><p><a href="https://x.com/HarryStebbings/status/2028225013120475598">Read more</a></p><h2>Anthropic&#8217;s AI Fluency Index: It&#8217;s Not How Much You Use AI &#8212; It&#8217;s How Well</h2><p><strong>What:</strong> Anthropic published the AI Fluency Index, tracking 11 observable behaviors across nearly 10,000 Claude conversations to measure how effectively people collaborate with AI. The key finding: 85.7% of conversations showed iteration and refinement &#8212; users building on previous exchanges rather than accepting the first response. Users who iterate exhibit 2.67 additional fluency behaviors on average, roughly double the rate of those who don&#8217;t.</p><p><strong>So What:</strong> This reframes the enterprise AI adoption conversation from &#8220;how many people are using it&#8221; to &#8220;how well are they using it.&#8221; Most organizations measure AI adoption by login counts and message volume. Anthropic is arguing those are vanity metrics. The behaviors that predict better outcomes &#8212; iterating, clarifying goals, questioning the model&#8217;s reasoning, identifying missing context &#8212; are teachable skills, not innate abilities. That makes AI fluency a training problem, not a technology problem.</p><p><strong>Now What:</strong> Stop measuring AI adoption by usage volume. Start measuring by behavior quality. The 11 fluency behaviors Anthropic identified are a ready-made rubric for enterprise training programs. If your team accepts Claude&#8217;s first response without iteration, you&#8217;re leaving most of the value on the table.</p><p><a href="https://www.anthropic.com/research/AI-fluency-index">Read more</a>-</p>]]></content:encoded></item><item><title><![CDATA[Weekly Headlines: Issue #11]]></title><description><![CDATA[February 20 - February 27, 2026]]></description><link>https://tsw.blankmetal.ai/p/weekly-headlines-issue-11</link><guid isPermaLink="false">https://tsw.blankmetal.ai/p/weekly-headlines-issue-11</guid><dc:creator><![CDATA[Blank Metal]]></dc:creator><pubDate>Fri, 27 Feb 2026 14:02:38 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!COtD!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0b1d4a4f-f3e4-4679-8d49-43bb615fab0e_1200x670.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!COtD!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0b1d4a4f-f3e4-4679-8d49-43bb615fab0e_1200x670.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!COtD!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0b1d4a4f-f3e4-4679-8d49-43bb615fab0e_1200x670.png 424w, https://substackcdn.com/image/fetch/$s_!COtD!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0b1d4a4f-f3e4-4679-8d49-43bb615fab0e_1200x670.png 848w, https://substackcdn.com/image/fetch/$s_!COtD!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0b1d4a4f-f3e4-4679-8d49-43bb615fab0e_1200x670.png 1272w, https://substackcdn.com/image/fetch/$s_!COtD!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0b1d4a4f-f3e4-4679-8d49-43bb615fab0e_1200x670.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!COtD!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0b1d4a4f-f3e4-4679-8d49-43bb615fab0e_1200x670.png" width="1200" height="670" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/0b1d4a4f-f3e4-4679-8d49-43bb615fab0e_1200x670.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:670,&quot;width&quot;:1200,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:1479886,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://tsw.blankmetal.ai/i/189356005?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0b1d4a4f-f3e4-4679-8d49-43bb615fab0e_1200x670.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!COtD!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0b1d4a4f-f3e4-4679-8d49-43bb615fab0e_1200x670.png 424w, https://substackcdn.com/image/fetch/$s_!COtD!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0b1d4a4f-f3e4-4679-8d49-43bb615fab0e_1200x670.png 848w, https://substackcdn.com/image/fetch/$s_!COtD!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0b1d4a4f-f3e4-4679-8d49-43bb615fab0e_1200x670.png 1272w, https://substackcdn.com/image/fetch/$s_!COtD!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0b1d4a4f-f3e4-4679-8d49-43bb615fab0e_1200x670.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Welcome to Blank Metal&#8217;s Weekly AI Headlines.</p><p>Each week, our team shares the AI stories that caught our attention&#8212;the articles, announcements, and insights we&#8217;re actually discussing internally. We curate the best of what we&#8217;re reading and add the context that matters: what happened, why it matters, and what to do about it.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://tsw.blankmetal.ai/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading The So What! Subscribe for free to get headlines in your inbox every Friday.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p>Short, sharp, and focused on impact.</p><h2>Anthropic Enterprise Event Rattles &#8212; Then Rallies &#8212; Software Stocks</h2><p><strong>What:</strong> Anthropic hosted an enterprise agents event in New York that initially spooked software investors, then calmed them. The company showcased Claude Cowork integrations across finance, legal, HR, and engineering &#8212; but emphasized that Claude needs data from existing software vendors to be useful. Software stocks that had been hammered 25-30% in 2026 rallied on the news.</p><p><strong>So What:</strong> Wall Street analysts from Deutsche Bank, Jefferies, and William Blair reached the same conclusion: Anthropic is positioning itself as an &#8220;intelligence infrastructure&#8221; layer on top of existing enterprise software, not a replacement for it. The &#8220;SaaSpocalypse&#8221; narrative may be overdone &#8212; model providers need the data and workflows that incumbents control.</p><p><strong>Now What:</strong> If your team has been waiting out the AI-disruption panic before making software purchasing decisions, this is a signal to reengage. The winning enterprise stack will likely be incumbents plus AI orchestration, not one replacing the other.</p><p><a href="https://www.investors.com/news/technology/software-stock-nemesis-anthropic-enterprise-market-event-news/">Read more</a></p><h2>OpenAI Partners with BCG, McKinsey, Accenture, and Capgemini to Deploy Enterprise Agents</h2><p><strong>What:</strong> OpenAI announced &#8220;Frontier Alliances&#8221; &#8212; multi-year partnerships with BCG, McKinsey, Accenture, and Capgemini to help enterprises deploy AI agents at scale through its Frontier platform. Each firm is building dedicated practice groups certified on OpenAI technology with access to product and research teams.</p><p><strong>So What:</strong> OpenAI is publicly acknowledging that model intelligence isn&#8217;t the bottleneck &#8212; implementation is. By enlisting the Big Four consulting firms, they&#8217;re conceding that enterprise AI adoption requires strategy, change management, workflow redesign, and systems integration that a model provider alone can&#8217;t deliver.</p><p><strong>Now What:</strong> Enterprise leaders should watch which consulting partners develop genuine AI deployment capability versus those just rebranding existing practices. The firms that invest in certified technical teams will separate from those selling AI strategy decks.</p><p><a href="https://openai.com/index/frontier-alliance-partners/">Read more</a></p><h2>OpenAI Ships a Product with Zero Manually-Written Code</h2><p><strong>What:</strong> OpenAI published &#8220;Harness Engineering&#8221; &#8212; a detailed account of building and shipping an internal product with zero lines of human-written code. Using Codex agents, a team of three engineers produced roughly a million lines of code across 1,500 merged PRs in five months, averaging 3.5 PRs per engineer per day.</p><p><strong>So What:</strong> This isn&#8217;t a demo &#8212; it&#8217;s a production product with daily internal users. The most revealing insight: their bottleneck shifted from writing code to building &#8220;scaffolding&#8221; &#8212; the docs, linters, architectural constraints, and feedback loops that let agents do reliable work. The engineer&#8217;s job became designing environments, not writing implementations.</p><p><strong>Now What:</strong> Start treating your AGENTS.md, CI configuration, and architectural documentation as first-class engineering artifacts. In an agent-heavy workflow, the quality of your scaffolding determines the quality of your output.</p><p><a href="https://openai.com/index/harness-engineering/">Read more</a></p><h2>Claude Code Security Finds 500+ Bugs That Humans Missed</h2><p><strong>What:</strong> Anthropic launched Claude Code Security, an AI vulnerability scanner that reasons about codebases like a human security researcher rather than pattern-matching against known CVEs. Using Opus 4.6, it found over 500 bugs in production open-source code that had survived expert review. It&#8217;s in limited preview for Enterprise/Team customers; open-source maintainers get free access.</p><p><strong>So What:</strong> This is now a two-horse race with OpenAI&#8217;s Aardvark security agent (launched four months earlier). As AI-generated code proliferates, AI-powered security review is shifting from &#8220;nice to have&#8221; to &#8220;essential counterbalance.&#8221; The human-in-the-loop design &#8212; nothing gets patched without developer approval &#8212; is the right trust model for enterprise adoption.</p><p><strong>Now What:</strong> If your team ships AI-generated code, you need AI-powered security review in the pipeline. Evaluate both Claude Code Security and Aardvark against your actual codebase &#8212; the tool that catches bugs your team missed is the one worth adopting.</p><p><a href="https://www.anthropic.com/news/claude-code-security">Read more</a></p><h2>Every Publishes Editorial Guidelines &#8212; Written for AI Agents</h2><p><strong>What:</strong> Media company Every published editorial guidelines explicitly stating they write for both human readers and AI agents. Technical guides are &#8220;specifically optimized to serve as instructions for agents.&#8221; They also use a tool called Proof to track text provenance &#8212; which text is human-written versus AI-generated.</p><p><strong>So What:</strong> This is the first major media company to publicly declare &#8220;agent-readable&#8221; as a design goal alongside &#8220;human-readable.&#8221; Just as &#8220;mobile-friendly&#8221; became a content standard a decade ago, &#8220;agent-friendly&#8221; content may be next. The provenance tracking via Proof signals that transparency about AI authorship is becoming table stakes.</p><p><strong>Now What:</strong> Audit your own content &#8212; documentation, knowledge bases, SOPs &#8212; through an agent-readability lens. If AI agents will consume your content to take action on behalf of your customers or employees, structure and clarity matter more than ever.</p><p><a href="https://every.to/guides/editorial-guidelines">Read more</a></p><h2>Notion Ships Custom Agents That Run Autonomously Across Tools</h2><p><strong>What:</strong> Notion launched Custom Agents &#8212; autonomous AI teammates that operate continuously across Notion, Slack, email, calendar, Figma, and Linear. Setup is describe-and-trigger: the agent writes its own instructions and wires up its own tools. Early adopters include Ramp (300+ agents) and Remote (saved 20 hours/week replacing their IT help desk).</p><p><strong>So What:</strong> The &#8220;agents as teammates&#8221; framing is becoming the default product paradigm for productivity software. Notion&#8217;s approach &#8212; agents that monitor channels, capture requests, enrich data, and route information without human prompting &#8212; shows how AI features are evolving from &#8220;ask a question&#8221; to &#8220;run a workflow.&#8221;</p><p><strong>Now What:</strong> If your team uses Notion, start with one high-volume, low-risk workflow (FAQ routing, sprint reporting, request triage) and build a Custom Agent. The learning curve is in identifying which workflows benefit from always-on monitoring versus on-demand AI assistance.</p><p><a href="https://www.notion.com/en-gb/blog/introducing-custom-agents">Read more</a></p><h2>Pete Koomen: Most AI Apps Are &#8220;Horseless Carriages&#8221;</h2><p><strong>What:</strong> YC Partner Pete Koomen argues that most AI applications are failing because they mimic old software design patterns instead of rethinking around AI capabilities. His central example: Gmail&#8217;s AI draft feature produces generic, formal emails that take longer to prompt than to write manually &#8212; while a properly designed system prompt would let users teach the AI their voice once and reuse it forever.</p><p><strong>So What:</strong> The core insight is about who should write the system prompt. In traditional software, developers define behavior and users provide input. But when an AI agent acts on your behalf, you should be teaching it how to behave &#8212; not accepting a one-size-fits-all version designed by committee. &#8220;Most AI apps should be agent builders, not agents.&#8221;</p><p><strong>Now What:</strong> If you&#8217;re building or buying AI tools, ask this question: does the product let users customize the system prompt, or does it force a generic experience? The tools that let users teach the AI their specific context will win.</p><p><a href="https://koomen.dev/essays/horseless-carriages/">Read more</a></p><h2>Devin Ships Its Biggest Update Since Launch</h2><p><strong>What:</strong> Cognition released the largest update to Devin &#8212; the AI software engineering agent &#8212; since its initial launch. The update expands Devin&#8217;s ability to handle multi-file changes, longer-running tasks, and more complex codebases autonomously.</p><p><strong>So What:</strong> The AI coding agent space is now a genuine multi-player competition: Codex, Claude Code, Devin, and Cursor are all shipping major capability updates within weeks of each other. Karpathy&#8217;s observation about the pace of change (see below) isn&#8217;t hyperbole &#8212; the tooling landscape is shifting faster than most engineering teams can evaluate.</p><p><strong>Now What:</strong> If you evaluated Devin six months ago and passed, it&#8217;s time to re-benchmark. The competitive pressure between these tools is driving capability improvements at a pace where quarterly reevaluation is more appropriate than annual.</p><p><a href="https://x.com/ScottWu46/status/2026350958213787903">Read more</a></p><h2>Aaron Levie: Jevons Paradox Means More Demand for Engineering, Not Less</h2><p><strong>What:</strong> Box CEO Aaron Levie argues that lowering the cost of engineering through AI won&#8217;t reduce demand &#8212; it will increase it. Citing Jevons Paradox (when a resource becomes cheaper, total consumption increases), he makes the case that cheaper software creation means more software gets built, not fewer engineers get hired.</p><p><strong>So What:</strong> This directly challenges the &#8220;AI will replace developers&#8221; narrative. If Levie is right, enterprises should be planning for a world where AI dramatically increases the surface area of what gets built &#8212; requiring more engineering judgment, architecture, and oversight, even as the per-unit cost of code drops. The services firms that help enterprises navigate this expansion will be busier, not obsolete.</p><p><strong>Now What:</strong> Reframe your AI investment thesis: instead of &#8220;how many developers can we cut,&#8221; ask &#8220;what could we build if development cost 10x less?&#8221; The organizations that treat AI coding tools as expansion enablers rather than headcount reducers will capture disproportionate value.</p><p><a href="https://x.com/levie/status/2026885050411745491">Read more</a></p><h2>Karpathy: Programming Changed More in Two Months Than in Ten Years</h2><p><strong>What:</strong> Andrej Karpathy &#8212; former Tesla AI chief, OpenAI founding member &#8212; states that programming has changed more in the last two months than in the previous decade, driven by the rapid advancement of AI coding tools.</p><p><strong>So What:</strong> When someone with Karpathy&#8217;s credibility and vantage point makes this claim, it&#8217;s worth taking seriously. The pace of change in developer tooling &#8212; Codex, Claude Code, Devin, Cursor &#8212; is compressing what used to be years of incremental improvement into weeks. For non-technical leaders, this means the assumptions behind your 2026 engineering plans may already be outdated.</p><p><strong>Now What:</strong> If your engineering team hasn&#8217;t fundamentally revisited their tooling and workflow in the last 90 days, they&#8217;re falling behind. The gap between teams leveraging AI coding tools and those that aren&#8217;t is widening fast.</p><p><a href="https://x.com/karpathy/status/2026731645169185220">Read more</a></p>]]></content:encoded></item><item><title><![CDATA[The Real Shift Behind Enterprise Agents ]]></title><description><![CDATA[Feb 24, 2026: Anthropic launches even more enterprise agentic capabilities]]></description><link>https://tsw.blankmetal.ai/p/the-real-shift-behind-enterprise</link><guid isPermaLink="false">https://tsw.blankmetal.ai/p/the-real-shift-behind-enterprise</guid><dc:creator><![CDATA[Blank Metal]]></dc:creator><pubDate>Tue, 24 Feb 2026 14:48:05 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!kJ7A!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F026301b5-baae-4c2f-b56f-31c79b15367d_6016x4016.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!kJ7A!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F026301b5-baae-4c2f-b56f-31c79b15367d_6016x4016.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!kJ7A!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F026301b5-baae-4c2f-b56f-31c79b15367d_6016x4016.jpeg 424w, https://substackcdn.com/image/fetch/$s_!kJ7A!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F026301b5-baae-4c2f-b56f-31c79b15367d_6016x4016.jpeg 848w, https://substackcdn.com/image/fetch/$s_!kJ7A!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F026301b5-baae-4c2f-b56f-31c79b15367d_6016x4016.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!kJ7A!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F026301b5-baae-4c2f-b56f-31c79b15367d_6016x4016.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!kJ7A!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F026301b5-baae-4c2f-b56f-31c79b15367d_6016x4016.jpeg" width="1456" height="972" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/026301b5-baae-4c2f-b56f-31c79b15367d_6016x4016.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:972,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:716499,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://tsw.blankmetal.ai/i/188386304?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F026301b5-baae-4c2f-b56f-31c79b15367d_6016x4016.jpeg&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!kJ7A!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F026301b5-baae-4c2f-b56f-31c79b15367d_6016x4016.jpeg 424w, https://substackcdn.com/image/fetch/$s_!kJ7A!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F026301b5-baae-4c2f-b56f-31c79b15367d_6016x4016.jpeg 848w, https://substackcdn.com/image/fetch/$s_!kJ7A!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F026301b5-baae-4c2f-b56f-31c79b15367d_6016x4016.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!kJ7A!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F026301b5-baae-4c2f-b56f-31c79b15367d_6016x4016.jpeg 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Today, February 24, 2026, Anthropic <a href="https://claude.com/blog/cowork-plugins-across-enterprise">announces the launch</a> of even more enterprise agentic capabilities with Claude that will enable highly useful task performance across all knowledge work, especially in Sales, Legal, Finance, and Operations. Blank Metal is pleased to announce our participation in this launch as an implementation partner.</p><p><strong>What We Thought Agents Would Be</strong></p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://tsw.blankmetal.ai/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading The So What! Subscribe for free to receive new posts and support our work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p>For two years, the market has talked about AI agents like they were digital employees. Point-and-click builders. Autonomous bots you&#8217;d deploy to handle discrete workflows&#8212;one for contracts, one for CRM maintenance, one for financial reporting.</p><p>The mental model was delegation: identify a task, hand it to a bot, receive the output. Same shape as human delegation, just faster and cheaper.</p><blockquote><p>What we got is so much better.</p></blockquote><p><strong>What Anthropic Built Instead</strong></p><p>Anthropic&#8217;s enterprise agentic capabilities launch builds on <a href="https://claude.com/product/cowork?utm_source=google&amp;utm_medium=pse_knwl&amp;utm_campaign=acq_cowork_us&amp;utm_content=use_txt_v1&amp;utm_term=claude%20cowork">Claude Cowork</a>&#8212;but the architecture isn&#8217;t &#8220;build and deploy discrete bots.&#8221; Instead, it deconstructs and reimagines work itself.</p><p>The building blocks:</p><p><strong><a href="https://claude.com/blog/complete-guide-to-building-skills-for-claude">Skills</a></strong>: Discrete capabilities you teach Claude&#8212;checking SOWs for margin, applying brand style, researching prospects.</p><p><strong><a href="https://claude.com/connectors">Connectors</a></strong>: Links to your systems via MCP, giving Claude direct access to your enterprise context.</p><p><strong>Commands</strong>: Workflow shortcuts that bundle common operations.</p><p><strong>Subagents</strong>: For complex work, Claude delegates to other Claudes with specialized configurations.</p><p>These bundle into <strong><a href="https://claude.com/blog/cowork-plugins">Plugins</a></strong>&#8212;shareable packages that turn Claude into a domain specialist. Here&#8217;s what makes this architecture powerful: Anthropic is shipping foundational plugins for domains like Sales, Legal, Finance, and Operations. These aren&#8217;t closed systems&#8212;they&#8217;re starting points. Each plugin establishes a baseline capability that elevates everyone&#8217;s Claude immediately. Your Sales team gets sophisticated pipeline analysis and deal coaching out of the gate. Your Finance team gets month-end reconciliation logic and variance analysis built in.</p><p>But then each person can extend those foundations with their own skills. Your Head of Sales can add your specific qualification criteria and competitive positioning. Your Controller can layer in your cost allocation rules and reporting templates. The published plugin gives everyone a sophisticated baseline. Your custom additions make it yours.</p><p>This is fundamentally different from building agents from scratch or buying point solutions. You&#8217;re not starting from zero, and you&#8217;re not locked into someone else&#8217;s complete vision. You&#8217;re building on a foundation that keeps getting better as Anthropic and the community contribute new capabilities.</p><p>The key insight: you&#8217;re not sharing agents that orchestrate workflows. You&#8217;re sharing the underlying skills and recipes that any agent can use.</p><p>One universal agent. Continuously uploadable capabilities. Not a workforce of specialized agents, but a single collaborator that keeps getting better at more things.</p><p>This requires a completely different relationship with the machine.</p><p><strong>The Capabilities Organizations Need to Develop </strong>Making this leap means building new organizational muscles:</p><p><strong>Capability decomposition</strong>. The old paradigm asks &#8220;which agent handles this task?&#8221; The new paradigm asks &#8220;what skills does this task require?&#8221; That means breaking work into teachable components&#8212;not &#8220;handle expense reports&#8221; but &#8220;verify receipt amounts against submitted totals &gt;&gt; apply company travel policy logic &gt;&gt; flag outliers for review &gt;&gt; format approvals for the finance system.&#8221; Many people have never articulated their work at this level.</p><p><strong>Taxonomy building</strong>. You&#8217;re not just teaching Claude one skill. You&#8217;re building a structured map of what your organization actually does&#8212;the underlying capabilities that combine into the workflows you run every day. This becomes an organizational asset that compounds over time.</p><p><strong>Real-time evaluation</strong>. With traditional agents, engineering teams could evaluate a shared agent&#8217;s outputs against expected outcomes. In the Claude Cowork model, each user runs their own configuration of capabilities. There&#8217;s no single &#8220;agent&#8221; to QA centrally. You become responsible for evaluating your own outputs&#8212;or for building skills that do evaluation for you.</p><p><strong>Non-linear delegation</strong>. When you delegate to a person, you&#8217;re renting a bundle of pre-existing capabilities. They know how to write emails, navigate systems, and apply judgment. With Enterprise Agents, you start with shared organizational context and connectors, but then you&#8217;re building capability bundles yourself&#8212;skill by skill, connector by connector. It&#8217;s not &#8220;hire someone who knows the job.&#8221; It&#8217;s &#8220;teach a universal intelligence your specific version of the job.&#8221;</p><p>The good news: Claude Cowork with access to your enterprise context can get you pretty far out of the gate. And these skills compound. Organizations that invest in shared context and capability decomposition now will have skill libraries that grow more valuable every month.</p><blockquote><p>It&#8217;s not automation. It&#8217;s capability architecture.</p></blockquote><p><strong>Why Blank Metal Is an Implementation Partner</strong></p><p>At Blank Metal, we&#8217;ve been living this transformation. Over the past year, Claude Code and similar AI coding assistants fundamentally changed how our engineering team works. We stopped thinking &#8220;write this code&#8221; and started thinking &#8220;solve this problem.&#8221; We built capability taxonomies, learned to decompose our workflows into teachable skills, and came out the other side thinking differently about what software development even is.</p><p>We recognize this pattern. Enterprise Agents is the same shift&#8212;now available to everyone, not just developers.</p><blockquote><p>Here&#8217;s what we know from experience: technical work isn&#8217;t the bottleneck. The hard work is helping teams shift from &#8220;automate this task&#8221; to &#8220;what capabilities does this task require?&#8221;</p></blockquote><p>The engagement pattern we&#8217;re seeing:</p><p>First, <strong>connector infrastructure</strong>&#8212;ensuring MCPs exist for internal systems so agents can access critical company knowledge.</p><p>Second, <strong>capability mapping</strong>&#8212;the difficult work of decomposing organizational processes into teachable skills.</p><p>Third, <strong>adoption enablement</strong>&#8212;helping people internalize a new mental model for human-AI collaboration.</p><p>This isn&#8217;t about building something and handing it off. It&#8217;s about changing how your organization thinks about work itself.</p><p><strong>The Moment of Choice</strong></p><p>Anthropics expanded enterprise agentic capabilities for Claude are live. The organizations that recognize this as a paradigm shift will build capability libraries that compound over time. The ones looking for discrete task automation will find plenty of options. They just won&#8217;t be the ones defining what enterprise AI looks like in two years.</p><p>The architecture is here. The question is whether your organization is ready to think about work differently.</p><blockquote><p>Ready to start mapping your capabilities? Let&#8217;s talk.</p></blockquote><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://tsw.blankmetal.ai/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading The So What! Subscribe for free to receive new posts and support our work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[Weekly Headlines: Issue #10]]></title><description><![CDATA[February 12 - February 19, 2026]]></description><link>https://tsw.blankmetal.ai/p/weekly-headlines-issue-10</link><guid isPermaLink="false">https://tsw.blankmetal.ai/p/weekly-headlines-issue-10</guid><dc:creator><![CDATA[Blank Metal]]></dc:creator><pubDate>Fri, 20 Feb 2026 14:03:18 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!OQ7k!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F68841b35-85be-4286-a61b-c53f60c4fe08_1200x670.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!OQ7k!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F68841b35-85be-4286-a61b-c53f60c4fe08_1200x670.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!OQ7k!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F68841b35-85be-4286-a61b-c53f60c4fe08_1200x670.png 424w, https://substackcdn.com/image/fetch/$s_!OQ7k!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F68841b35-85be-4286-a61b-c53f60c4fe08_1200x670.png 848w, https://substackcdn.com/image/fetch/$s_!OQ7k!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F68841b35-85be-4286-a61b-c53f60c4fe08_1200x670.png 1272w, https://substackcdn.com/image/fetch/$s_!OQ7k!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F68841b35-85be-4286-a61b-c53f60c4fe08_1200x670.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!OQ7k!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F68841b35-85be-4286-a61b-c53f60c4fe08_1200x670.png" width="1200" height="670" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/68841b35-85be-4286-a61b-c53f60c4fe08_1200x670.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:670,&quot;width&quot;:1200,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:1479807,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://tsw.blankmetal.ai/i/188539783?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F68841b35-85be-4286-a61b-c53f60c4fe08_1200x670.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!OQ7k!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F68841b35-85be-4286-a61b-c53f60c4fe08_1200x670.png 424w, https://substackcdn.com/image/fetch/$s_!OQ7k!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F68841b35-85be-4286-a61b-c53f60c4fe08_1200x670.png 848w, https://substackcdn.com/image/fetch/$s_!OQ7k!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F68841b35-85be-4286-a61b-c53f60c4fe08_1200x670.png 1272w, https://substackcdn.com/image/fetch/$s_!OQ7k!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F68841b35-85be-4286-a61b-c53f60c4fe08_1200x670.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Welcome to Blank Metal&#8217;s Weekly AI Headlines.</p><p>Each week, our team shares the AI stories that caught our attention&#8212;the articles, announcements, and insights we&#8217;re actually discussing internally. We curate the best of what we&#8217;re reading and add the context that matters: what happened, why it matters, and what to do about it.</p><p>Short, sharp, and focused on impact.</p><h2>NVIDIA Open-Sources Two-Way Voice Model for Real-Time Conversation</h2><p><strong>What:</strong> NVIDIA released an open-source voice model capable of simultaneous listening and speaking&#8212;mimicking natural human conversation dynamics rather than turn-based exchanges.</p><p><strong>So What:</strong> This removes a major friction point in voice AI applications; enterprises building customer service agents, copilots, or voice interfaces now have a free, production-ready foundation for more natural interactions.</p><p><strong>Now What:</strong> If you&#8217;re evaluating voice AI vendors, benchmark this against paid alternatives&#8212;open-source parity is accelerating faster than most procurement cycles assume.</p><p><a href="https://x.com/HuggingModels/status/2022995332058251548">Read more</a></p><h2>Vertical SaaS Founder Says LLMs Will Gut His Own Industry</h2><p><strong>What:</strong> A founder who built traditional vertical SaaS argues that LLMs are collapsing core software moats&#8212;proprietary UI, workflow complexity, data aggregation&#8212;into simple chat interfaces, reducing years of engineering to &#8220;one week of writing.&#8221;</p><p><strong>So What:</strong> If this 12-24 month disruption timeline holds, enterprise leaders buying or building vertical software need to reassess whether they&#8217;re investing in durable value or soon-to-be-commoditized features.</p><p><strong>Now What:</strong> Audit your current vertical software stack through this lens&#8212;which vendors are truly differentiated by domain expertise versus UI complexity that AI could flatten?</p><p><a href="https://x.com/nicbstme/status/2023501562480644501?s=20">Read more</a></p><h2>OpenAI Open-Sources GABRIEL for Automated Qualitative Research</h2><p><strong>What:</strong> OpenAI released an open-source Python toolkit that uses GPT to convert qualitative data like interviews, social media posts, and images into quantitative measurements at scale&#8212;replacing manual coding work.</p><p><strong>So What:</strong> Enterprises sitting on mountains of unstructured customer feedback, support transcripts, or internal surveys now have a legitimate pathway to extract structured insights without building custom pipelines or hiring research teams.</p><p><strong>Now What:</strong> If your org has qualitative data gathering dust, pilot GABRIEL on a contained dataset to see if it can surface insights your current analytics miss.</p><p><a href="https://openai.com/index/scaling-social-science-research/">Read more</a></p><h2>OpenAI Bets Codex&#8217;s Future on GUI, Not Terminal</h2><p><strong>What:</strong> In a new interview, OpenAI&#8217;s Codex team revealed 5x growth since January to over a million weekly users, shipped GPT-5.3 Codex alongside their fastest coding model &#8220;Spark,&#8221; and explained why they&#8217;re prioritizing graphical interfaces over terminal-based workflows.</p><p><strong>So What:</strong> The explicit contrast with Claude Code&#8217;s terminal-first approach signals a strategic fork in how major AI labs think enterprise developers want to interact with coding agents&#8212;and their emphasis on code review (not generation) as the next bottleneck suggests where tooling investments may shift.</p><p><strong>Now What:</strong> If you&#8217;re evaluating coding agents, test both paradigms with your actual workflows&#8212;the GUI vs. terminal split may matter more for adoption than underlying model capability.</p><p><a href="https://open.spotify.com/episode/6bVrjHG2evanjiXgM1UNDF?si=42d27d6525a94780">Read more</a></p><h2>OpenAI Acquires OpenClaw Creator to Boost Agent Push</h2><p><strong>What:</strong> Peter Steinberger, creator of OpenClaw, is joining OpenAI to work on agentic AI development.</p><p><strong>So What:</strong> OpenAI is aggressively recruiting founders with deep experience building developer tools and document processing&#8212;capabilities that matter for enterprise agents that need to read, manipulate, and act on business documents.</p><p><strong>Now What:</strong> Watch for OpenAI&#8217;s agent capabilities to improve around document handling, a common pain point in enterprise automation workflows.</p><p><a href="https://www.theverge.com/ai-artificial-intelligence/879623/openclaw-founder-peter-steinberger-joins-openai">Read more</a></p><h2>Sinofsky: AI-Native Companies Will Define the Next Era</h2><p><strong>What:</strong> Former Microsoft exec Steven Sinofsky argues that companies building their core products <em>with</em> AI&#8212;not just adding AI features&#8212;will become the platform leaders of this generation, comparable to how Microsoft owned Windows, Google owned web, and Facebook/Uber owned mobile.</p><p><strong>So What:</strong> This framing challenges enterprises to honestly assess whether they&#8217;re treating AI as a feature bolt-on or a foundational capability&#8212;a distinction that may determine who leads and who follows in the next decade.</p><p><strong>Now What:</strong> Audit where AI sits in your org: is it enhancing existing workflows, or fundamentally reshaping how your core product gets built and delivered?</p><p><a href="https://x.com/stevesi/status/2021701369640759601?s=20">Read more</a></p><h2>Perplexity&#8217;s Model Council Pits Three AI Giants Against Each Other</h2><p><strong>What:</strong> Perplexity now runs queries across Claude, GPT, and Gemini simultaneously, then uses a fourth model to synthesize where they agree, disagree, and what each uniquely contributes.</p><p><strong>So What:</strong> The feature itself is basic, but it validates a strategic bet: as model performance varies by task, the real value shifts to the orchestration layer&#8212;knowing which model to use when and how to reconcile conflicting outputs.</p><p><strong>Now What:</strong> If you&#8217;re building AI applications, start thinking about multi-model routing and synthesis as a core capability, not an edge case.</p><p><a href="https://www.perplexity.ai/hub/use-cases/model-council-strategic-analysis">Read more</a></p><h2>Former GitHub CEO Raises $60M to Reimagine Developer Tools for AI Agents</h2><p><strong>What:</strong> Nat Friedman&#8217;s new startup Entire has raised $60M to build a developer platform designed from the ground up for AI agents, not human coders.</p><p><strong>So What:</strong> This is a serious signal that foundational dev infrastructure may need rebuilding&#8212;GitHub, built for human collaboration, may not be optimized for how AI agents read, write, and manage code at scale.</p><p><strong>Now What:</strong> Engineering leaders should start asking whether their current toolchains will bottleneck agent-assisted development as adoption accelerates.</p><p><a href="https://entire.io/blog/hello-entire-world/">Read more</a></p><h2>Box CEO Calls for New Agent Identity Standards</h2><p><strong>What:</strong> Aaron Levie argues that AI agents need their own distinct identities within enterprise platforms, requiring a fundamental rethink of authentication and authorization frameworks.</p><p><strong>So What:</strong> As agents increasingly act on behalf of employees&#8212;accessing systems, making decisions, moving data&#8212;current identity models built for humans won&#8217;t cut it, creating both security gaps and audit nightmares.</p><p><strong>Now What:</strong> Start mapping which systems your AI tools access today and whether your IAM framework can distinguish between human and agent actions.</p><p><a href="https://x.com/levie/status/2024335500283420836">Read more</a></p><h2>Figma and Anthropic Bridge AI Code to Visual Design</h2><p><strong>What:</strong> Figma&#8217;s new Code to Canvas feature lets designers import Claude Code output directly into Figma as editable design components.</p><p><strong>So What:</strong> This closes a critical gap in AI-assisted product development&#8212;code generated by AI can now flow back into design tools, potentially accelerating the prototype-to-production loop for teams using both platforms.</p><p><strong>Now What:</strong> If your product team spans design and engineering, explore whether this integration could reduce handoff friction in your current workflow.</p><p><a href="https://x.com/Techmeme/status/2023803589052035260">Read more</a></p>]]></content:encoded></item><item><title><![CDATA[Weekly Headlines: Issue #9]]></title><description><![CDATA[February 06 - February 12, 2026]]></description><link>https://tsw.blankmetal.ai/p/weekly-headlines-issue-9</link><guid isPermaLink="false">https://tsw.blankmetal.ai/p/weekly-headlines-issue-9</guid><dc:creator><![CDATA[Blank Metal]]></dc:creator><pubDate>Fri, 13 Feb 2026 15:02:54 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!FiTo!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fae9c9dd1-5836-48f5-9fd7-beae08afce68_1200x670.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!FiTo!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fae9c9dd1-5836-48f5-9fd7-beae08afce68_1200x670.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!FiTo!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fae9c9dd1-5836-48f5-9fd7-beae08afce68_1200x670.png 424w, https://substackcdn.com/image/fetch/$s_!FiTo!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fae9c9dd1-5836-48f5-9fd7-beae08afce68_1200x670.png 848w, https://substackcdn.com/image/fetch/$s_!FiTo!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fae9c9dd1-5836-48f5-9fd7-beae08afce68_1200x670.png 1272w, https://substackcdn.com/image/fetch/$s_!FiTo!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fae9c9dd1-5836-48f5-9fd7-beae08afce68_1200x670.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!FiTo!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fae9c9dd1-5836-48f5-9fd7-beae08afce68_1200x670.png" width="1200" height="670" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/ae9c9dd1-5836-48f5-9fd7-beae08afce68_1200x670.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:670,&quot;width&quot;:1200,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:1480170,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://tsw.blankmetal.ai/i/187792201?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fae9c9dd1-5836-48f5-9fd7-beae08afce68_1200x670.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!FiTo!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fae9c9dd1-5836-48f5-9fd7-beae08afce68_1200x670.png 424w, https://substackcdn.com/image/fetch/$s_!FiTo!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fae9c9dd1-5836-48f5-9fd7-beae08afce68_1200x670.png 848w, https://substackcdn.com/image/fetch/$s_!FiTo!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fae9c9dd1-5836-48f5-9fd7-beae08afce68_1200x670.png 1272w, https://substackcdn.com/image/fetch/$s_!FiTo!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fae9c9dd1-5836-48f5-9fd7-beae08afce68_1200x670.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Welcome to Blank Metal&#8217;s Weekly AI Headlines.</p><p>Each week, our team shares the AI stories that caught our attention&#8212;the articles, announcements, and insights we&#8217;re actually discussing internally. We curate the best of what we&#8217;re reading and add the context that matters: what happened, why it matters, and what to do about it.</p><p>Short, sharp, and focused on impact.</p><h1>Agent Infrastructure &amp; Governance</h1><p><em>The bottleneck isn&#8217;t building agents &#8212; it&#8217;s running them reliably, safely, and at scale.</em></p><h2>Former GitHub CEO Raises $60M to Manage AI Agent Fleets</h2><p><strong>What:</strong> Thomas Dohmke launched Entire, a dev platform designed to track and govern code produced by AI agents, starting with an open-source CLI that captures the full reasoning context behind AI-generated commits.</p><p><strong>So What:</strong> This validates what many teams are discovering firsthand&#8212;the real bottleneck isn&#8217;t generating code with AI, it&#8217;s reviewing and governing what actually ships. Existing Git workflows weren&#8217;t built for machine-speed output.</p><p><strong>Now What:</strong> If your engineering org is scaling AI coding tools, start auditing where human review is already becoming the constraint&#8212;that&#8217;s likely where you&#8217;ll need new tooling or processes first.</p><p><a href="https://entire.io/blog/hello-entire-world">Read more</a></p><p><em> </em></p><h2>Warp Bets Agent Orchestration Is the Real Enterprise Bottleneck</h2><p><strong>What:</strong> Warp launched Oz, cloud infrastructure for scheduling, governing, and running coding agents at scale&#8212;complete with cron triggers, sandboxed environments, and audit trails. The platform already writes 60% of Warp&#8217;s own PRs.</p><p><strong>So What:</strong> The hard part isn&#8217;t getting agents to work. It&#8217;s getting them to work reliably, safely, and repeatedly without human babysitting. Warp is betting that orchestration&#8212;not the agents themselves&#8212;is where the real enterprise value sits.</p><p><strong>Now What:</strong> If you&#8217;re running agents in production (or planning to), audit your current orchestration stack. The gap between &#8220;demo-ready&#8221; and &#8220;enterprise-ready&#8221; is exactly where tools like this aim to live.</p><p><a href="https://www.warp.dev/oz">Read more</a></p><p></p><h2>Claude Cowork Comes to Windows&#8212;Leveling the AI Desktop Playing Field</h2><p><strong>What:</strong> Anthropic shipped Claude Cowork for Windows, bringing the same AI desktop assistant that&#8217;s been a big unlock for Mac users to the PC ecosystem.</p><p><strong>So What:</strong> Mac users are used to having first access to tools, while PC users have been largely limited to Microsoft-supported options. This matters in enterprise: most corporate desktops are Windows. Getting AI that feels like a real collaborator&#8212;not just a chat window&#8212;onto PCs opens the door for millions of knowledge workers who&#8217;ve been watching from the sideline.</p><p><strong>Now What:</strong> If your org has been waiting for AI desktop tools that aren&#8217;t locked into the Microsoft ecosystem, this is worth a pilot. The &#8220;pick a folder&#8221; simplicity may move faster than a Copilot rollout stuck in security review.</p><p><a href="https://x.com/claudeai/status/2021336313979625910?s=20">Read more</a></p><p></p><h1>The SaaS Reckoning</h1><p><em>SaaS isn&#8217;t dead &#8212; but the business model that sustained it is under structural pressure.</em></p><h2>The Big 4 Consulting Unbundling Has Started</h2><p><strong>What:</strong> Bitwise CEO Hunter Horsley draws a parallel between the Craigslist unbundling of 2006 and what&#8217;s happening to professional services firms like PwC&#8212;every service line on their website is work that agentic systems can now do faster and cheaper.</p><p><strong>So What:</strong> The difference from 2006: enterprises don&#8217;t have to wait for a startup to build the disruption and hope M&amp;A works out. They can build the agentic version themselves, now. The path is clearer&#8212;hire a team, build the capability, own the asset.</p><p><strong>Now What:</strong> Most enterprises know they need to move. They&#8217;re just stuck on where to start. Identify one consulting-heavy workflow and scope what the agentic version looks like.</p><p><a href="https://x.com/HHorsley/status/2021486174767096091?s=20">Read more</a></p><p></p><h2>Ben Thompson: The SaaS Wall Is Structural, Not Cyclical</h2><p><strong>What:</strong> Ben Thompson argues the SaaS downturn isn&#8217;t a dip&#8212;it&#8217;s a permanent shift from growth companies to stable businesses. Seat-based pricing breaks when headcount stagnates or shrinks. Systems of record remain defensible, but discretionary tools face disruption from AI-native alternatives that do the same job without the per-seat tax.</p><p><strong>So What:</strong> This is the distinction enterprise buyers need to internalize: your CRM and ERP aren&#8217;t going anywhere, but the layer of tools around them&#8212;the ones your teams adopted during the growth era&#8212;are vulnerable. When agents can perform tasks across systems, the &#8220;good enough&#8221; SaaS tool that lives on inertia loses its moat overnight.</p><p><strong>Now What:</strong> Audit your software stack in two buckets: systems of record (defensible, keep) and discretionary tools (exposed, renegotiate or replace). Your leverage as a buyer has never been higher.</p><p><a href="https://share.transistor.fm/s/25f9c622">Listen here</a></p><p></p><h2>a16z&#8217;s Anish Acharya: The &#8220;SaaS Apocalypse&#8221; Is a Myth&#8212;But the Moats Are Changing</h2><p><strong>What:</strong> a16z general partner Anish Acharya calls the &#8220;SaaS is dead&#8221; narrative overblown, but argues the real shift is significant: AI agents are breaking the lock-in legacy software relied on. Meanwhile, consumers are happily paying $200+/month for tools like Claude and Grok&#8212;not because they&#8217;re for everyone, but because they&#8217;re 100x better for someone. He also frames the dev tools market (Cursor vs. Claude Code) as looking more like Cloud than Uber vs. Lyft.</p><p><strong>So What:</strong> Two things to watch: (1) SaaS as a delivery model survives, but SaaS as a moat erodes when agents can move data between systems and perform tasks across tools. Switching costs are dropping. (2) The willingness to pay $200+/month for AI tools that actually work signals that the market is bifurcating&#8212;power users will pay dramatically more for dramatically better tools, while commodity features race to zero.</p><p><strong>Now What:</strong> If you&#8217;re evaluating enterprise software, the new buying criteria isn&#8217;t &#8220;what does this tool do?&#8221; It&#8217;s &#8220;how well does this tool work with agents?&#8221; And if you&#8217;re selling software, watch your per-seat pricing&#8212;the market is moving toward value-based models fast.</p><p><a href="https://www.thetwentyminutevc.com/anish-acharya">Listen here</a></p><p></p><h1>Models &amp; Code Abundance</h1><p><em>Model capabilities are commoditizing fast &#8212; the strategic question is shifting from &#8220;which model?&#8221; to &#8220;what do you build on top?&#8221;</em></p><h2>Six Major AI Releases in a Single Day &#8212; The Pace Is the Headline</h2><p><strong>What:</strong> February 12 saw six major AI releases hit simultaneously: <a href="https://openai.com/index/introducing-gpt-5-3-codex-spark/">OpenAI shipped GPT-5.3-Codex-Spark</a> on Cerebras hardware (1,000+ tokens/sec for real-time coding), <a href="https://blog.google/innovation-and-ai/models-and-research/gemini-models/gemini-3-deep-think/">Google launched Gemini 3 Deep Think</a> (new #1 on math/science benchmarks), <a href="https://officechai.com/ai/chinas-minimax-releases-m2-5-beats-gemini-3-pro-and-gpt-5-2-on-swe-bench/">MiniMax dropped M2.5</a> at 96% cheaper than competitors, <a href="https://www.reuters.com/business/media-telecom/bytedances-new-ai-video-model-goes-viral-china-looks-second-deepseek-moment-2026-02-12/">ByteDance&#8217;s Seedance 2.0</a> video model went viral in China, <a href="https://www.cnbc.com/2026/02/12/chinese-ai-stocks-new-model-and-agent-releases-zhipu-minimax.html">Zhipu hiked prices 30%</a>, and <a href="https://www.techrepublic.com/article/news-amazon-engineers-revolt-over-ai-tool-restrictions/">Amazon engineers revolted internally</a>&#8212;choosing Claude Code over Amazon&#8217;s own Kiro.</p><p><strong>So What:</strong> No single release here is the story. The story is that six shipped on the same Tuesday and nobody blinked. Model capabilities are commoditizing so fast that &#8220;best model&#8221; rotates weekly. The strategic question is shifting from &#8220;which model is best?&#8221; to &#8220;which infrastructure lets you swap models without rebuilding?&#8221;</p><p><strong>Now What:</strong> If your AI strategy is built around a single model provider, the lock-in risk isn&#8217;t going away&#8212;it&#8217;s inverting. The moat is in your orchestration layer and data, not the model underneath.</p><p></p><h2>Scott Belsky: Exponential Code Won&#8217;t Kill SaaS&#8212;It&#8217;ll Reshape Who Wins</h2><p><strong>What:</strong> Adobe CPO Scott Belsky argues that AI-generated code abundance won&#8217;t destroy enterprise software&#8212;it will make foundational infrastructure (security, data graphs, shared memory) more valuable, while &#8220;private-equity-owned niche clunkware&#8221; gets disrupted.</p><p><strong>So What:</strong> Three big implications: (1) &#8220;Disposable software&#8221;&#8212;temporary, single-use apps&#8212;will proliferate, creating new security surface area. (2) Per-seat pricing is dead; usage-based and outcome-based models are coming. (3) The apprenticeship pipeline breaks when AI automates entry-level tasks, and companies need to deliberately rebuild knowledge transfer.</p><p><strong>Now What:</strong> The apprenticeship point is the sleeper insight. If AI handles the grunt work that used to train junior people, who&#8217;s building the next generation of senior talent? Every enterprise needs an answer to this.</p><p><a href="https://www.implications.com/p/exponential-code-network-effects">Read more</a></p><p></p><h1>The Narrative vs. The Reality</h1><p><em>The hype says everything is about to change. The data says the people who already changed are breaking.</em></p><h2>Matt Shumer&#8217;s &#8220;Something Big Is Happening&#8221; Goes Mainstream</h2><p><strong>What:</strong> AI startup founder Matt Shumer&#8217;s open letter comparing AI&#8217;s current moment to February 2020 Covid went viral outside the tech bubble&#8212;mainstream media picked it up and non-technical audiences are now reading it.</p><p><strong>So What:</strong> The capability claims are real. But the fear framing and the Covid analogy are doing all the heavy lifting. Covid happened <em>to</em> people&#8212;a pathogen hitting zero immunity. AI is happening <em>for</em> people to build with. Better analogy: the internet in 1998. Clearly going to change everything. Unclear exactly how. The people who leaned in early did fine.</p><p><strong>Now What:</strong> When clients forward this (and they will), don&#8217;t amplify the fear or dismiss it. Translate it: which of your workflows has AI already outpaced current tools, and which are 18 months out? That&#8217;s the useful conversation.</p><p><a href="https://shumer.dev/something-big-is-happening">Read more</a></p><p></p><h2>The First Signs of AI Burnout Are Hitting the Early Adopters</h2><p><strong>What:</strong> A Berkeley Haas study of 200 employees over 9 months found that AI doesn&#8217;t reduce work&#8212;it intensifies it. Workers managed more parallel threads, checked AI outputs constantly, and revived long-deferred tasks, creating cognitive overload disguised as productivity.</p><p><strong>So What:</strong> The study&#8217;s warning: organizations can&#8217;t distinguish genuine productivity gains from unsustainable intensity. People are losing sleep because &#8220;just one more prompt&#8221; is irresistible. Work bleeds into lunches and late evenings not because of deadlines, but because AI makes it feel like you <em>could</em> do more.</p><p><strong>Now What:</strong> This is the contrarian signal in a week full of AI optimism. If your teams are adopting AI aggressively, check in on sustainability&#8212;not just output. The most engaged users may be the ones burning out fastest.</p><p><a href="https://simonwillison.net/2026/Feb/9/ai-intensifies-work/">Read more</a></p><p></p>]]></content:encoded></item><item><title><![CDATA[Stop Demoing AI. Start Building With It.]]></title><description><![CDATA[Blank Metal offers Claude Code Workshops to enterprise organizations]]></description><link>https://tsw.blankmetal.ai/p/stop-demoing-ai-start-building-with</link><guid isPermaLink="false">https://tsw.blankmetal.ai/p/stop-demoing-ai-start-building-with</guid><dc:creator><![CDATA[Blank Metal]]></dc:creator><pubDate>Wed, 11 Feb 2026 21:27:38 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!dANY!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F45407a61-8e7a-45da-8ac1-484578ccdc0a_1830x956.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!dANY!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F45407a61-8e7a-45da-8ac1-484578ccdc0a_1830x956.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!dANY!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F45407a61-8e7a-45da-8ac1-484578ccdc0a_1830x956.png 424w, https://substackcdn.com/image/fetch/$s_!dANY!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F45407a61-8e7a-45da-8ac1-484578ccdc0a_1830x956.png 848w, https://substackcdn.com/image/fetch/$s_!dANY!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F45407a61-8e7a-45da-8ac1-484578ccdc0a_1830x956.png 1272w, https://substackcdn.com/image/fetch/$s_!dANY!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F45407a61-8e7a-45da-8ac1-484578ccdc0a_1830x956.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!dANY!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F45407a61-8e7a-45da-8ac1-484578ccdc0a_1830x956.png" width="1456" height="761" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/45407a61-8e7a-45da-8ac1-484578ccdc0a_1830x956.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:761,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:3123503,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://tsw.blankmetal.ai/i/187677766?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F45407a61-8e7a-45da-8ac1-484578ccdc0a_1830x956.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!dANY!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F45407a61-8e7a-45da-8ac1-484578ccdc0a_1830x956.png 424w, https://substackcdn.com/image/fetch/$s_!dANY!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F45407a61-8e7a-45da-8ac1-484578ccdc0a_1830x956.png 848w, https://substackcdn.com/image/fetch/$s_!dANY!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F45407a61-8e7a-45da-8ac1-484578ccdc0a_1830x956.png 1272w, https://substackcdn.com/image/fetch/$s_!dANY!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F45407a61-8e7a-45da-8ac1-484578ccdc0a_1830x956.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><h2><strong>We&#8217;re running Claude Code Labs with Anthropic &#8212; and capacity is limited.</strong></h2><p>Here&#8217;s what happens at so many enterprise &#8220;AI workshops&#8221;. . .  someone presents slides about what AI <em>could</em> do. Maybe there&#8217;s a live demo where an engineer types a prompt and everyone nods. People leave feeling inspired and do absolutely nothing different on the next day.</p><p>We&#8217;ve watched this pattern enough times to know it doesn&#8217;t work. Inspiration without execution is just. . . time spent.</p><h2><strong>What Claude Code Labs actually are</strong></h2><p>Claude Code Labs are in-person workshops where enterprise teams spend three hours <em>building with Claude Code</em> &#8212; Anthropic&#8217;s command-line tool for agentic coding. Not just watching.</p><p>Blank Metal runs these in partnership with Anthropic&#8217;s Applied AI team. The format is simple: you bring your laptop, we bring the curriculum. By the end of the session, every person in the room has created real code with Claude Code &#8212; on their own machine, against real problems, with support from people who do this for a living.</p><h2><strong>Who these are for</strong></h2><p>These workshops are for technical teams and practitioners at enterprise organizations who want to move past the &#8220;should we use AI?&#8221; conversation and into the &#8220;how do we actually adopt this?&#8221; phase.</p><p>The people who get the most out of it tend to have an input into their company&#8217;s AI tooling adoption. They&#8217;re architects and engineers who are tired of evaluating tools through blog posts and vendor demos and want to feel what it&#8217;s actually like to work with Claude Code against real problems. We also encourage technical Product Managers to attend - it could change their whole approach to work.</p><p>To attend, you need basic command line familiarity, a laptop you can install the software on (that can access the internet), and a Claude Code console account.</p><h2><strong>Why we do this</strong></h2><p>We&#8217;ve been helping enterprise organizations build with AI &#8212; not as a concept, but as deployed, running software. The most common pattern we see is misalignment: leadership has bought into AI, but engineers still don&#8217;t know what to do with it day-to-day. Or they haven&#8217;t had time to really dig in and focus on learning it.</p><p>That gap doesn&#8217;t close with a webinar. It closes when people sit down and build something.</p><p>Claude Code Labs compresses that learning curve into a structured afternoon: 90-minute setup, three hours of building, and an optional hour to keep experimenting.</p><p>30 to 50 people per session. Small enough that nobody hides in the back. Large enough to make it worth your organization&#8217;s time to host.</p><h2><strong>Limited capacity </strong></h2><p>We&#8217;re running a limited number of these in partnership with Anthropic. Each one requires coordination between our team, Anthropic&#8217;s Applied AI group, and the host organization. That means limited slots, and demand fills the calendar fast.</p><p>If your organization is serious about adopting AI coding tools at scale &#8212; not as a pilot, not as a &#8220;let&#8217;s see&#8221; experiment, but as part of how your engineering team actually works &#8212; these labs are designed for you.</p><h2><strong>Get on the list!</strong></h2><p>We&#8217;re booking Claude Code Labs now for enterprise organizations. If you want to host one for your team, <strong><a href="https://41exaw.share-na2.hsforms.com/2fL1RFeACT6eO1opts3DHLg">reach out to us here</a></strong>. We&#8217;ll tell you if it&#8217;s a fit and when the next available session is.</p><p></p>]]></content:encoded></item><item><title><![CDATA[AI Real Talk Event]]></title><description><![CDATA[April 9, 2026 in Minneapolis | 4 pm - 7 pm]]></description><link>https://tsw.blankmetal.ai/p/ai-real-talk-event</link><guid isPermaLink="false">https://tsw.blankmetal.ai/p/ai-real-talk-event</guid><dc:creator><![CDATA[Blank Metal]]></dc:creator><pubDate>Fri, 06 Feb 2026 19:24:02 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!EvG3!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F38ad0185-587a-4cc2-90a4-8508583d9140_2048x1463.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!EvG3!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F38ad0185-587a-4cc2-90a4-8508583d9140_2048x1463.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!EvG3!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F38ad0185-587a-4cc2-90a4-8508583d9140_2048x1463.jpeg 424w, https://substackcdn.com/image/fetch/$s_!EvG3!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F38ad0185-587a-4cc2-90a4-8508583d9140_2048x1463.jpeg 848w, https://substackcdn.com/image/fetch/$s_!EvG3!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F38ad0185-587a-4cc2-90a4-8508583d9140_2048x1463.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!EvG3!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F38ad0185-587a-4cc2-90a4-8508583d9140_2048x1463.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!EvG3!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F38ad0185-587a-4cc2-90a4-8508583d9140_2048x1463.jpeg" width="1456" height="1040" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/38ad0185-587a-4cc2-90a4-8508583d9140_2048x1463.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1040,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:439918,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://tsw.blankmetal.ai/i/187113885?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F38ad0185-587a-4cc2-90a4-8508583d9140_2048x1463.jpeg&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!EvG3!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F38ad0185-587a-4cc2-90a4-8508583d9140_2048x1463.jpeg 424w, https://substackcdn.com/image/fetch/$s_!EvG3!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F38ad0185-587a-4cc2-90a4-8508583d9140_2048x1463.jpeg 848w, https://substackcdn.com/image/fetch/$s_!EvG3!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F38ad0185-587a-4cc2-90a4-8508583d9140_2048x1463.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!EvG3!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F38ad0185-587a-4cc2-90a4-8508583d9140_2048x1463.jpeg 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p><strong>AI Real Talk</strong> is a quarterly gathering for executives who are doing the hard work of bringing AI into their organizations &#8212; not just talking about it. They know that transforming their organization starts with honest conversation.</p><p>The format is simple: real challenges, real outcomes, and real conversation about what you haven&#8217;t figured out yet. Everything runs under <a href="https://www.chathamhouse.org/about-us/chatham-house-rule">Chatham House</a> rules so people can speak freely about what they&#8217;re working on.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!CgqW!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fac737f78-ddc9-4454-84cc-6d2e4d2b0855_560x560.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!CgqW!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fac737f78-ddc9-4454-84cc-6d2e4d2b0855_560x560.jpeg 424w, https://substackcdn.com/image/fetch/$s_!CgqW!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fac737f78-ddc9-4454-84cc-6d2e4d2b0855_560x560.jpeg 848w, https://substackcdn.com/image/fetch/$s_!CgqW!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fac737f78-ddc9-4454-84cc-6d2e4d2b0855_560x560.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!CgqW!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fac737f78-ddc9-4454-84cc-6d2e4d2b0855_560x560.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!CgqW!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fac737f78-ddc9-4454-84cc-6d2e4d2b0855_560x560.jpeg" width="282" height="282" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/ac737f78-ddc9-4454-84cc-6d2e4d2b0855_560x560.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:560,&quot;width&quot;:560,&quot;resizeWidth&quot;:282,&quot;bytes&quot;:54997,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://tsw.blankmetal.ai/i/187113885?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F184a1ac3-8d89-4ced-aa8c-76999eaee7a0_560x560.jpeg&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!CgqW!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fac737f78-ddc9-4454-84cc-6d2e4d2b0855_560x560.jpeg 424w, https://substackcdn.com/image/fetch/$s_!CgqW!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fac737f78-ddc9-4454-84cc-6d2e4d2b0855_560x560.jpeg 848w, https://substackcdn.com/image/fetch/$s_!CgqW!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fac737f78-ddc9-4454-84cc-6d2e4d2b0855_560x560.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!CgqW!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fac737f78-ddc9-4454-84cc-6d2e4d2b0855_560x560.jpeg 1456w" sizes="100vw"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>On April 9th, join Blank Metal and <strong>Anand Francis</strong> (Head of AI, US Bank) for a live fireside chat on how he&#8217;s driving real AI adoption inside a highly regulated enterprise&#8212;what&#8217;s working, what isn&#8217;t, and what he&#8217;s learning 90 days in.</p><p><strong>This event is invite-only. <br><a href="https://pages.blankmetal.ai/april-9-2026-rsvp">Sign up here to RSVP to the April 9th event</a></strong></p>]]></content:encoded></item><item><title><![CDATA[Weekly Headlines: Issue #8]]></title><description><![CDATA[January 29 - February 5, 2026]]></description><link>https://tsw.blankmetal.ai/p/weekly-headlines-issue-8</link><guid isPermaLink="false">https://tsw.blankmetal.ai/p/weekly-headlines-issue-8</guid><dc:creator><![CDATA[Blank Metal]]></dc:creator><pubDate>Fri, 06 Feb 2026 15:02:04 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!BZ1i!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2f6af09a-e860-4c03-8718-cd8c13c36b7e_1200x670.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!BZ1i!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2f6af09a-e860-4c03-8718-cd8c13c36b7e_1200x670.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!BZ1i!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2f6af09a-e860-4c03-8718-cd8c13c36b7e_1200x670.png 424w, https://substackcdn.com/image/fetch/$s_!BZ1i!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2f6af09a-e860-4c03-8718-cd8c13c36b7e_1200x670.png 848w, https://substackcdn.com/image/fetch/$s_!BZ1i!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2f6af09a-e860-4c03-8718-cd8c13c36b7e_1200x670.png 1272w, https://substackcdn.com/image/fetch/$s_!BZ1i!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2f6af09a-e860-4c03-8718-cd8c13c36b7e_1200x670.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!BZ1i!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2f6af09a-e860-4c03-8718-cd8c13c36b7e_1200x670.png" width="1200" height="670" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/2f6af09a-e860-4c03-8718-cd8c13c36b7e_1200x670.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:670,&quot;width&quot;:1200,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:1480096,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://tsw.blankmetal.ai/i/187046665?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2f6af09a-e860-4c03-8718-cd8c13c36b7e_1200x670.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!BZ1i!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2f6af09a-e860-4c03-8718-cd8c13c36b7e_1200x670.png 424w, https://substackcdn.com/image/fetch/$s_!BZ1i!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2f6af09a-e860-4c03-8718-cd8c13c36b7e_1200x670.png 848w, https://substackcdn.com/image/fetch/$s_!BZ1i!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2f6af09a-e860-4c03-8718-cd8c13c36b7e_1200x670.png 1272w, https://substackcdn.com/image/fetch/$s_!BZ1i!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2f6af09a-e860-4c03-8718-cd8c13c36b7e_1200x670.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><h2>Anthropic Launches Claude Opus 4.6 with Finance-First Features</h2><p><strong>What:</strong> Anthropic released Claude Opus 4.6, which now tops the Finance Agent benchmark at 60.7%&#8212;a 5.5% jump from Opus 4.5&#8212;and outperforms GPT-5.2 on knowledge work tasks in finance and legal.</p><p><strong>So What:</strong> This isn&#8217;t just another model bump. Opus 4.6 can combine regulatory filings, market reports, and internal data to produce analyses that would otherwise take analysts days. First-pass deliverables are now genuinely usable, not just rough drafts.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://tsw.blankmetal.ai/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading The So What! Subscribe for free to receive new posts and support our work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p><strong>Now What:</strong> If your finance or legal teams are still treating AI as a research assistant, it&#8217;s time to test it as a first-draft analyst. The &#8220;vibe working&#8221; era means reviewing AI output, not creating from scratch.</p><p><a href="https://www.zdnet.com/article/anthropic-claude-opus-4-6-first-try-work-deliverables/">Read more</a></p><p></p><h2>Alibaba Open-Sources Speech Models That Beat GPT-4o</h2><p><strong>What:</strong> Alibaba released Qwen3-ASR, a pair of open-source speech recognition models supporting 52 languages that match or outperform GPT-4o Transcribe and Whisper-large-v3, with the smaller version achieving 92ms latency.</p><p><strong>So What:</strong> Enterprise teams building voice interfaces, transcription pipelines, or multilingual support tools now have a high-performance open-source option that sidesteps API costs and vendor lock-in.</p><p><strong>Now What:</strong> If you&#8217;re paying per-minute for transcription APIs or building latency-sensitive voice features, benchmark Qwen3-ASR against your current stack&#8212;the cost and control benefits could be substantial.</p><p><a href="https://link.mail.beehiiv.com/ss/c/u001.VUuH6R6zI0G5BkbXvz91_GAyPOWiE-on8J799p4fhR76Qqf_kdor_uXjef0Uq8JOBxphMrkCbqX5IbjGqnqErZ691EyF0WAJumPKYvWpxqN7-0qRzSo3EucBUzDJGYABWKITU0bEl92eqJtSwmTjshGK_Mvbp-9BwtmeNmRskgkuYDlfqX1mUn8-w_X6pOHFUmv3YRDd092TttoRBB0k67ZoJe-VXXCh9vrdzNhwpXceZ8zcLT3o_yy7m4i-R6U-RrHy-_fJ-BbQ0lYJhbiA9H2qDyfnVeV4geljiCpS7ewyx-KtP008IoGshITgXZ0mo6RASkNTaakA85CQFz6FzA/4nq/o2WIiBxrSn2Q0euIovxe-g/h3/h001.m1yAvkcCJdf3jzfTJkGr1ymU4WtCLiIJilqO9gDCQME">Read more</a></p><p></p><h2>OpenAI Codex Mac App Now Free to Try</h2><p><strong>What:</strong> OpenAI released a native Mac desktop app for Codex, its AI coding assistant, with free trial access for ChatGPT Plus subscribers.</p><p><strong>So What:</strong> This signals OpenAI&#8217;s push to embed AI coding tools directly into developer workflows&#8212;enterprise teams evaluating coding assistants now have another serious contender alongside GitHub Copilot and Claude.</p><p><strong>Now What:</strong> If your engineering team is already paying for ChatGPT Plus, have a few developers test Codex against your current tooling to see if consolidation makes sense.</p><p><a href="https://www.zdnet.com/article/openai-codex-mac-app-free-trial/">Read more</a></p><p></p><h2>Codex vs. Opus Showdown Reveals the &#8220;Ur-Coding Model&#8221; Race</h2><p><strong>What:</strong> Every&#8217;s head-to-head comparison of GPT-5.3 Codex and Opus 4.6 found both models converging toward similar capabilities, with Opus excelling on complex, open-ended tasks while Codex delivers more consistent, reliable execution.</p><p><strong>So What:</strong> The finding that matters isn&#8217;t which model won&#8212;it&#8217;s the thesis that great coding agents become great <em>general</em> work agents, meaning AI coding infrastructure may be foundational business infrastructure, not just a dev tools expense.</p><p><strong>Now What:</strong> If you&#8217;re running multiple AI models in production, consider formalizing a model selection framework that matches task complexity to model strengths rather than defaulting to one provider.</p><p><a href="https://every.to/vibe-check/codex-vs-opus">Read more</a></p><p></p><h2>Apple Brings Agentic Coding to Xcode 26.3</h2><p><strong>What:</strong> Apple&#8217;s latest Xcode update introduces agentic AI capabilities that can autonomously write, debug, and refactor code within its native development environment.</p><p><strong>So What:</strong> This signals Apple&#8217;s serious entry into AI-assisted development tooling&#8212;enterprise teams building iOS/macOS apps now have a first-party option competing with Copilot and Cursor, potentially tightening Apple&#8217;s ecosystem lock-in further.</p><p><strong>Now What:</strong> If your org ships Apple platform apps, evaluate whether this native integration outweighs your current third-party coding assistant&#8212;ecosystem alignment often wins on friction alone.</p><p><a href="https://www.apple.com/newsroom/2026/02/xcode-26-point-3-unlocks-the-power-of-agentic-coding/?cid=ADC-DM-c00377-M00827">Read more</a></p><p></p><h2>OpenAI Retires GPT-4o as It Doubles Down on GPT-5.2</h2><p><strong>What:</strong> Starting February 13th, ChatGPT users will lose access to GPT-4o, GPT-4.1, GPT-4.1 mini, and o4-mini&#8212;though API access remains unchanged for developers.</p><p><strong>So What:</strong> With only 0.1% of users still choosing GPT-4o daily, this signals OpenAI&#8217;s aggressive push to consolidate around newer models, reducing maintenance overhead while accelerating GPT-5.2 development.</p><p><strong>Now What:</strong> Audit any internal tools or workflows that reference specific model versions in ChatGPT (not API)&#8212;and use this as a reminder that model availability is never guaranteed.</p><p><a href="https://link.mail.beehiiv.com/ss/c/u001.wZN1XY49ssmkxVgHdsx183UNH4UTy36RoFEzZc0N1g-dEKSwCadxLET19dh0H1ErXlCv-p1IY1GL82wRCnXeJLSJ0JIFZSMLa-dkAgZsDkJi8tcY3bDjncg-uaj4wc1x2neTSbonmCRhIIhsnBxAniYgmCQeK5TgY9isWXm7j0kd9k85LLqwd_hesALQmtNZZh1Rr1VzpktziSs0Yks4HCqYvE07-oyepoveLFsarB9ByWwanyRGHUG0LlsaEFuLhWuZA_HcScmMDK3BEZKugfwmZP3cVGulcGJo1j24en3k6S4YUxFzlzfxyIkObKYu7yKwclVU7I_DHr8no4uH4RLTcCC8za9d_lf0iYCvYj1WcoohkWyg5_AZ1aRZdK6e/4nq/o2WIiBxrSn2Q0euIovxe-g/h7/h001.gNfuYg2oN1KEPsAbAAVQ9uH453Bp5eRkZJaq-ns4Hoo">Read more</a></p><p></p><h2>GitHub Brings Claude and Codex AI Agents to Its Platform</h2><p><strong>What:</strong> GitHub is integrating Anthropic&#8217;s Claude and OpenAI&#8217;s Codex as AI coding agents directly into its platform, expanding beyond its existing Copilot offering.</p><p><strong>So What:</strong> This signals GitHub&#8217;s shift from single-vendor AI to a multi-model marketplace approach&#8212;enterprise teams may soon choose which AI agent handles their coding workflows rather than being locked into one provider.</p><p><strong>Now What:</strong> Evaluate whether your current Copilot agreements allow flexibility to test competing agents as they become available.</p><p><a href="https://www.theverge.com/news/873665/github-claude-codex-ai-agents">Read more</a></p><p></p><h2>A16Z Maps AI&#8217;s Winners: Leaders, Gainers, and Surprise Breakouts</h2><p><strong>What:</strong> Andreessen Horowitz published an analysis categorizing AI companies into &#8220;leaders&#8221; (dominant incumbents), &#8220;gainers&#8221; (fast-rising challengers), and &#8220;unexpected winners&#8221; (companies benefiting from AI tailwinds without being AI-native).</p><p><strong>So What:</strong> The framework offers enterprise leaders a useful mental model for evaluating vendors and partnerships&#8212;distinguishing between established players with staying power, aggressive upstarts worth watching, and traditional companies quietly leveraging AI to pull ahead of competitors.</p><p><strong>Now What:</strong> Use this lens when assessing your own vendor stack: are you over-indexed on &#8220;leaders&#8221; who may move slowly, or missing &#8220;gainers&#8221; who could deliver faster innovation?</p><p><a href="https://www.a16z.news/p/leaders-gainers-and-unexpected-winners">Read more</a></p><p></p><h2>Williams F1 Team Partners with Anthropic and Atlassian on AI</h2><p><strong>What:</strong> Williams Racing announced a multi-year partnership with Anthropic&#8217;s Claude and Atlassian to integrate AI across team operations, from race strategy to engineering workflows.</p><p><strong>So What:</strong> F1 teams are data-intensive operations with split-second decision requirements&#8212;this signals enterprise AI moving into high-stakes, real-time environments where the margin for error is measured in milliseconds.</p><p><strong>Now What:</strong> Watch how AI performs in domains where speed and precision are non-negotiable; successful use cases here could inform time-critical enterprise applications in your own operations.</p><p><a href="http://www.thedrum.com/news/anthropic-s-claude-and-atlassian-williams-f1-team-announce-multi-year-partnership">Read more</a></p><p></p><h2>China&#8217;s Kimi K2 Claims Top Open-Source LLM Crown</h2><p><strong>What:</strong> Moonshot AI released Kimi K2, a trillion-parameter open-source model that benchmarks above Claude Opus 4.5 on coding and agentic tasks, available free via API and Hugging Face.</p><p><strong>So What:</strong> The open-source frontier is now a multi-geography race&#8212;enterprises gain another high-capability option outside US providers, but must weigh geopolitical considerations alongside performance.</p><p><strong>Now What:</strong> If you&#8217;re building agentic workflows, benchmark Kimi K2 against your current stack&#8212;the cost-performance math on open models keeps getting more competitive.</p><p><a href="https://venturebeat.com/orchestration/moonshot-ai-debuts-kimi-k2-5-most-powerful-open-source-llm-beating-opus-4-5">Read more</a></p><p></p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://tsw.blankmetal.ai/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading The So What! Subscribe for free to receive new posts and support our work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[Weekly Headlines: Issue #7]]></title><description><![CDATA[January 23 - 27, 2026]]></description><link>https://tsw.blankmetal.ai/p/weekly-headlines-issue-7</link><guid isPermaLink="false">https://tsw.blankmetal.ai/p/weekly-headlines-issue-7</guid><dc:creator><![CDATA[Blank Metal]]></dc:creator><pubDate>Fri, 30 Jan 2026 15:03:22 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!3Q2x!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8ca13492-b8a3-4f6e-98c7-92733414f674_1200x670.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!3Q2x!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8ca13492-b8a3-4f6e-98c7-92733414f674_1200x670.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!3Q2x!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8ca13492-b8a3-4f6e-98c7-92733414f674_1200x670.png 424w, https://substackcdn.com/image/fetch/$s_!3Q2x!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8ca13492-b8a3-4f6e-98c7-92733414f674_1200x670.png 848w, https://substackcdn.com/image/fetch/$s_!3Q2x!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8ca13492-b8a3-4f6e-98c7-92733414f674_1200x670.png 1272w, https://substackcdn.com/image/fetch/$s_!3Q2x!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8ca13492-b8a3-4f6e-98c7-92733414f674_1200x670.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!3Q2x!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8ca13492-b8a3-4f6e-98c7-92733414f674_1200x670.png" width="1200" height="670" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/8ca13492-b8a3-4f6e-98c7-92733414f674_1200x670.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:670,&quot;width&quot;:1200,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:1480264,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://tsw.blankmetal.ai/i/186260079?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8ca13492-b8a3-4f6e-98c7-92733414f674_1200x670.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!3Q2x!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8ca13492-b8a3-4f6e-98c7-92733414f674_1200x670.png 424w, https://substackcdn.com/image/fetch/$s_!3Q2x!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8ca13492-b8a3-4f6e-98c7-92733414f674_1200x670.png 848w, https://substackcdn.com/image/fetch/$s_!3Q2x!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8ca13492-b8a3-4f6e-98c7-92733414f674_1200x670.png 1272w, https://substackcdn.com/image/fetch/$s_!3Q2x!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8ca13492-b8a3-4f6e-98c7-92733414f674_1200x670.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Welcome to Blank Metal&#8217;s Weekly AI Headlines.</p><p>Each week, our team shares the AI stories that caught our attention&#8212;the articles, announcements, and insights we&#8217;re actually discussing internally. We curate the best of what we&#8217;re reading and add the context that matters: what happened, why it matters, and what to do about it.</p><p>Short, sharp, and focused on impact.</p><h2>Amazon&#8217;s One Medical Launches AI Health Assistant for Members</h2><p><strong>What:</strong> One Medical introduced an AI-powered health assistant that helps members get personalized answers, book appointments, and prepare for visits&#8212;all integrated with their medical records.</p><p><strong>So What:</strong> Amazon is quietly building the AI-native healthcare stack, and this signals that consumer-facing AI health tools backed by real clinical data (not just chatbots) are becoming table stakes for healthcare operators.</p><p><strong>Now What:</strong> If you&#8217;re in healthcare or benefits, watch how members respond to AI triage&#8212;this could reshape expectations for how employees interact with any health-adjacent enterprise service.</p><p><a href="https://www.aboutamazon.com/news/retail/one-medical-ai-health-assistant">Read more</a></p><p><em> </em></p><h2>OpenAI and Leidos Partner to Deploy AI Across Federal Government</h2><p><strong>What:</strong> OpenAI announced a partnership with defense contractor Leidos to bring ChatGPT and agentic AI capabilities to federal government agencies, marking OpenAI&#8217;s most significant push into the public sector.</p><p><strong>So What:</strong> This signals AI moving from pilot projects to production infrastructure in government&#8212;and Leidos&#8217; involvement means this is about deployment at scale, not innovation theater. Enterprise vendors should expect federal AI procurement to accelerate.</p><p><strong>Now What:</strong> If you serve federal customers, understand that AI capabilities are moving from &#8220;nice to have&#8221; to table stakes faster than procurement cycles typically allow.</p><p><a href="https://fedscoop.com/openai-chatgpt-leidos-agentic-ai-artificial-intelligence-llm-large-language-models-government-mission-efficiency/">Read more</a></p><p></p><h2>Vercel Launches Marketplace for Shareable AI Agent Skills</h2><p><strong>What:</strong> Vercel released skills.sh, a marketplace for portable &#8220;skill&#8221; files that can be easily installed across multiple AI coding tools, including skills that teach one AI model how to orchestrate another.</p><p><strong>So What:</strong> This signals a shift toward modular, composable AI tooling where enterprises can mix capabilities across models&#8212;potentially letting teams route tasks to the best-fit model rather than being locked into a single provider.</p><p><strong>Now What:</strong> Explore whether standardized skill files could simplify how you manage AI agent capabilities across your stack, especially if you&#8217;re already juggling multiple coding assistants.</p><p><a href="https://skills.sh/">Read more</a></p><p></p><h2>OpenAI Pulls Back the Curtain on Codex Agent Architecture</h2><p><strong>What:</strong> OpenAI published a detailed technical breakdown of how its Codex coding agent works internally, explaining the loop structure that powers its autonomous code generation.</p><p><strong>So What:</strong> This transparency helps enterprise teams understand what&#8217;s actually happening under the hood of AI coding tools&#8212;useful for setting realistic expectations and identifying where human oversight should plug in.</p><p><strong>Now What:</strong> Use this as a reference point when evaluating any agent-based coding tool; understanding the loop architecture helps you spot limitations before they become production problems.</p><p><a href="https://openai.com/index/unrolling-the-codex-agent-loop/?utm_source=tldrai">Read more</a></p><p></p><h2>Claude Gets Interactive Tools for Live Data and Code</h2><p><strong>What:</strong> Anthropic launched interactive tools that let Claude connect to Google apps, run code, create visualizations, and work with files directly within conversations.</p><p><strong>So What:</strong> This moves Claude from chatbot to workspace&#8212;enterprise teams can now build live dashboards, analyze real-time data, and automate multi-step workflows without leaving the interface.</p><p><strong>Now What:</strong> Audit your current workflow gaps where context-switching slows teams down; these native integrations may eliminate the need for custom middleware.</p><p><a href="https://claude.com/blog/interactive-tools-in-claude">Read more</a></p><p></p><h2>Software Engineer Argues SRE Is the Future of the Field</h2><p><strong>What:</strong> Swizec Teller makes the case that as AI handles more code generation, the real value in software engineering shifts to running and maintaining systems reliably&#8212;the domain of Site Reliability Engineering.</p><p><strong>So What:</strong> For enterprise leaders, this suggests your AI coding investments may accelerate a talent shift: engineers who can keep complex systems running become more valuable than those who only write new code.</p><p><strong>Now What:</strong> Audit whether your team&#8217;s skills&#8212;and hiring criteria&#8212;are weighted toward building versus operating, and adjust accordingly.</p><p><a href="https://swizec.com/blog/the-future-of-software-engineering-is-sre/">Read more</a></p><p></p><h2>Alibaba&#8217;s Qwen-3 Becomes First AI Model to Run in Orbit</h2><p><strong>What:</strong> China&#8217;s Adaspace launched Alibaba&#8217;s Qwen-3 model on a satellite, completing a full inference cycle in under two minutes as part of a planned 2,800-satellite AI compute network.</p><p><strong>So What:</strong> This is less about space and more about China&#8217;s long-term bet on distributed AI infrastructure&#8212;a signal that major players are thinking beyond earthbound data centers for compute capacity and resilience.</p><p><strong>Now What:</strong> File this under &#8220;strategic awareness&#8221; rather than action items&#8212;it&#8217;s a useful reference point when evaluating where global AI infrastructure investment is heading.</p><p><a href="https://ground.news/article/alibabas-qwen-3-becomes-first-general-purpose-ai-to-run-in-orbit_c6af03">Read more</a></p><p></p><h2>MCP Gets a UI Layer: Tools Can Now Return Interactive Interfaces</h2><p><strong>What:</strong> Anthropic and partners launched MCP Apps, an extension to the Model Context Protocol that lets tools return interactive UI components&#8212;dashboards, forms, visualizations&#8212;that render directly in conversations rather than plain text.</p><p><strong>So What:</strong> This solves a real gap in agentic workflows: instead of re-prompting for every data exploration step, users can interact with rich interfaces while keeping the AI model in the loop. The &#8220;build once, deploy across Claude, ChatGPT, VS Code&#8221; promise signals MCP maturing into genuine infrastructure.</p><p><strong>Now What:</strong> If you&#8217;re building MCP tools, evaluate whether adding UI components could dramatically improve the user experience&#8212;especially for data-heavy or configuration-intensive workflows.</p><p><a href="https://blog.modelcontextprotocol.io/posts/2026-01-26-mcp-apps/">Read more</a></p><p></p><h2>ChatGPT Can Now Analyze Your Apple Watch Health Data</h2><p><strong>What:</strong> OpenAI enabled ChatGPT to import and analyze Apple Watch health data, letting users ask questions about their sleep patterns, heart rate trends, and activity metrics in natural language.</p><p><strong>So What:</strong> This is the first major consumer AI integration with personal health data at scale&#8212;a proving ground for how AI assistants will handle sensitive, longitudinal personal information and a preview of the &#8220;AI as personal health analyst&#8221; future.</p><p><strong>Now What:</strong> Watch how users respond to AI having access to intimate health data. The trust patterns established here will shape enterprise health AI expectations.</p><p><a href="https://www.washingtonpost.com/technology/2026/01/26/chatgpt-health-apple/">Read more</a></p><p></p><h2>OpenAI Launches Free AI Research Tool, Signals Vertical Playbook</h2><p><strong>What:</strong> OpenAI released Prism, a free AI-powered workspace for scientists built on an acquired LaTeX platform, explicitly modeling the approach Cursor and Windsurf took with code editors.</p><p><strong>So What:</strong> The pattern matters more than the product&#8212;OpenAI is telegraphing that &#8220;acquire specialized workflow tool + add deep AI context&#8221; is the winning formula, which means every vertical-specific SaaS tool is now either a platform for this play or a target.</p><p><strong>Now What:</strong> Audit your team&#8217;s specialized workflow tools (design, legal, finance) and ask which ones have full context of the work being done&#8212;those are where AI integration will hit hardest.</p><p><a href="https://openai.com/prism/">Read more</a></p><p></p>]]></content:encoded></item></channel></rss>