Welcome to Blank Metal’s Weekly AI Headlines.
Each week, our team shares the AI stories that caught our attention—the articles, announcements, and insights we’re actually discussing internally. We curate the best of what we’re reading and add the context that matters: what happened, why it matters, and what to do about it.
Short, sharp, and focused on impact.
The Reckoning
Three stories this week share a throughline: the costs of moving fast with AI are becoming visible. Token bills, comprehension gaps, and bubble economics are all different faces of the same question—what happens when the honeymoon ends?
You’ve Figured Out AI at Work—Now Comes the Bill
What: The Wall Street Journal reports that enterprises are hitting a new phase of AI adoption: the token bill. Companies that moved aggressively from pilots to production are discovering that AI inference costs scale faster than they expected. The productivity gains are real, but so is the compute bill—and most organizations didn’t budget for what production-scale AI actually costs.
So What: This is the hangover after the honeymoon. The first wave was “look what AI can do.” The second wave was “let’s put it everywhere.” The third wave—happening now—is “who’s paying for all these tokens?” This isn’t a reason to slow down, but it is a reason to be intentional about where AI creates enough value to justify the cost. Not every workflow needs a frontier model.
Now What: Audit your AI usage against actual business value. The 80/20 rule applies: a small number of AI-powered workflows are probably driving most of your value, while a long tail of lower-value uses are burning tokens. Right-size your model selection—use smaller, faster models for routine tasks and save frontier models for high-stakes decisions.
Comprehension Debt: The Hidden Cost Nobody’s Measuring
What: Addy Osmani coined “comprehension debt”—the growing gap between how much code exists in your system and how much any human genuinely understands. Unlike technical debt, which creates visible friction, comprehension debt grows silently until your system breaks and nobody can fix it. An Anthropic study found developers using AI assistance scored 17% lower on comprehension quizzes than control groups.
So What: Your team just shipped 10x faster. Congratulations—you now have 10x more code that nobody fully understands. Tests pass, CI is green, but when something breaks at 2am, the person on call has to reason about code they never wrote, never reviewed, and never internalized. This is a fundamentally different failure mode than technical debt.
Now What: Treat genuine understanding—not passing tests—as non-negotiable. One practical step: require that AI-generated code gets the same review depth as human-written code. If your team is skimming AI output because “it looks right,” that’s the debt accumulating. The teams building comprehension discipline now will be better positioned when the reckoning arrives.
Yes, AI Is a Bubble. The Interesting Question Is What Kind.
What: Derek Thompson and Paul Kedrosky make the case that AI is definitively a bubble—private AI spending will exceed $700 billion in 2026, representing 50-80% of quarterly GDP growth, more than the combined historical spending on 1930s public works, the Manhattan Project, Apollo, and the Interstate Highway System. But they argue it’s a “rational bubble”: each individual actor is behaving rationally, even as the collective outcome is economically unsustainable.
So What: The historical parallel that matters isn’t dot-com—it’s railroads. By 1900, railroads were 62% of U.S. market capitalization despite massive overbuilding, with half of peak-period track eventually abandoned. Tech now represents roughly 60% of the index. The bubble will pop, but the infrastructure will remain and reshape everything it touches. Anthropic doubled revenue in two months. OpenAI added $1B annualized revenue per week. Stripe reports AI companies growing faster than any previous generation.
Now What: Build on the infrastructure while the bubble funds it, but don’t mistake bubble economics for sustainable economics. The companies that thrive post-correction will be the ones generating real revenue from real workflows—not the ones burning venture capital on AI features nobody asked for. If your AI investment can’t justify itself on unit economics today, it won’t survive the correction.
The Human Variable
AI’s biggest open question isn’t technical—it’s human. How do 81,000 users actually feel about it? What happens to the people who built the systems? And why does every organization think it’s further along than it actually is?
What 81,000 People Actually Want from AI
What: Anthropic published the largest multilingual qualitative study of AI users ever conducted—80,508 Claude users across 159 countries. The headline finding: people don’t split cleanly into optimists and pessimists. Those who want emotional AI support are 3x more likely to also fear dependency on it. 81% say AI has already delivered on some aspect of their vision.
So What: The framing of “AI believers vs. skeptics” is wrong. Real users hold both simultaneously—they want the productivity gains (32% cite this as the primary delivered benefit) while worrying about job displacement (22.3%) and loss of autonomy (21.9%). Lower-income countries are significantly more optimistic than wealthy ones, which inverts the usual tech adoption narrative.
Now What: If you’re rolling out AI tools internally, don’t segment your workforce into supporters and resisters. Design adoption programs that acknowledge both the excitement and the anxiety—because the same people feel both. The “cognitive partnership” framing (17% of users describe AI this way) resonates more than “productivity tool.”
What Do Coders Do After AI?
What: Anil Dash, writing for the New York Times Magazine, draws a line that most AI commentary misses: “In the creative disciplines, LLMs take away the most soulful human parts of the work and leave the drudgery to you. In coding, LLMs take away the drudgery and leave the human, soulful parts to you.” He identifies two cohorts of coders—the 9-to-5 professionals facing devastating displacement, and the craftspeople watching their medium transform into something unrecognizable.
So What: 700,000 tech workers have been laid off in the last few years. We’ll be at a million soon. But the displacement isn’t uniform. The “journeyman coders” writing standardized business logic are the most vulnerable—that’s exactly the code LLMs generate best. Meanwhile, coders who see it as craft are experiencing a different kind of loss: their job is becoming “describing software” rather than writing it. Both are painful, but they require completely different responses.
Now What: If you manage engineering teams, this framework matters for retention and hiring. Your most valuable people aren’t the ones who write the most code—they’re the ones who understand why the system works. As Osmani’s comprehension debt concept makes clear, the ability to reason about code is becoming more valuable than the ability to write it. Hire for judgment, not velocity.
What’s Your AI Adoption Level?
What: Steve Yegge published an AI adoption maturity framework that’s resonating across the industry—a clear progression from “Not Using AI” through “AI-Assisted” to “AI-Native” with specific behaviors at each level. The framework maps where individuals and organizations actually sit versus where they think they are.
So What: Most organizations overestimate their AI maturity because they conflate tool access with adoption. Having ChatGPT licenses doesn’t make you AI-assisted any more than having a gym membership makes you fit. The framework exposes the gap between “we have AI tools” and “our workflows have fundamentally changed.”
Now What: Use this as a self-assessment. Where does your team actually sit—not where leadership thinks they sit? The honest answer shapes whether you need more tools, more training, or more workflow redesign. Most organizations discover they need the third one.
The Agent Economy
Design tools that replace designers. Enterprise leaders planning agent deployments. A strategist declaring the bubble debate over. The agent economy isn’t emerging—it’s arriving, and the market is repricing everything around it.
Google Launches “Vibe Design” with Stitch—Figma Drops 8%
What: Google Labs unveiled Stitch, an AI-native UI design platform with an AI canvas, smarter design agent, voice input, instant prototyping, and built-in design system support. The market reacted immediately—Figma’s stock dropped 8% on the announcement, now down 80% from its August 2025 IPO.
So What: This is the design tool version of what happened to coding: AI collapses the gap between intent and artifact. Stitch doesn’t just assist designers—it lets non-designers produce high-fidelity UI through natural language and voice. The stock reaction tells you the market believes this shift is structural, not incremental.
Now What: If your team is evaluating design tooling or hiring designers, watch this space closely. The question is shifting from “which design tool?” to “do we need the same number of designers?”—and the answer will look different in six months than it does today.
Aaron Levie: What 20+ Enterprise IT Leaders Are Actually Saying About AI
What: Box CEO Aaron Levie sat down with 20+ enterprise AI and IT leaders—particularly from regulated industries—and shared the emerging consensus. Agents are “clearly the big thing,” with enterprises moving from experimental chatbots to production agent deployments. But the infrastructure isn’t ready: governance models are immature, payment rails for machine-to-machine transactions don’t exist, and most organizations are still figuring out where agents fit in their org charts.
So What: When the CEO of a $5B enterprise software company reports from the field, it’s a demand signal. The shift from “chatbot pilots” to “agent deployments” is happening, but the gap between ambition and infrastructure is widening. Only one in five companies has a mature governance model for agent deployments. The rest are flying blind or moving slowly.
Now What: If you’re planning enterprise AI rollouts, governance and observability should be in your architecture from day one—not bolted on after agents are already running. The organizations that get agent governance right early will move faster later. The ones that skip it will hit a wall when the first production agent does something unexpected.
Ben Thompson: Why Agents Mean This Isn’t a Bubble
What: Ben Thompson makes his most definitive macro call on AI yet: we’re not in a bubble. His argument rests on three LLM paradigm shifts—ChatGPT (2022), reasoning models like o1 (2024), and agents via Opus 4.5/Claude Code (late 2025). Each shift addressed a core LLM weakness, and agents are the inflection that changes the economics. The key insight: agents don’t just require a better model—they require integration between model and harness, which means Anthropic and OpenAI are becoming the differentiated point in the value chain, not commoditized infrastructure.
So What: Thompson identifies two dynamics that separate agents from prior AI hype. First, agents dramatically reduce the number of humans needed to drive compute demand—a small number of people wielding agents creates exponentially more economic output than chatbot adoption ever could. Second, Microsoft’s decision to bundle Anthropic’s Claude into its new $99/seat E7 enterprise tier (via Copilot Cowork) is an admission that model-agnostic strategies don’t work for agents. If agents require integrated model+harness, the companies building that integration capture the profits.
Now What: If Thompson is right, the strategic question for enterprises shifts. It’s not “which model should we use?” but “which agent platform are we building on?” The model-agnostic approach that seemed prudent a year ago may now be a liability—because agents aren’t modular. For organizations evaluating AI investments, this argues for deeper commitment to fewer platforms rather than hedging across many.
The Practitioner’s Edge
Two tools this week that separate the people talking about AI from the people building with it.
The MCP Debate Settles: CLI for Developers, MCP for Organizations
What: A viral blog post declared “MCP is Dead” in favor of CLI tools, arguing that LLMs already know jq and curl so MCP wrappers add unnecessary complexity. Cloudflare responded with “Code Mode”—a new approach where AI agents write TypeScript against MCP tool APIs instead of using specialized tool-calling syntax, improving both performance and token efficiency by 47%.
So What: Both sides are right about different problems. CLI tools win for individual developers who already have the right access and know the tools. But MCP over streamable HTTP solves the enterprise problem: centralized tool servers with proper auth, shared infrastructure across teams, and audit trails. That’s the difference between one developer vibe-coding and an org shipping agents at scale.
Now What: Stop debating MCP vs. CLI as a binary. Use CLI tools where the developer already has access and the LLM already knows the tool. Use MCP servers where you need centralized governance, shared access, and auditability. Cloudflare’s Code Mode suggests the best of both worlds: MCP infrastructure with code-native invocation patterns.
Defuddle: The Markdown Converter LLM Workflows Need
What: Defuddle is a lightweight tool that converts any web page into clean Markdown with YAML frontmatter. Available as an API, browser extension, and bookmarklet—it also handles YouTube transcription. Think of it as a universal adapter between the messy web and the structured context that LLMs prefer.
So What: LLMs—especially in coding and workflow contexts—perform dramatically better with Markdown input than raw HTML or copy-pasted text. Every time you paste a URL into an AI tool and get a mediocre response, the problem is often the input format, not the model. Tools like Defuddle solve the “last mile” problem of getting clean context into AI workflows.
Now What: Add this to your AI toolkit. When feeding articles, documentation, or web content into AI workflows, convert to Markdown first. The token efficiency gains alone are worth it—but the real win is better AI output from cleaner input. For engineering teams, consider wrapping this in an MCP server for agent workflows.



