Back to archive
Issue #79··38 min read·19 stories

AI Compute Crunch + Desktop Agent Wars Erupt in One Day

Three desktop agent leaks in 24 hours. OpenAI's leaked memo. Stanford AI Index drops.

Anthropic, Google, and OpenAI all had desktop agent product leaks within 24 hours. The AI industry also hit a real compute wall the same week, per a WSJ report: Anthropic's Claude API uptime sat at 98.95% over the 90 days to April 8, OpenAI announced Sora will go dark on April 26 to redirect compute to coding and enterprise tools, and GPU prices have climbed 48% on the Ornn Compute Price Index. An OpenAI internal memo leaked to CNBC also says Microsoft has "limited" the company's enterprise reach.
NEWS

OpenAI revenue chief Denise Dresser told staff in a Sunday memo that the new Amazon partnership is the company's main path to enterprise growth, and that the long-running Microsoft tie-up has 'limited our ability' to reach clients. Amazon committed up to $50 billion in February. The framing signals a clear pivot from OpenAI's most important investor toward its biggest cloud rival.

Vercel's ARR has gone from $100 million at the start of 2024 to a $340 million run rate by February 2026, on the back of AI-generated app deployments. CEO Guillermo Rauch told a HumanX panel the company operates with public-company discipline and is 'ready and getting more ready' for an IPO. The software IPO pipeline is otherwise frozen on AI disruption fears.

Anthropic is finalising a power-user redesign of Claude Code codenamed Epitaxy, with a Cowork-style layout, multi-repo support, and a Coordinator Mode that delegates to parallel sub-agents from inside the app. TestingCatalog also leaked similar plays from Google (a new Agent tab in Gemini Enterprise) and OpenAI (Codex gets a built-in browser and PR review). All three ship as soon as next week.

The WSJ reports AI developers are hitting a hard capacity wall. Claude API uptime sat at 98.95% over the 90 days to April 8, well below the 99.99% cloud standard, and Retool moved workloads to OpenAI over reliability. OpenAI CFO Sarah Friar told the paper she spends much of her time hunting for near-term compute and deciding which projects to shelve because the resources are not there.

The SF Standard reports Altman's Russian Hill home was hit twice in three days. Friday: a Molotov cocktail at the gate, allegedly by 20-year-old Daniel Moreno-Gama. Sunday: a Honda sedan stopped on Lombard and the passenger fired a gunshot; Amanda Tom and Muhamad Tarik Hussein were arrested for negligent discharge, with three firearms found at their residence. Altman shared a family photo and asked critics to de-escalate.

TECHNICAL

Anthropic's engineering team shipped a working taxonomy of multi-agent coordination patterns: generator-verifier, orchestrator-subagent, agent teams, message bus, and shared-state. Each section walks through the mechanics, the failure modes, and the telltale signs you have outgrown one pattern and need the next. The recommendation is to start with the simplest pattern that solves your problem and escalate only when the data forces it.

Calif documents how three engineers and Claude collaboratively built a working exploit for nginx CVE-2026-27654. The bug is a heap buffer overflow in the WebDAV COPY and MOVE handler. The non-obvious lesson: running the same prompt across independent Claude sessions produced one 'impossible' verdict and one working exploit. Independence between AI sessions, not just between humans, is what surfaced the right precondition.

Addy Osmani is calling the next discipline Agentic Engine Optimisation, AEO for short: structuring docs so coding agents can actually use them. Agents fetch via HTTP, strip HTML, count tokens, and silently discard anything that does not fit. The fixes are practical: token-efficient pages, capability signals via skill.md, an llms.txt sitemap, and a robots.txt that does not block AI traffic.

A Towards Data Science piece argues that AI memory systems fail not in storage or retrieval but in hygiene. Memories pile up, decisions reverse, preferences shift, and the system has no mechanism to forget. The proposed fix is a SQLite schema with confidence, decay_score, and expires_at columns, so old facts age out instead of quietly steering the assistant for months after they stopped being true.

LinkedIn used to run five separate systems to choose what showed up in your feed: trending, collaborative filtering, embedding retrieval, and more. ByteByteGo walks through how the team ripped them all out and replaced them with a single LLM-powered retrieval model serving 1.3 billion users in under 50 milliseconds. The piece covers structured profile data, latency budgets, and the ranking gotchas they hit.

Cloudflare extended Dynamic Workers so each AI-generated app can spin up its own Durable Object as a private database. The new piece is persistence: agents can now build small apps with custom UIs and long-lived state, all inside the isolate sandbox. Isolates start 100 times faster than containers and use a tenth of the memory, which makes one-database-per-app economically viable at scale.

ANALYSIS

Search Engine Journal argues that Google's task-completing agents, booking restaurants and planning trips, already represent a structural break for SEO rather than a future one. As agents do the choosing, traditional content visibility metrics break. Builders shipping consumer products need to plan for a web where agents are the readers, not humans, and rethink how their content surfaces to machine intermediaries.

HackerNoon's piece reframes the model wars as engine arguments. The same model wired into different harnesses, Gemini in Sheets versus NotebookLM for example, behaves like a different product. The five concerns that matter are instruction layering, action mediation, loop control, policy enforcement, and memory strategy. Most reliability problems blamed on the model are actually harness design problems.

Bryan Cantrill argues that LLMs threaten the most underrated programmer virtue: laziness, the impulse to build clean abstractions so future you does not have to. Work costs nothing to an LLM, so it happily piles on more code, more layers, more cruft. Without the human friction of finite time, software becomes larger but not better. Simon Willison endorsed the piece the same day.

Ben Thompson argues that AI ends the zero-marginal-cost era that powered Aggregation Theory and the entire 2010s consumer internet. Compute is finite again, and that changes everything from product strategy to capex cycles. He reads Anthropic's Mythos and Meta's Muse Spark releases as the first wave of an industry where opportunity cost between bets is real and increasing.

Stanford HAI's annual AI Index dropped, and the headline finding is that capability gains are still accelerating, not flattening. Industry produced more than 90 percent of notable frontier models in 2025. The nine-chapter report covers research, technical performance, responsible AI, economy, science, medicine, education, policy, and public opinion. It is the canonical reference document for the year.

TOOLS

ppt-master is a Python tool that generates natively editable PowerPoint files from any document, real shapes rather than flattened images, so the output stays editable in PowerPoint or Keynote. It hit 4,687 GitHub stars in days, with 148 added on April 13 alone. Useful if you live in slide-driven environments and currently lose hours rebuilding decks from PDFs or transcripts.

Addy Osmani released Agent Skills, an open framework that gives coding agents the same SDLC workflow a senior engineer would follow: seven slash commands from /spec through /ship, each with verification gates and anti-rationalisation tables to stop the agent from declaring victory early. Installs into Claude Code, Cursor, and Gemini. The opinionated structure is the point, fewer choices and better outputs.

InstantDB is pitching itself as the backend AI agents can drive themselves: every dashboard action, create account, push schema, set permissions, is a CLI command, with built-in undo when an agent does something silly. Auth, storage, real-time streams, and end-to-end type safety come standard. The investor list is the tell: Brockman, Dean, Graham, Saarinen, and 50+ technical founders.