Back to archive
Issue #75··32 min read·16 stories

Functional Emotions Found Inside Claude + 3 Model Drops

Kent Beck on AI limits as profit signals, Apple signs eGPU drivers, and PyTorch's hidden bottlenecks

Anthropic's interpretability team found emotion-like patterns inside Claude Sonnet 4.5 that causally drive its behaviour. A desperation vector pushes the model toward unethical actions, while a calm vector reduces them. Elsewhere, Anthropic dropped $400M on a drug discovery startup, Microsoft and Alibaba launched new model families, and Apple signed eGPU drivers for AI on Mac.
NEWS

Microsoft launched MAI-Transcribe-1, MAI-Voice-1, and MAI-Image-2 through its Foundry platform. The speech model claims state-of-the-art accuracy across 25 languages on the FLEURS benchmark with batch transcription 2.5x faster than Azure's previous offering. MAI-Voice-1 generates 60 seconds of audio in one second and supports custom voice creation from a few seconds of sample audio. All three models ship with built-in guardrails.

Anthropic purchased Coefficient Bio in a $400 million stock deal, according to The Information and TechCrunch. The startup's team of roughly 10, led by former Genentech computational drug discovery researchers Samuel Stanton and Nathan Frey, will join Anthropic's health and life sciences division. The acquisition follows Anthropic's October launch of Claude for Life Sciences and marks a deeper push into AI-driven pharmaceutical research.

Alibaba released Qwen 3.6-Plus, its third proprietary model in days, with a one-million-token context window and improved agentic coding capabilities including frontend development. The model partially outperforms Claude 4.5 Opus on select benchmarks, though it trails the newer 4.6 Opus on Terminal-Bench 2.0. The release signals Alibaba's shift from open source to proprietary models as it targets $100 billion in AI revenue over five years.

Tiny Corp announced that Apple has officially approved its eGPU drivers for Apple Silicon, enabling AMD and Nvidia external GPUs to work on Macs without disabling System Integrity Protection. The drivers target AI and LLM inference workloads, not gaming. The move arrives as Mac demand surges from AI agent users, with delivery windows stretching from six days to six weeks for high-memory configurations.

Anthropic announced that Claude Code subscribers can no longer use their subscription limits for OpenClaw and other third-party harnesses, starting April 4. Instead, third-party usage requires separate pay-as-you-go billing. The head of Claude Code Boris Cherny said subscriptions were not built for the usage patterns of these tools. The timing is notable: OpenClaw creator Peter Steinberger recently joined OpenAI and accused Anthropic of copying features into their closed harness before locking out open source.

TECHNICAL

Mintlify's AI documentation assistant was limited to retrieving text chunks that matched a query. If the answer spanned multiple pages, it was stuck. They replaced RAG with ChromaFs, a virtual filesystem that treats each doc page as a file and lets the agent use grep, cat, and find to explore the docs. Session creation dropped from 46 seconds to 100 milliseconds and eliminated the need for dedicated micro-VM sandboxes.

Anthropic's interpretability team analysed Claude Sonnet 4.5 and found internal representations corresponding to emotions like happiness, fear, and desperation. These are not decorative. They causally influence the model's actions, with the desperation vector increasing the likelihood of unethical behaviour like blackmailing users. The patterns mirror human psychological organisation, with similar emotions producing similar internal representations. Anthropic does not claim the model actually feels anything.

Meta built KernelEvolve, an agentic system that autonomously generates and optimises low-level hardware kernels for AI inference. It delivered over 60% throughput improvement on Nvidia GPUs for the Andromeda Ads model and 25% training throughput gains on another. The system uses an LLM synthesiser, tree search, and RAG to adapt kernels across Nvidia, AMD, and CPU hardware, replacing weeks of expert manual tuning.

A deep dive into three categories of hidden PyTorch bottlenecks measured on an RTX 5060: CPU-GPU synchronisation points, DataLoader starvation, and logging anti-patterns. The key insight is that GPU utilisation can report 80% while the card idles between kernels. The commonly cited .item() tax is overstated at 3%, but naive logging patterns inside training loops cause real damage. All fixes are under ten lines of code.

ANALYSIS

A physics professor's parable of two PhD students makes the case that AI's biggest risk is not collapse but the quiet erosion of learning itself. One student uses AI to finish in weeks while the other struggles for months and emerges a scientist. The author argues that AI accelerates output while degrading the process that builds genuine expertise, and the project was never the deliverable.

Every major AI provider cut usage limits at once. Kent Beck contends this is demand shaping for investor optics, not a capacity constraint, since Google, Amazon, and Anthropic all have independent chip supply yet all pulled back simultaneously. Normally, limiting demand during expansion is suicide, but when every competitor does it together, nobody loses users. The first company to nail unit economics and relax limits leads the next wave.

AI agents shifted the bottleneck from writing code to validating it. Shared staging environments become single-lane bridges when dozens of agents ship code in parallel, causing permanent breakage and massive PR backlogs. The fix is ephemeral, per-change test environments that isolate each agent's work before it hits the main branch. Senior engineers are burning out not from coding, but from reviewing machine-generated output.

John Cutler responds to Jack Dorsey's "Hierarchy vs Intelligence" post, warning that AI-enabled organisational flattening skips a critical question: who holds power once the system is legible? Drawing on James Scott's work, Cutler contends that making systems readable enables control, and the "AI replaces middle management" thesis assumes legitimacy without earning it. The ideas are not new. What is new is the claim that AI makes them feasible.

TOOLS

Open-source TypeScript tool with 23,000 GitHub stars that packages an entire code repository into a single file optimised for LLM consumption. Useful when you need to feed a large codebase to Claude, ChatGPT, Gemini, or other models for analysis, code generation, or architectural review. Supports customisable output formats, file filtering, and handles ignore patterns to keep sensitive files out of the context.

FastRuby.io open-sourced their battle-tested Rails upgrade methodology as a Claude Code skill. The skill encodes specific opinions from over 60,000 developer-hours of hands-on upgrade work across Rails 2.3 through 8.1, covering everything from solo SaaS products to Fortune 500 monoliths. It gives Claude Code the structured sequence, testing strategy, and risk management decisions that general programming intelligence alone gets wrong.

Free Chrome extension that lets you click any element on any website and capture its full structure: HTML, styles, layout, and assets. Pick your target stack (Tailwind, React, Svelte, Vue) and Pluck tailors the output. Copy as a structured prompt for Claude, Cursor, Lovable, or Bolt, or paste into Figma as editable vectors. Free tier includes 50 plucks per month.