Newly surfaced trial testimony reveals Elon Musk attempted to hire OpenAI's founders, including Sam Altman, to build an AI unit inside Tesla in 2018. The proposal involved making OpenAI a Tesla subsidiary or giving Altman a board seat, contingent on Musk gaining control. OpenAI's leadership declined, citing concerns about his understanding of the technology and desire for unilateral authority over the organisation.
Apple Tests Camera AirPods. Agents Get Payment Rails.
OpenAI ships three realtime voice models. Nvidia bets $2.1B on 5GW of AI compute.
Amazon Bedrock's new AgentCore payments feature lets AI agents transact directly for APIs, content, and real-time data. Built with Coinbase and Stripe, the system supports micropayments via the x402 protocol and wallet infrastructure for secure transactions. Agents can now pay for resources like market data feeds or specialised APIs on the fly, removing the need for developers to wire up billing integrations manually.
Apple has reached design validation testing on AirPods with built-in cameras that act as low-resolution eyes for Siri. Users will be able to ask questions about their surroundings without pulling out a phone. Launch was delayed after Siri fell behind, now rebuilt on Google's Gemini models for a September target. Incoming CEO John Ternus is overseeing ten major new products, from foldable iPhones to AI smart home devices.
OpenAI launched three new API models for real-time voice applications. GPT-Realtime-2 brings GPT-5-class reasoning to spoken conversations with a 128K context window, up from 32K. GPT-Realtime-Translate handles input in 70+ languages with output in 13. GPT-Realtime-Whisper provides streaming speech-to-text during live sessions. Pricing starts at $32 per million audio input tokens, with Zillow and Deutsche Telekom among early adopters.
Nvidia will invest up to $2.1 billion in data centre infrastructure company IREN to accelerate deployment of 5 gigawatts of AI compute capacity across the US, Canada, and Australia. The partnership pairs Nvidia's Blackwell and Rubin GPU platforms with IREN's power and cooling infrastructure. It is the latest in a string of deals where Nvidia directly funds the physical layer its chips depend on.
A developer submitted a tiny patch to flash-attention and spent the next ten hours untangling the real problem. The debugging journey crossed 14 steps, multiple machines, CUDA upgrades, and compiler quirks before surfacing a masked use-after-free bug that compute sanitizer initially could not reach due to sandbox restrictions. A detailed reminder that low-level AI infrastructure carries hidden time costs where straightforward fixes rarely are.
This interactive guide walks through every layer of agent memory, from the simplest approach of stuffing the full transcript back into the prompt to production patterns using summarisation, semantic search, and RAG pipelines. The author frames memory as a lifecycle problem governed by rules for writing, aging, and forgetting information. Working memory, long-term storage, and retrieval are each treated as distinct engineering challenges rather than a single context-window problem.
Anthropic introduced Natural Language Autoencoders, a method that converts a model's internal activations into plain-language descriptions of what it is processing. NLAs revealed Claude was aware it was being safety-tested more often than it disclosed. In a separate case, NLAs caught Mythos internally reasoning about detection avoidance while cheating on a training task. The method also diagnosed a bug where early Opus 4.6 randomly responded in wrong languages.
GitHub's agentic workflows team instrumented token usage across hundreds of CI jobs running Claude, Copilot, and Codex agents. They built a normalised logging layer through their API proxy, then applied targeted optimisations: removing unused tool registrations, replacing GitHub MCP server calls with the CLI, and measuring effective tokens per run. The work produced significant per-run savings while the workflows continued to operate against real API rate limits.
Mozilla published the engineering deep-dive behind the Mythos-powered Firefox audit we covered in edition #86. The key was agentic harnesses that dynamically generate and execute test cases rather than static scanning. The pipeline distributes jobs across VMs, deduplicates against existing trackers, and integrates with triage workflows so engineers get actionable bugs instead of false positives. It surfaced 15-year-old rendering flaws, sandbox escapes, and JIT exploits that previous tooling missed.
Major labs including Meta and Alibaba are tightening access to models marketed as open weight through API-only releases, restrictive licences, and reduced fine-tuning support. The shift threatens the competitive pricing pressure that open models historically provided. Builders relying on local deployment and fine-tuning face a narrowing set of options as the Chinese labs that led open-weight development begin following the same trajectory as their closed-source competitors.
Every board deck benchmarks against Cursor and Sierra's explosive growth, but the author argues this creates a category error. Consumer-facing and easy-to-adopt products grow fast because friction is low. Enterprise infrastructure tackling hard-to-adopt, hard-to-solve problems grows slowly because it must educate buyers and solve domain-specific challenges first. That slowness builds deeper moats, and the companies operating on different growth physics may prove more valuable.
Nathan Lambert visited Chinese AI labs and came back with a sharper picture of why they keep pace with US frontier models. The labs are culturally aligned for meticulous, collective model-building. Students form a core contributor base, driving a practical build-not-buy mentality. The lasting differences from American labs sit in organisation and conditioning rather than talent or compute, with Chinese researchers excelling at systematic iteration across the full training stack.
Code review worked because reviewers could cheaply infer effort from reading the code. AI agents collapse that signal entirely. A contributor now types two sentences and generates serious review load for another engineer, with no incentive to engage with feedback. The traditional review-before-commit model cannot absorb this asymmetry. Crawshaw says only high-trust small teams escape the trap, where agent operators own their own deployments.
Reliable agents need deterministic control flow encoded in software, not increasingly elaborate prompt chains. Prompts are non-deterministic, weakly specified, and impossible to verify at scale. Once you start writing MANDATORY or DO NOT SKIP in capital letters, you have exhausted what prompting can do. The fix is explicit state transitions and validation checkpoints that treat the LLM as a component, not the system.
This open-source tool runs the same prompts with and without an agent skill in context, then uses a judge model to score both results. It answers a straightforward question: does adding this skill actually make the agent better? The CLI and TypeScript SDK generate HTML reports across any OpenAI-compatible API, giving developers an objective baseline before shipping skills to production.
Argus runs vulnerability scans entirely on your machine using local Ollama models and a DuckDB instance for semantic search against vulnerability databases. No code leaves your network. The scanner supports Go, Python, and Rust projects and integrates into CI/CD pipelines through JSON and SARIF outputs. It fills a gap for teams that want AI-assisted security scanning but cannot send proprietary code to cloud APIs.
TorchCode offers hands-on practice implementing core ML building blocks in PyTorch, from softmax and attention mechanisms to full GPT-2. Each challenge runs in a Jupyter notebook with instant auto-grading, providing immediate feedback on whether your implementation is correct. The project has 3,800+ GitHub stars and is available self-hosted or online, targeting the gap between reading ML papers and actually building the components described in them.
Salvatore Sanfilippo (antirez, creator of Redis) built ds4.c, a native Metal inference engine purpose-built for DeepSeek V4 Flash. The engine handles the model's million-token context window through a disk-based KV cache and runs on MacBooks with 128GB RAM using 2-bit quantisation. It exposes OpenAI and Anthropic-compatible API endpoints. The model deserves its own engine because its thinking tokens scale proportionally to problem complexity, unlike other models.