Back to archive
Issue #42··16 min read·8 stories

Pentagon Threatens Anthropic Over AI Safeguards

One team uncovers what fills your AI coding context window. Plus: advanced prompt caching, and AI as an exoskeleton.

The Pentagon yesterday threatened to cut off Anthropic, escalating a dispute over AI safeguards and signaling increasing geopolitical pressure on major model providers. Separately, one team intercepted thousands of API calls to reveal what actually fills your AI coding tool's context window. This comes as a new analysis frames AI not as a coworker, but as an exoskeleton for builders.

NEWS
6 stories

1M Context Window Ships in Claude Sonnet 4.6

Anthropic released Claude Sonnet 4.6, aiming for Opus 4.6 capabilities with a 1M token context window in beta. While Anthropic claims preference wins, independent analysis suggests higher token usage for certain tasks, potentially increasing costs. Early reliability issues were reportedly fixed.

2

Gemini 3.1 Pro Doubles ARC-AGI-2 Reasoning

Google announced Gemini 3.1 Pro, an upgraded AI model that more than doubles its predecessor's reasoning performance on the ARC-AGI-2 benchmark. It's rolling out to developer products and can generate animated SVGs from text or synthesize complex data into dashboards. The release is in preview to validate agentic workflows.

3

AI Hardware Development Accelerates at Apple: Glasses, Pendant, Camera AirPods

Apple is reportedly accelerating development on AI-focused smart glasses, a pendant, and enhanced AirPods designed to leverage Siri with visual context for AI actions. Expect Apple to drive new multimodal AI interaction patterns in consumer hardware, potentially influencing future platform APIs and user expectations.

4

$100B Funding Round Reportedly Nears for OpenAI

OpenAI is reportedly close to a $100 billion funding round, the largest in history, boosting its valuation. This investment, with major tech players involved, signals the immense infrastructure costs and focus on rapid growth over immediate profit in advanced AI development.

5

Pentagon Threatens Anthropic Over AI Safeguards

The Pentagon is reportedly threatening to cut ties with Anthropic due to a dispute over AI safeguards. While the Pentagon seeks models for 'all lawful purposes,' Anthropic insists on limits for mass surveillance and autonomous weaponry, a conflict highlighted by Claude's use in the Maduro operation.

6

20VC x SaaStr: Enterprises Will AI Into Existence, ROI or Not

A 20VC x SaaStr discussion argues Corporate America prioritizes AI adoption over proven ROI, effectively 'willing AI into existence.' This sentiment fuels massive investments in AI companies, exemplified by Anthropic's $380 billion post-money valuation, and creates a 'gravity well' for traditional B2B SaaS. The key insight: enterprise AI spending is locked in for the next 1-2 years, irrespective of immediate returns.

TECHNICAL
2 stories
1

AI Coding Agents Waste Context Tokens

An experiment intercepting 3,177 API calls across four AI coding tools found Gemini uses vastly more tokens by aggressively dumping files and history. Claude models also incur an "architectural tax" from tool definitions. This reveals diverse, often inefficient, context handling strategies, such as aggressive data dumping and fixed architectural overheads, which directly affect token costs and agent performance.

2

Prompt Caching Slashes OpenAI Latency, Costs

OpenAI's prompt caching reuses key/value tensors for identical prompt prefixes, especially for prompts over 1024 tokens. Stabilizing prefixes and consistent tool definitions improve cache hit rates, reducing latency and costs. The Responses API offers better cache utilization than Chat Completions.

ANALYSIS
3 stories
1

Autonomy Doubles in Claude Code, Users Still Intervene

Anthropic research shows Claude Code's autonomous turn duration nearly doubled, hitting the 99.9th percentile. Experienced users auto-approve more often but also intervene more frequently, implying agent systems require effective monitoring and intervention tools.

2

Kasava: AI Should Be an Exoskeleton, Not a Coworker

Kasava argues AI should amplify human capacity like an 'exoskeleton', not operate as an autonomous 'coworker'. The piece criticizes current agentic AI for often failing due to a lack of human context, advocating instead for tools that assist human decision-making and execution.

3

Octoverse Data Shows AI Drives Devs to Typed Languages

GitHub's analysis suggests AI tools like Copilot are shifting developer language choices, creating a 'convenience loop.' This loop makes devs favor AI-integrated tech, with Octoverse 2025 data showing TypeScript usage now surpasses Python and JavaScript. The piece argues AI compatibility is a critical factor in future tech stack decisions.

TOOLS
1 story
1

Agent Tracking, Observability Added to MLflow Platform

MLflow, an open-source platform, now offers end-to-end tracking, observability, and evaluation for building AI agents and models. It integrates these features to manage the AI development lifecycle, supporting Python with an active community.