Back to archive
Issue #34··14 min read·7 stories

xAI-SpaceX Valued at $1.25 Trillion

OpenAI eyes $100B raise; LLMjacking threatens self-hosted AI; Anthropic readies Sonnet 5.

Elon Musk's xAI and SpaceX are reportedly combining with a $1.25 trillion valuation, reflecting an aggressive push for compute and market dominance. OpenAI is also reportedly trying to raise $100 billion at an $830 billion valuation, signaling sustained investor confidence in foundational model companies. Meanwhile, LLMjacking emerged yesterday as a costly new threat for teams running self-hosted AI infrastructure.

NEWS
6 stories

Stronger Coding, Competitive Math, Detailed Visual Generation Emerge from Sonnet 5 Leaks

Anthropic is reportedly launching Claude Sonnet 5 during Super Bowl week. Early tests suggest Sonnet 5 shows competitive performance in math and better coding output than previous Claude versions, particularly in structured visual generation. The model is expected to feature a 128k context window, targeting a faster, lower-cost tier than Opus.

2

Orbital Data Centers Drive $1.25T SpaceX-xAI Merger

SpaceX acquired Elon Musk's AI startup xAI in a $1.25 trillion merger. The deal, structured as a share exchange, values SpaceX at $1 trillion and xAI at $250 billion. A stated driver for the merger is the development of "orbital data centers" to meet AI infrastructure demand. The acquisition also provides capital to xAI as it expands infrastructure to compete with rivals.

3

Open-Source Jarvis AI Delivers Local, Private Voice Dictation

Jarvis AI Assistant is a free, open-source voice dictation tool that includes automatic punctuation, filler word removal, grammar correction, and text generation, with full prompt engineering customization. Users choose between local models (Whisper, Parakeet, Ollama) for privacy or cloud services (Deepgram, Gemini) for speed. The project ensures privacy with zero tracking and telemetry, available for Mac and iOS.

4

AST Parsing Cuts Agent Context Tokens 90% vs. RAG

Archimyst's agentic runtime cuts token consumption by up to 90% compared to RAG systems, accelerating AI reasoning. It uses PageRank-weighted AST graph parsing for codebase understanding. The platform provides full-repository awareness and local-first architecture for complex agentic tasks.

5

OpenAI Reportedly Seeks $100B at $830B Valuation

OpenAI is reportedly in talks for a $100 billion funding round at an $830 billion valuation, aiming to close by Q1 next year. This capital would fund AI development and compute costs to maintain a competitive edge. The move comes as investor sentiment cools on AI's growth and chip supply tightens.

6

Personal AI Agent Connects Messaging to Apps

Clawd is a self-hosted AI assistant that runs on messaging apps like WhatsApp, Telegram, and iMessage. It integrates with over 500 services via Composio, uses Claude models, and offers browser automation through an isolated Chromium instance or existing Chrome sessions. Builders set it up with API keys and a config file.

TECHNICAL
4 stories
1

RePo Repositions LLM Context, Outperforming Baselines on Long-Context Tasks

Sakana AI researchers introduced Context Re-Positioning (RePo), a technique allowing LLMs to dynamically re-organize their internal view of input data for long-context tasks. RePo uses a lightweight neural network to dynamically re-organize token positions by content, instead of fixed positional embeddings. This helps LLMs cluster related information, outperforming baselines on complex tasks like retrieval-augmented generation and question answering over extended documents. RePo outperformed standard models by over 11 points on the RULER benchmark and maintained high accuracy when extrapolating to 4x longer contexts (up to 16,000 tokens).

2

Build AI Product Sense with Agents: A Hands-On Guide from Lenny's Newsletter

Lenny's Newsletter presents a hands-on approach to building 'AI product sense' using AI coding agents like Cursor. The article shows readers how to set up Cursor, experiment with AI models, and inspect tool calls, culminating in building a personal AI operating system. It explains concepts like RAG, agent memory, and context engineering through practical application, enabling builders to directly grasp AI product development.

3

Weaviate: AI Agents Need Memory as Infrastructure

Weaviate argues that stateless AI agent interactions are a systems problem, not a model limitation, causing repetitive input and inefficiency. Their solution treats memory as core infrastructure, not a feature. This requires active management like write control, deduplication, and purposeful forgetting to build durable, programmable memory systems for evolving agents.

4

LLMjacking Hijacks Exposed Self-Hosted LLMs

A new threat called LLMjacking targets self-hosted LLM infrastructure, where attackers hijack exposed endpoints for illicit monetization. The attack scans for vulnerable services like Ollama, validates access, then sells it on dark web markets. Insecure default configurations are the main vulnerability. Solutions include reverse proxies with authentication (e.g., Nginx with API keys), network isolation for MCP servers, and token-based rate limiting.

ANALYSIS
2 stories
1

Alderson: Two AI User Types Create Enterprise Gap

Martin Alderson argues two AI user types are emerging: 'power users' on advanced tools like Claude Code, and 'ChatGPT users' sticking to basic conversational AI. He states enterprise environments, often limited to less capable tools, fall behind smaller companies where employees freely adopt sophisticated AI. Corporate IT policies limiting local scripting and legacy system integration worsen this gap. Alderson predicts companies offering accessible APIs and secure AI sandboxes will gain a competitive edge.

2

Analysis: Google Wins Chatbot Race Via Default Integration

An analysis argues Google is outperforming OpenAI by integrating Gemini into existing internet defaults like search, browsers, and mobile OS. This makes AI interaction ambient rather than a separate destination. Google's financial ability to subsidize AI at scale further supports this strategy.

TOOLS
1 story
1

Microsoft AutoGen Orchestrates AI Agent Teams

AutoGen is a framework for building AI applications with multiple collaborating agents. Developers define agent roles and capabilities, letting them communicate to handle tasks like coding, debugging, and problem-solving through conversation. It enables customizable agent workflows for complex systems.