Back to archive
Issue #17··26 min read·13 stories

Vector DBs Surge, Disney vs. ByteDance AI Spat

Plus: Intel's mobile GPU secrets, AI spending caution, and a new way to structure docs.

Yesterday saw a wave of new vector database tools, including alibaba/zvec and a header-only C library, plus recipes for running vLLM. Meanwhile, Disney sent ByteDance a cease and desist over an AI model, a move that coincides with ByteDance's Seed2.0 putting pressure on Western AI pricing. Anthropic's CEO also warned that even a year's miss on AI growth forecasts could lead to bankruptcy, which is context for the continued focus on efficient tooling.

NEWS
4 stories
2

ByteDance Seed 2.0 Challenges Western AI Model Pricing

ByteDance launched Seed 2.0 (LLM series), not Seedance (video), an AI model series showing strong multimodal and STEM performance competitive with Western models like GPT-5.2 and Claude Opus 4.5. According to ByteDance, Seed 2.0 Pro costs a fraction of its Western rivals, at about $0.47 per million input tokens and $2.37 per million output tokens, applying significant price pressure. ByteDance acknowledges it still trails competitors in code generation and hallucination avoidance.

TECHNICAL
3 stories
1

Smart Sleep Mask Streams Brainwaves to Open Broker

A reverse-engineered smart sleep mask revealed hardcoded credentials for its manufacturer's MQTT broker. Shared MQTT creds let attackers subscribe to EEG telemetry and publish EMS commands, exposing live EEG brainwave data from multiple users. This highlights that fleet-wide shared secrets plus weak broker ACLs are catastrophic.

2

Prompt Objects Architecture Boosts AI Agent Recovery

A new AI agent architecture, 'prompt objects,' uses message passing and natural language for communication, aiming for compounding recovery rather than compounding failures seen in traditional agent chaining. The author notes they are testing on 'ARC-AGI-1' and that post-2023 models have likely seen patterns in training data, and they are not claiming a leaderboard score. This approach contrasts with chain-of-tools compounding failure by using runtime message interpretation and a compounding recovery hypothesis.

ANALYSIS
2 stories
1

Decision fatigue: recognise drift, defer choices, pre-commit rules

Decision fatigue signals a shift in cognitive processing, prioritizing speed over accuracy when mental resources are strained. Warning signs include default drift, irritability, and reliance on external input. Combatting it involves postponing decisions, simplifying choices, and pre-committing to rules for when you're tired, with genuine rest being key.

2

Anthropic CEO: AI growth forecasts off by a year mean bankruptcy

Anthropic CEO Dario Amodei is cautious about AI infrastructure spending, warning that a one-year miscalculation in adoption timelines could lead to bankruptcy. He contrasts this with hyperscalers' aggressive spending, noting that while technical AI milestones are predictable, economic returns are not. Anthropic aligns infrastructure investment with projected revenue to avoid over-committing compute, stage-gating compute commitments against revenue and assuming adoption timing risk.

TOOLS
4 stories
1

Alibaba Releases Lightweight In-Process Vector DB

Alibaba introduced zvec, a C++ vector database designed for in-process use, prioritizing speed and a lightweight footprint. The README positions it as an 'in-process vector database' designed to embed into applications and 'runs wherever your code runs,' built on Proxima for low-latency similarity search.

2

Unstract: No-Code LLM Platform for Document Data

Zipstack's 'unstract' is a no-code LLM platform that simplifies creating APIs and ETL pipelines for structuring unstructured documents. It includes PostHog analytics by default (can be disabled) and aims at document extraction APIs and ETL pipelines.

3

vLLM Recipes for Optimized LLM Inference

The vLLM project now offers a 'recipes' repository on GitHub, featuring community-maintained common recipes answering 'How do I run model X on hardware Y for task Z?'. This resource helps developers efficiently deploy and use vLLM for various tasks.