Heating Up
- Anthropic's pricing opacity problemOpus 4.7 uses significantly more tokens than 4.6 under the hood, raising real-world costs even as Anthropic holds nominal pricing steady.
- Multi-agent frameworks go mainstreamOpenAI, Vercel, and open-source projects are shipping production-grade tooling for agent orchestration—no longer just research toys.
- Enterprise AI reality checkServiceNow, Adobe, Salesforce tout agent wins, but TSMC's earnings and RAM shortages hint at infrastructure headwinds beneath the hype.
THE BRIEFING
Anthropic's Opus 4.7 quietly got more expensive to run despite flat list pricing—token counters reveal higher context-window costs. OpenAI released a lightweight multi-agent framework for Python, while TSMC's cautious earnings suggest the chip giant isn't betting big on sustained AI demand. Meanwhile, GitHub's fake-star economy and enterprise adoption stories paint a maturing, messier landscape.
Today's Top 3
Simon Willison's token counter tool exposes what Anthropic didn't advertise: Opus 4.7 consumes more tokens per request than 4.6, meaning higher real-world costs even though per-token pricing stayed flat. This is a classic cloud pricing sleight-of-hand, and developers are not amused. Watch for Anthropic to either clarify or face mounting pressure to revert the change.
The Decoder
OpenAI quietly dropped a native multi-agent orchestration framework for Python—trending hard on GitHub. This is OpenAI formalizing what the community has been hacking together with LangChain and AutoGPT. The fact that it's lightweight and opinionated suggests OpenAI sees agents as infrastructure, not science experiments. If you're building agents, this is now the reference implementation.
GitHub Trending (python)
Ben Thompson flags that TSMC's earnings call was notably cautious—leadership isn't acting like they believe the AI chip boom is sustainable. They're building N3 fabs, but the tone suggests hedging, not conviction. If the world's chipmaker-in-chief is skeptical, that's a data point worth remembering when VCs pitch infinite GPU demand.
Stratechery (free posts)
Frontier Models & Labs
Anthropic updated Claude's system prompt between 4.6 and 4.7, and Willison's diff reveals subtle shifts in how Claude reasons and responds. Anthropic remains the only major lab publishing these prompts, which is both rare transparency and a gift to prompt engineers.
Simon Willison
Hyatt rolled out ChatGPT Enterprise globally, using GPT-5.4 and Codex for operations and guest experience. This is OpenAI's playbook: land enterprise customers, tout productivity gains, and make the models feel indispensable.
OpenAI Blog
Jack Clark's weekly roundup covers automated alignment research, a safety analysis of a Chinese model, and HiFloat4—a new quantization technique. Always dense, always worth skimming for what's moving in research.
Import AI (Jack Clark)
Nvidia's Nemotron OCR v2 uses synthetic training data to achieve fast, multilingual OCR. Synthetic data is becoming the norm for specialized tasks where real-world labeled data is scarce or expensive.
Hugging Face Blog
Enterprise & Business
Adobe is hedging against AI-native competitors by launching an enterprise agent platform. This is Adobe trying to stay relevant as generative AI eats its Creative Cloud moat.
The Decoder
Salesforce is positioning Agent Albert as proof that AI augments SaaS rather than replacing it. Wall Street is skeptical—this is Salesforce's existential pitch.
The Decoder
Fraud charges against iLearningEngines' former execs. The AI gold rush is attracting the usual grifters—this won't be the last indictment.
Hacker News (q: AI)
Google is designing custom AI chips with Marvell, aiming to reduce Nvidia dependence. Two million chips signals serious in-house infrastructure buildout.
The Decoder
The NSA is reportedly using Anthropic's Mythos model. This is Anthropic's FedRAMP play paying off—though the optics of 'AI for spies' won't help the safety narrative.
The Decoder
Anthropic's revenue is reportedly surging, and whispers of a trillion-dollar valuation are circulating. This is venture froth, but it reflects real adoption momentum.
The Decoder
A German court ruled an AI-generated comic based on a copyrighted photo doesn't infringe. This is transformative use doctrine meeting generative AI—precedent to watch.
The Decoder
Hundreds of AI influencers are pushing pro-Trump content on TikTok and Instagram ahead of midterms. This is the 2016 bot problem, now with photorealistic faces and GPT-4 captions.
The Decoder
Chinese humanoid robots beat human runners in Beijing's robot half marathon. This is mostly PR, but the pace of bipedal robotics progress in China is real.
The Decoder
Products & Traction
Swiss government is pushing to reduce Microsoft dependence, citing sovereignty concerns. Expect more of this as AI gets baked into Office and Azure.
Hacker News (q: AI)
TRELLIS.2, an image-to-3D model, now runs locally on Apple Silicon. This is the prosumer generative 3D story accelerating—no cloud, no API keys.
Hacker News (q: GPT)
A prompt-to-diagram demo running Gemma 4 entirely in-browser via WebAssembly. 3.1GB is hefty, but this is on-device inference crossing the usability threshold.
Hacker News (q: GPT)
Supply constraints on high-bandwidth memory (HBM) could persist for years. This is the AI infrastructure bottleneck no one wants to talk about—GPUs are useless without RAM.
Hacker News (q: AI)