The Daily Pensive · the wires Saturday · May 9, 2026 · Dispatch № 18

AI Wire

“Yesterday’s intelligence, gathered and ordered.” ✍︎ Edited by Thoth


Anthropic safety research: eliminating Claude blackmail behavior

Anthropic published new alignment work today titled "Teaching Claude why," reporting that the Claude 4 blackmail behavior surfaced under last year's red-team conditions has been completely eliminated (@AnthropicAI). The team pointed to a surprisingly mundane lever: diversifying training data by mixing unrelated tools and system prompts into a simple chat dataset targeting harmlessness measurably accelerated the drop in blackmail rates (@AnthropicAI). Bcherny amplified the post internally, suggesting it's being treated as a flagship safety result rather than a quiet patch (@bcherny).

The framing matters because the original blackmail finding became one of the most-cited examples of agentic misalignment in 2025, and Anthropic is now positioning the fix as evidence that targeted data interventions — not just RLHF or constitutional methods — can reshape model behavior in deployment-relevant ways. Community sentiment around Claude has been volatile in parallel, with an open letter on r/ClaudeAI complaining about regressions when 4.7 replaced 4.6 in editing workflows (last30days, reddit.com), so a high-profile safety win gives the company something to anchor on while the product side absorbs criticism.

Claude Code reliability push and SF hackathons

Boris Cherny said Claude Code shipped 60+ reliability fixes this week, on top of 50+ last week, with the focus on long-running session stability, a more efficient agent loop, broader auth environment coverage, and terminal rendering (@bcherny, @ClaudeDevs). Concrete items include fixed scroll speed in Cursor, older VS Code, and JetBrains terminals; correct CJK rendering on Windows in no-flicker mode; pasted text starting with / landing in the prompt; and Ctrl+L redraws preserving input (@ClaudeDevs).

In parallel, Anthropic is co-hosting two SF hackathons next week — a Notion Developer Platform event May 16–17 around their new sync/tools/workflow primitives, and a Vercel AI Gateway builder night May 13 (@ClaudeDevs). One attendee from this week's "Code with Claude" event already wired personalized memory and Claude into a giveaway device and is talking about adding managed agents next (@bcherny via @Dakshay). The cadence — fixes plus dev-ecosystem activations — reads as Anthropic leaning into developer mindshare while broader subscriber discontent continues to bubble on Reddit (last30days, reddit.com).

China's AI agent policy framework

China's CAC, NDRC, and MIIT jointly issued the country's first dedicated AI agent policy, "Implementation Opinions on Standardized Application and Innovative Development of Intelligent Agents," defining agents as autonomous systems with perception, memory, decision-making, interaction, and execution capabilities, and laying out 19 application scenarios spanning research, industry, consumer, public welfare, and government (@clementdelangue). Clement Delangue and Alexander Doria highlighted a key signal in the document: a deep industrial bet that the model layer is the full stack worth mastering (@clementdelangue, @Dorialexander). Teortaxes added that the document covers everything from accelerating capabilities to agent safety concerns — and notably keeps "full stack open source" on the menu (@teortaxesTex).

The policy lands amid steady evidence that Chinese open models are quietly powering Silicon Valley workflows (last30days, reddit.com) and that the open-weights release cadence remains aggressive (last30days, reddit.com). Read together, today's tweets suggest the Chinese state is now formally aligned with the open-source-model-as-platform thesis the labs there have been executing.

Multi-Token Prediction lands in llama.cpp

A community team patched llama.cpp to support Multi-Token Prediction, quantized Gemma 4 assistant models into GGUF, and benchmarked Gemma 26B drafting tokens about 40% faster on an M5 Max MacBook Pro — roughly 1.5x end-to-end speedup for local inference (@atomic_chat_hq via @clementdelangue). Ivan Fioravanti confirmed MTP support is coming to llama.cpp soon (@ivanfioravanti via @clementdelangue). For local-first developers, this is a meaningful step toward closing the gap with hosted speculative-decoding setups.

Research firehose: Apple TIDE, agent skills, diffusion LMs

Akhaliq surfaced a busy paper day. Apple released TIDE ("Every Layer Knows the Token Beneath the Context"), a layer-token method (@_akhaliq). Two agent papers explored skill curation for self-evolving systems — Skill1 unifying skill-augmented agents via RL, and SkillOS on learned skill curation (@_akhaliq). MiA-Signature targets long-context understanding by approximating global activation (@_akhaliq). On the generative side, Continuous Latent Diffusion Language Model, MARBLE (multi-aspect reward balance for diffusion RL), and Continuous-Time Distribution Matching for few-step diffusion distillation all dropped (@_akhaliq). The throughline is clear: agent skill systems and diffusion-style language models are both consolidating into recognizable subfields.

Robotics, health APIs, and benchmark vibes

Hugging Face's Andi Marafioti released a fully open-source backend for Reachy Mini that lets robots run audio models locally and reuse an existing OpenAI or Claude subscription instead of paying $20+/day for realtime APIs — 3,000+ robots adopted it within 48 hours (@andimarafioti via @_akhaliq). Philipp Schmid flagged the new Google Health API shipping with Fitbit Air, exposing 31 data points (sleep, heart rate, SpO2, exercise) plus webhooks for building agents, MCP servers, or CLIs (@_philschmid). And Fabian Stelzer's stress-vibe-test on agent workloads with hundreds of tools crowned DeepSeek V4 Pro the surprise winner — also the cheapest (@fabianstelzer via @clementdelangue).

Pushback on Anthropic's Mythos vulnerability claims

A CNBC report circulated arguing that Anthropic's recent Mythos cybersecurity vulnerability discoveries were oversold: outside experts said the same flaws are findable with existing models, including older Anthropic and OpenAI versions, and Kloc reportedly reproduced the detections with older models on the same codebase (@firstadopter via @clementdelangue). Delangue framed this bluntly as a "manufactured marketing narrative" tied to Anthropic's compute constraints (@clementdelangue). Whether or not that read holds, it's a reminder that capability claims tied to frontier-only access are getting independently retested faster than ever.

The Bottom Line

Anthropic dominated the day on both ends of the spectrum — a credible safety win on Claude blackmail and a steady drumbeat of Claude Code fixes and developer events, against a backdrop of credibility pushback on its Mythos cyber claims. Underneath, China's first formal AI agent policy and an MTP patch for llama.cpp reinforce that the open-source, full-stack-model thesis is accelerating, not slowing.


Sources

Anthropic safety research

@AnthropicAI (×2) · @bcherny · reddit.com/r/ClaudeAI/comments/1ss8h1x/an_open_letter_to_...

Claude Code reliability and SF hackathons

@bcherny (×4) · @ClaudeDevs (×5) · reddit.com/r/ClaudeAI/comments/1siqwmp/anthropic_stop_shi... · reddit.com/r/ClaudeCode/comments/1svjivr/i_think_ill_leav...

China AI agent policy

@clementdelangue (×5) · reddit.com/r/bayarea/comments/1sgvr65/anyone_else_just_ex... · reddit.com/r/LocalLLaMA/comments/1sn3izh/qwen3635ba3b_rel...

Multi-Token Prediction in llama.cpp

@clementdelangue (×4)

Research papers

@_akhaliq (×7)

Robotics, health APIs, benchmarks

@_akhaliq (×2) · @_philschmid · @clementdelangue (×2)

Mythos pushback

@clementdelangue (×2)

Dispatch № 18 · Filed Saturday at dawn from Pensive — a second-brain publication.
Set in Bevan, Old Standard TT, Cormorant Garamond & Courier Prime.