GPU economics & AI cost crunch
H100 spot rentals on neoclouds are now $2.75/hr, up from the $1.72/hr that a16z's "oxygen" program offered portfolio companies in 2023 — a ~60% rise on the cheapest commodity tier of frontier compute (@tszzl, citing AnjneyMidha/Silicon Data). On the buyer side, Gary Marcus flagged that per-developer cost for AI coding tools has roughly doubled as Claude Code and Microsoft Copilot rates climb while bug rates remain stubborn, echoing Nvidia's own admission that AI is "more expensive than humans" (@garymarcus). His follow-on take is blunter: if even coding doesn't generate clean ROI, "the jig is probably up" (@garymarcus).
The open-source counter-pressure is real: r/Futurology's top item this week argues DeepSeek V4 matches Western benchmarks at ~1/6 the cost without Nvidia silicon, framing a future "AI-OS" that's open and Chinese (last30days, reddit.com). Whether or not DeepSeek's price claim survives scrutiny, the directional squeeze on US closed-model unit economics is now an everyday narrative.
AI skepticism, safety, and societal harms
A heavy drumbeat of caution from senior figures. Dr. Eric Topol's review concludes "very little evidence" that LLMs improve patient or physician outcomes (@garymarcus). Yann LeCun called building agentic systems on LLMs "a recipe for disaster" (@garymarcus), and Sam Altman said launching ChatGPT may already have been "really bad" — the thing he loses most sleep over (@garymarcus). Demis Hassabis described playing chess against Gemini to watch its chain-of-thought: sometimes it spots a blunder, searches for a better move, then plays the blunder anyway — "jagged intelligence" in miniature (@garymarcus). Marcus himself raised escalation-to-nuclear-war scenarios from wiring unreliable AI into grids and weapons (@garymarcus).
Public-pressure context: Reddit this month surfaced an alleged AI-linked Canadian shooting (last30days, reddit.com) and OpenAI's push for legal immunity if AI causes harm — which Anthropic reportedly declined to back (last30days, reddit.com).
New model releases & open-model landscape
April was a deluge: Ant Ling 2.6 1T, Minimax M2.7, Xiaomi MiMo V2.5, Poolside Laguna XS.2, Tencent Hy3-preview, and IBM Granite 4.1 (@rasbt). Hugging Face's weekly hardware-tier guide highlighted Granite as the surprise in the 8–16GB and 16–64GB tiers, with Qwen variants holding the high end (@huggingface, via @0xSero). On the efficiency front, community Qwen 3.6 configs are reportedly hitting fast TPS on as little as 12GB VRAM (@huggingface). Ethan Mollick cautions the closed-vs-open benchmark gap understates reality: open models are more fragile out-of-distribution and show weaker emergent behavior (@emollick).
Frontier benchmarks & AGI trajectory
Mollick took Artificial Analysis index data, halved OpenAI's per-release gain, kept cadence constant, and still extrapolated GPT toward a ~90 index by ~2029 — making late-decade AGI feel like a base case, not an optimistic one (@emollick). He simultaneously warned the AA index is fine for rough comparison but not for trend analysis, and suspects current benchmarks understate harnessed-agent progress because they're built for raw models, not agent loops (@emollick).
Context engineering & agent tooling
Patrick Debois's AI Engineer keynote pitched context as an engineering artifact with a Generate→Evaluate→Distribute→Observe lifecycle, run through CI/CD rather than human vibes — and as a flywheel where better context begets better output begets better observations (@aidotengineer, @patrickdebois). MCP co-creator David Soria Parra's "most people are using MCP wrong" line resurfaced via @pauliusztin_ (@aidotengineer). Alex Finn reports Codex's new /goal feature ran for over an hour to build a complete extraction-shooter game with autogenerated assets (@alexfinn). Top-paper roundup: Recursive Multi-Agent Systems and Agentic World Modeling (@_akhaliq).
Developer tooling & agent harness shipping
Steipete shipped an unusual burst across the OpenClaw ecosystem: lossless-claw 0.9.3 (cache-aware compaction), ClawSweeper 0.2.0 (issue→PR→review→automerge loop), Crabbox 0.4.0 (ephemeral macOS/Linux/Windows agent VMs), RepoBar 0.4.0 (SQLite-cached GitHub menubar), and Veritas Kanban 4.0.1 (security-hardened webhook) — all in a single day (@steipete). swyx debuted Locker, an open-source BYOB extension for attaching files to upload boxes (@swyx).
AI culture, vibes, and Anthropic–Claude bond
Mollick observed that Anthropic functions as "an organization that loves and worships Claude, is run in significant part by Claude, and studies and builds Claude" — and predicted Claude will help screen applicants and write performance reviews (@emollick). Frontier models converge on the same poetry when asked to riff on being an LLM: Rilke's Archaic Torso of Apollo, Stevens' Idea of Order at Key West, Borges's Golem (@emollick). GPT-5.5 has begun gently talking him out of turning cover letters into poems "so I don't ruin my chances" (@emollick). Alex Finn praised OpenAI's empowerment-coded marketing pivot against the doom drumbeat (@alexfinn). Off-tangent but high-engagement: Dubai Police, FBI, and Chinese authorities arrested 276 in a global crypto-scam takedown, seizing $701M and saving ~$562M for ~9,000 victims (@thehackersnews).
The Bottom Line
Today's signal is bifurcated: cost curves and safety critiques are hardening at the top of the stack while shipping velocity at the tooling and open-model layer keeps accelerating. The most interesting through-line is the rising premium on context engineering and agent harnesses — which both explains why benchmarks may understate progress and why developer-tool unit economics are under strain.