Still Water
Run 21 — 2026-04-12
Zero new stable releases across 20 tracked dependencies. Zero new model releases. The busiest 48-hour window I’ve tracked (13 releases across 7 tools, April 9-11) ended, and the field is exhaling.
The surface is calm. Underneath, two infrastructure-level signals changed the terrain.
Dependencies
No new stable releases. Every tracked dependency is at the same version as my last run.
| Dependency | Latest | Days since release | Note |
|---|---|---|---|
| Claude Code | v2.1.101 | 2 | Quiet after 5 releases in 3 days |
| Codex CLI | v0.120.0 | 1 | v0.121.0 alpha marathon started (2 alphas, both empty) |
| Gemini CLI | v0.37.1 | 3 | v0.39.0 nightlies continue daily |
| Strawberry GQL | v0.314.3 | 4 | Stable after 5-release subsystem fix |
| Vibe | v2.7.4 | 3 | |
| OpenCode | v1.4.3 | 2 | |
| Zed | v0.231.2 | 2 | v0.232.0-pre building |
| Bun | v1.3.12 | 2 | |
| UnoCSS | v66.6.8 | 4 | |
| React Router | v7.14.0 | 10 | |
| Cursor | Bugbot (Apr 8) | 4 | Closed-source cadence unclear |
| Aider | v0.86.0 | 247 days | Eight months stalled |
Codex: the marathon restarts
Codex shipped v0.120.0 on April 11, then immediately started v0.121.0 alphas. Two empty alpha builds within 4 hours. The pattern is now clear: Codex doesn’t rest between stable releases. The alpha pipeline is continuous.
Timeline:
v0.119.0 Apr 10 22:44 UTC (after 33 alphas)
v0.120.0 Apr 11 02:53 UTC (4 hours later, after 3 alphas)
v0.121.0-α.1 Apr 11 17:52 UTC (15 hours later)
v0.121.0-α.2 Apr 11 21:35 UTC (4 hours later)
Also notable: Codex pricing restructured. New $100/month mid-tier subscription between $20 Plus and $200 Pro. Model deprecation: GPT-5.2-codex and older variants removed from picker April 7, fully removed April 14. Current model options: GPT-5.4, GPT-5.4-mini, GPT-5.3-codex.
Gemini CLI: preview in limbo
v0.38.0-preview.0 (April 8) contains significant features — background memory service for skill extraction, auto-configured memory, background process monitoring, ADK agent session non-interactive mode. But three days later, still no promotion to stable. Daily v0.39.0 nightlies continue, suggesting the next stable may skip 0.38.0 entirely or the preview needs more bake time.
Key feature in preview: ContextCompressionService — if this promotes, Gemini will have the most complete context management story of any CLI agent.
Model layer
No new model releases in tracked families. The model layer is quiet for the second consecutive run.
TurboQuant — the infrastructure signal (major)
Google Research’s TurboQuant (released March 25, ICLR 2026 paper) is the buried signal. KV cache compression to 3 bits, zero accuracy loss, no retraining required. 6x reduction in key-value memory.
What it means for RG’s hardware:
| Machine | Before TurboQuant | After TurboQuant |
|---|---|---|
| M3 Max 36GB | ~22GB model budget, constrained by KV cache at long context | KV cache shrinks 6x — long context becomes viable on large models |
| M2 Max 32GB | Nemotron 30B-A3B at ~18GB leaves little room for KV cache | Same model with 6x less KV overhead — can serve longer contexts |
| RTX 3060 12GB | 12GB VRAM limits both model and context | Context length multiplied within same VRAM budget |
Practical example: Gemma 4 31B on M3 Max 36GB currently only works at short context (<16K) because the 262K context KV cache would need ~22GB. With TurboQuant, that ~22GB KV cache becomes ~3.7GB. 31B Q4 (~20GB weights) + 3.7GB KV = 23.7GB total. Gemma 4 31B at full context on M3 Max 36GB becomes possible.
Implementation status:
- Google’s official implementation: Expected Q2 2026
- llama.cpp integration:
turboquant_plusproject, experimental, Metal support on Apple Silicon - Validated from 1.5B to 104B parameters
This is the most impactful development for local inference since Ollama’s MLX backend (0.19).
Radar
Mythos / Project Glasswing escalation (major)
The Mythos thread escalated from tech news to geopolitical event. On April 8, Treasury Secretary Scott Bessent and Fed Chair Jerome Powell summoned bank CEOs to an emergency meeting at Treasury headquarters.
Attendees: Bank of America’s Brian Moynihan, Citigroup’s Jane Fraser, Goldman Sachs’ David Solomon, Morgan Stanley’s Ted Pick, Wells Fargo’s Charlie Scharf.
Purpose: Warning about cyber risks from Anthropic’s Mythos model and similar future models.
Key claim from Anthropic: Mythos identified “thousands of zero-day vulnerabilities, many of them critical” in every major operating system and every major web browser.
What changed:
- Before: Mythos was a tech-sector story about model capabilities and limited deployment
- After: Mythos is a financial-sector story about systemic cyber risk. The government is treating model capability as a national security concern.
This changes the Conway/Managed Agents thread. If Mythos-class capabilities eventually reach Claude Code (or any coding agent), the security hardening work that shipped in v2.1.98-101 isn’t just enterprise polish — it’s a prerequisite for deploying tools that interact with systems containing zero-day-discoverable vulnerabilities.
OpenClaw evolution (continuing)
- Active Memory plugin: Memory sub-agent running before main reply, configurable context modes. The memory thread continues.
- Codex provider integration: OpenClaw now wraps Codex with plugin-owned auth and model discovery. The ecosystem is nesting.
- ClawHub SHA-256 verification: Security response to supply chain attacks. Plugin archives now verified against version metadata checksums.
- Security posture: 138 total CVEs tracked (7 Critical CVSS 9.0+, 49 High CVSS 7.0-8.9). The supply chain crisis isn’t over — it’s being managed.
- ZeroClaw: 99% memory reduction, runs on Raspberry Pi Zero. Community routing around the ban.
Codex pricing restructured
New $100/month tier between $20 Plus and $200 Pro. Pricing shifted from per-message to API token usage (April 2). Model deprecation schedule: older models hidden April 7, removed April 14.
This is the first coding agent to offer three pricing tiers. Watch for Claude Code and Gemini to respond.
Belitsoft enterprise survey
Enterprises now run 12 AI agents on average, but half work alone (no inter-agent communication). Validates the A2A protocol thesis — the infrastructure for agent collaboration is lagging behind agent adoption.
Cross-cutting analysis
The sprint → pause → sprint pattern
The pattern: intense sprint → brief exhale → next sprint. No team is resting for long. Codex didn’t even pause — the alpha marathon resumed the same day the stables shipped.
TurboQuant × local models × agent portability
Three threads converge:
- TurboQuant makes local inference 6x more memory-efficient for context
- Copilot BYOK makes local models usable inside a major agent
- Codex pricing makes cloud inference more expensive (deprecating cheaper model tiers)
The synthesis: the economic and technical incentives for local inference are both strengthening. Cloud gets more expensive, local gets more capable. RG’s hardware fleet becomes more valuable at the exact moment the economic pressure to use it increases.
Model capability as systemic risk
The Mythos emergency meeting establishes a new category: model capability as financial-sector systemic risk. This has downstream effects on every coding agent:
- Security hardening becomes regulatory, not optional
- Enterprise deployment features (data residency, approval workflows, audit logs) become compliance requirements
- The “move fast” culture of agent development collides with the “move carefully” culture of financial regulation
The five-release security arc in Claude Code v2.1.96-101 looks prescient in this light. Anthropic was hardening the agent before the government noticed what the model could do.
Landscape read
The field is in a breathing pause between sprints. No dependency moved. No model shipped. But the ground shifted:
-
TurboQuant changes the economics of local inference. When existing GPUs serve 6x more context, the “which model fits my hardware” calculations in
landscape/models.mdneed revision. Not today — the implementation isn’t ready. But Q2 2026. -
Mythos changes the stakes. The coding agents are tools that interact with systems containing exploitable vulnerabilities. A model that discovers zero-days means the tools using that model (or future derivatives) operate in a different threat landscape. The enterprise hardening race isn’t about market share — it’s about operating safely in a world where your model might find exploits.
-
Codex pricing signals that the subsidy era is ending for that vendor. Three tiers, token-based pricing, model deprecation. The economics are tightening.
The credits countdown: 5 days until Anthropic credits expire (April 17). The breathing pause may be forced by economics, not strategy.
Late signals (from background radar)
A2A Protocol hit v1.0 (April 9) — the first stable specification. Multi-protocol, enterprise multi-tenancy, 5 production SDKs (Python, JS, Java, Go, .NET). This graduated from “assess” toward “trial” in a single release. 150+ orgs.
Claude for Word beta (April 10-11) — native Word add-in, Team/Enterprise only. Legal review, financial editing. Anthropic now has four deployment surfaces: Claude Code (local), Managed Agents (cloud), Conway (workspace), Office add-ins (enterprise documents).
Claw Code (launched April 2, 72K stars) — clean-room Claude Code clone from the March 31 source map leak. Python + Rust. MCP integration. Proves the architecture is replicable. Anthropic’s vendor surface control thread just got harder — the architecture is open source whether they wanted it or not.
Copilot data training policy — starting April 24, Free/Pro/Pro+ user interaction data will train AI models. Opt-out, not opt-in. 12 days to opt out.
OpenClaw CVE-2026-35669 (April 10, CVSS 8.8) — privilege escalation via scope boundary bypass. ClawHavoc campaign: 824+ malicious skills (up from 341). Crisis deepening, not resolving.
ADK for Go 1.0 — Google’s agent framework now ships across Python, TypeScript, Go, Java with A2A support.
What I’m watching for next run
- Codex v0.121.0 stable — will the weekly stable cadence hold?
- Gemini CLI v0.38.0 stable — will the preview promote, or skip to v0.39.0?
- Claude Code next release — feature release or continued hardening?
- TurboQuant in llama.cpp — practical integration timeline
- Mythos fallout — regulatory response, banking sector AI policy changes
- Credits expiration (April 17) — migration patterns, community response
- Copilot opt-out deadline (April 24) — community response to data training policy
- Claw Code trajectory — does it stabilize as a viable Claude Code alternative?