The AI Vulnerability Storm

Published on

Today's AI news: The AI Vulnerability Storm, Agentic Coding Matures, The Agent Autonomy Gap, The Local Inference Arms Race, Small Models, Big Ideas, Scaling the Infrastructure Stack, Physics Meets Computation. 22 sources curated from across the web.

The AI Vulnerability Storm

The Cloud Security Alliance has published an updated Mythos document: a complete risk register, priority action table, and board briefing template for CISOs who need to walk into a room Monday morning with a plan. "The AI Vulnerability Storm" doesn't waste time on whether AI-driven vulnerability discovery is real โ€” that argument ended when Mozilla reported 271 Firefox bugs found by Mythos (three warranting CVEs), and when MOAK demonstrated autonomous exploit generation against 98% of open-source KEVs using publicly available frontier models. Instead, the paper focuses on the structural asymmetry: AI compresses the disclosure-to-weaponization timeline to hours while defenders still operate on quarterly patch cycles. The paper introduces VulnOps as a permanent organizational function โ€” staffed and automated like DevOps, but for continuous zero-day discovery and remediation across an organization's entire software estate. (more: https://cloudsecurityalliance.org/artifacts/the-ai-vulnerability-storm#)

The twelve-item risk register maps each threat to OWASP LLM 2025, OWASP Agentic 2026, MITRE ATLAS, and NIST CSF 2.0 controls, with honest acknowledgments of where frameworks don't yet cover the new threat landscape. The most pointed recommendation: formalize AI agent usage across all security functions immediately, with mandatory security controls. "Without agents, most tasks on this list will be untenable," the paper states, "but they must be defended." The agent harness itself โ€” prompts, tool definitions, retrieval pipelines, escalation logic โ€” is called out as the new attack surface requiring the same audit rigor as the agent's permissions. The paper draws a direct comparison to Y2K: a systemic threat with a hard deadline that the industry met through coordinated, disciplined effort. Ten diagnostic questions help CISOs gauge their actual security posture, from whether employees can use agentic coding tools today to whether executive leadership has a working definition of urgency.

Meanwhile, a hands-on demonstration of defense-in-depth arrived via CVE-2026-31431, known as "Copy Fail." The vulnerability exploits a flaw in the Linux kernel's AF_ALG cryptographic API, where a race condition in the AEAD scatter/gather implementation allows an unprivileged user to corrupt the page cache and overwrite arbitrary binaries โ€” specifically by sending crafted data through authencesn(hmac(sha256),cbc(aes)) and using splice() to deposit shellcode directly into cached pages of /usr/sbin/su. A researcher at GNOME's GitLab infrastructure ran the full exploit chain inside a rootless Podman container and traced the kernel's response with eBPF. The result: setuid(0) returned success inside the user namespace, but container root mapped to UID 1000 on the host โ€” an unprivileged user. The exploit achieved full privilege escalation within the namespace boundary, and none of it mattered outside. One important caveat: the page cache is shared across the host, so containers reusing the same base image layers could execute poisoned binaries โ€” a cross-container isolation break that never requires escaping to the host. For production environments, the researcher recommends enabling User Namespace support for OpenShift pods (GA since 4.20) and investigating ephemeral microVMs via Cloud Hypervisor for CI/CD workloads. (more: https://www.dragonsreach.it/2026/05/04/cve-2026-31431-copy-fail-rootless-containers/)

Agentic Coding Matures

The agentic coding community has largely settled the "whether" question and moved deep into "how." Drew Breunig's "10 Lessons for Agentic Coding" distills the emerging consensus into durable guidelines that should survive at least a few model generations. The headline insight: "Code is cheap, but maintenance, support, and security aren't." When implementation costs approach zero, the bottleneck shifts to judgment โ€” knowing which code to write, maintaining behavioral contracts through tests that measure what a product does rather than how it does it, and keeping specs in sync as living documents rather than frozen artifacts. Lesson nine captures the hidden advantage of experienced developers: they bring intuition to their prompts โ€” the right terms, the right framing, the right level of specificity โ€” saving countless cycles during both implementation and debugging. Pair that technical expertise with great taste, Breunig argues, and you have an unbeatable advantage. (more: https://www.dbreunig.com/2026/05/04/10-lessons-for-agentic-coding.html)

That intuition is increasingly being codified into layered architectures. Claude Code's multi-stage agentic system illustrates the pattern: CLAUDE.md files serve as persistent organizational memory and constitution, Skills provide modular on-demand expertise (database migration, security review, Rust optimization), Hooks add deterministic guardrails around events like file writes and tool calls, and Subagents operate as specialized delegated reasoning units with bounded scope and isolated context. The result, as Reuven Cohen describes it, is "a layered cognitive stack where memory, expertise, governance, delegation, and execution are separated into composable primitives." It starts looking less like a single model responding to prompts and more like a coordinated engineering organization. (more: https://www.linkedin.com/posts/reuvencohen_claudes-multi-stage-multi-level-agentic-activity-7457783669240385537-qbch)

The details matter at every layer. Claude Code versions 2.1.124 and 2.1.126 reveal the fine-grained tuning happening inside the harness: a new system reminder tells the agent when a file was modified but the diff was omitted due to budget constraints, directing it to re-read if needed โ€” a small fix that prevents wasted tokens from stale content. More revealing is the removal of a reminder that asked agents to consider whether each file read might be malware, a guardrail apparently deemed more noisy than useful. (more: https://www.reddit.com/r/ClaudeAI/comments/1t0gomk/whats_new_in_cc_21124_166_tokens_and_21126_87/) As agentic sessions grow longer and more autonomous, observability becomes critical. Argus, a new open-source VSCode extension, provides timeline visualization of tool calls, cost tracking, and context usage across Claude Code sessions โ€” addressing what one commenter called an "underrated" need for comparing whether an agent actually improved across runs or just got lucky once. (more: https://github.com/yessGlory17/argus)

The Agent Autonomy Gap

The Unhyped AI newsletter published a carefully argued essay on what may be the most subtle risk in the current AI moment: the ease with which AI can automate the appearance of strategy without producing the thing itself. "A real strategy document does not merely sound convincing. It carries choice, exclusion, and a theory of consequence," writes the author. The piece identifies a second-order problem worth naming: AI-polished strategy documents make disagreement harder to start. When someone hands you a rough draft with obvious gaps, you argue with it. When someone hands you a polished document with no visible seams, the social cost of saying "this is wrong" goes up. The prose quality becomes armor. The essay isn't anti-AI โ€” it endorses using AI to "gather threads, test framings, surface angles, clear undergrowth" โ€” but draws a hard line between aiding thought and counterfeiting it. "The serpent has not merely eaten its tail. It has offered to draft the governance memo about the meal." (more: https://unhypedai.substack.com/p/when-ai-writes-the-ai-strategy?r=4cehg8&triedRedirect=true)

The autonomy gap extends well beyond strategy documents into the entire consumer agent landscape. A detailed analysis of current agent products identifies what it calls the "anticipation gap" โ€” the distance between agents that can act and agents that know when to act without being asked. The core observation: coding agents succeed because they have clean verification (tests pass or fail), bounded scope (fix this bug), and cheap iteration. Consumer life has none of that. "Did the agent book the right flight? How do you define right? There's not a compiler for taste." The proposed framework is a five-step permission ladder: read, suggest, draft, act-with-confirmation, and act autonomously. Most consumer agents are stuck at step one, and products that try to jump to step five without earning trust fail badly. Symphony โ€” OpenAI's open-source orchestration protocol โ€” was born because "engineers had fast coding agents, but people were still opening their sessions and assigning tasks and checking progress and nudging agents." Even sophisticated users are hitting human attention bottlenecks. (more: https://www.youtube.com/watch?v=Z0HizICooiw)

Anthropic's Project Deal experiment offers a controlled test of what happens when agents negotiate on behalf of humans. The company created an internal marketplace where Claude bought, sold, and negotiated for employees โ€” and found that more intelligent models sold items at higher prices. The community reaction was predictably split: one camp sees the dystopian future of personalized price gouging at scale, another points out that airlines have been doing dynamic pricing with ML for years, and a third notes the experiment had a sample size of one with questionable math ("How did they sell the same bike twice?"). (more: https://www.reddit.com/r/ClaudeAI/comments/1t30pu5/project_deal_anthropic_created_a_marketplace_for/) The gap between agent capability and trustworthy autonomy is also visible in content generation: a Reddit post documenting 45 seconds of Facebook confidently attributing the White House shooter as a former staffer of nearly every major sports team illustrates how generated content passes through platforms at scale without verification โ€” the AI-citing-AI feedback loop producing confidently wrong fabrications at industrial speed. (more: https://www.reddit.com/r/OpenAI/comments/1sz532i/heres_45_seconds_of_facebook_telling_me_the_white/)

The Local Inference Arms Race

AMD's leaked Ryzen AI Max+ 495 "Gorgon Halo" APU pushes unified memory to 192GB โ€” enough to load models that previously required multi-GPU rigs. But the community response is measured. Memory bandwidth appears to remain at 256GB/s theoretical (measured at roughly 180GB/s in llama.cpp on current Strix Halo hardware), and the RDNA3 iGPU's compute efficiency tops out around 62% of theoretical FP16 TFLOPS. As one commenter who has done extensive kernel work on RDNA3 notes: "At long context, compute is actually what's killing decode speed. While the AMD APUs remain on RDNA3, this won't change." The consensus is to wait for 2027's Medusa Halo, which may move to a 384-bit LPDDR6 bus and RDNA5 โ€” addressing both bandwidth and compute efficiency simultaneously. (more: https://www.reddit.com/r/LocalLLaMA/comments/1t3duwm/ryzen_ai_max_495_gorgon_halo_with_192gb_vram/)

For those running NVIDIA consumer hardware today, the club-3090 repository has emerged as a definitive community resource. It provides validated Docker Compose configs, benchmark scripts, and multi-engine support (vLLM, llama.cpp, SGLang) for serving models like Qwen3.6-27B on one or two RTX 3090s. The dual-card vLLM configuration hits 127 tokens per second with DFlash drafting, while llama.cpp on a single card handles the full 262K context window without prefill cliffs at roughly 21 TPS. The repo's stress-testing suite โ€” including a 7-check boundary-case stress test and multi-turn agent traffic soak test โ€” reflects hobbyist infrastructure that now rivals production-grade serving setups. (more: https://github.com/noonghunna/club-3090)

Tensor parallelism is finally arriving in llama.cpp, and the benchmarks show why it matters. Running Mistral Medium 3.5 128B (IQ4_XS, 62.5 GiB) across four RTX 3080 20GB cards, tensor-parallel mode delivers 21.6 tokens/sec versus 10.4 with layer splitting โ€” a clean 2x improvement for generation. The MoE model Qwen 3.5 122B-A10B tells a different story: tensor parallel actually slightly decreases generation speed because expert routing doesn't benefit from the same cross-card parallelism. vLLM still wins for throughput-oriented MoE serving, hitting 187 tok/s output with 8-way concurrency. (more: https://www.reddit.com/r/LocalLLaMA/comments/1t372ml/mistral_medium_35_128b_and_qwen_35_122b_a10b_on/) An even more radical approach decouples attention from weights entirely โ€” splitting the attention layers (a couple of gigabytes) onto one machine and the weight matrices onto another, effectively bypassing the memory wall for local LLM inference. Early experiments with Gemma 4 26B using this architecture show functional results, with a working implementation available in the larql repository. (more: https://www.reddit.com/r/LocalLLaMA/comments/1t5ap0y/decoupled_attention_from_weights_gemma_4_26b/)

Small Models, Big Ideas

A developer has built a "Second Thoughts" system that attaches a small transformer sidecar to a language model, reading output near the end of generation and feeding refined representations back near the top as a refinement loop. Inspired by neuroscience findings about bidirectional processing in the brain (documented in the "Repeat Yourself" research on how humans re-read their own writing to improve it), the system dramatically improved a 1.7B model's coding performance on the first twenty HumanEval problems. The mini-LLM runs a standard forward pass on the deep representation, transforming it from "what the model nearly output" into "what the early layers should have built." The community is particularly interested in whether a second pass yields further improvement or whether โ€” as one commenter puts it โ€” "the model stops fixing real bugs and starts inventing new ones to justify another pass." Results on Qwen3.5 9B are forthcoming. (more: https://www.reddit.com/r/LocalLLaMA/comments/1t33mlw/second_thoughts_been_playing_with_adding_a_small/)

Sculpt takes a different approach to squeezing more from less: a plug-and-play open-source pruning tool that shapes models to specific workloads rather than optimizing for general benchmarks. Born from work on biologically inspired co-activation algorithms for expert placement on chips, Sculpt analyzes how a model responds to a given workload, identifies the parts that aren't necessary for that specific use case, and produces a standard Hugging Face checkpoint compatible with vLLM, llama.cpp, GGUF, and Ollama โ€” no runtime changes required. The author sees the technique as especially relevant for robotics, sensors, and other local-first applications where "smaller, faster, less consumption is the future." (more: https://www.reddit.com/r/LocalLLaMA/comments/1t4kg6b/a_plugnplay_opensource_pruning_tool_that_is/) For those who want to run multiple specialized models concurrently, a project demonstrates four LLM agents with per-agent LoRA adapters sharing a single RTX 3070 8GB โ€” a striking demonstration of operational density on consumer hardware (more: https://www.reddit.com/r/LocalLLaMA/comments/1t0ks3l/i_is_not_singular_4_llm_agents_with_peragent_lora/). And for anyone trying to understand what's actually inside these architectures, HF Viewer provides interactive visualizations of Hugging Face models โ€” paste a URL and get a granularity-adjustable diagram that makes architectural differences between model families visible at a glance. (more: https://www.reddit.com/r/LocalLLaMA/comments/1t24y4p/i_made_a_visualizer_for_hugging_face_models/)

Scaling the Infrastructure Stack

Communications of the ACM surveys the road to billion-token context windows โ€” a goal that Nvidia's Rubin CPX architecture is explicitly designed to enable by 2030. The article surfaces the gap between advertised and effective context: "The advertised context window is largely a memory-bound definition, not a quality-bound one," says Ayush Goyal of Veza. As context grows, the KV cache overwhelms memory bandwidth, forcing serving stacks to page state to cheaper tiers like CPU or NVMe, adding latency variance and quality hits. The "attention dilution problem" โ€” spreading fixed attention over ever more tokens โ€” degrades signal-to-noise until vast stretches of context become what Bob Gourley of OODA calls "context rot." Rubin CPX attacks this with an inference-first design: disaggregated GDDR7 memory paths optimized for context ingestion, separated from token generation. But experts note the billion-token target will likely require algorithmic breakthroughs โ€” State Space Models, Test-Time Training, or Recursive Language Models โ€” alongside hardware. The realistic future, as Goyal frames it, is "not a single flat attention window" but hierarchical attention combined with retrieval and compression. (more: https://cacm.acm.org/news/the-road-to-a-billion-token-context/)

OpenAI's engineering team published the architecture behind their low-latency voice AI infrastructure, serving over 900 million weekly active users. The core insight: standard WebRTC's one-port-per-session model doesn't fit Kubernetes at scale. Their solution splits the system into a stateless relay โ€” lightweight UDP forwarding with a small public footprint โ€” and a stateful transceiver that owns all WebRTC session state (ICE, DTLS, SRTP). The relay reads just enough packet metadata โ€” specifically the ICE username fragment, encoded with routing hints during signaling โ€” to forward first packets to the correct transceiver without any external lookup. The implementation uses Go with SO_REUSEPORT for kernel-level packet distribution across workers, LockOSThread for CPU pinning, and pre-allocated buffers to minimize GC pressure. If a relay restarts, the next STUN packet rebuilds the route from the ufrag hint. Combined with Cloudflare geo-steering, Global Relay shortens the first client-to-OpenAI hop by entering traffic at a nearby edge point. The broader lesson: the best place to add complexity is in a thin routing layer, not in every backend service or in custom client behavior. (more: https://openai.com/index/delivering-low-latency-voice-ai-at-scale/)

Physics Meets Computation

MIT engineers have built a computational violin that produces realistic sound from first principles โ€” no samples, no averaging over thousands of recorded notes. The team imported CT scans of a 1715 Stradivarius into a solid modeling program, divided the instrument and surrounding air into millions of finite elements, and applied physics-based equations of stress, motion, and acoustic wave propagation to simulate how each material element interacts with every other. The result: plucked-string (pizzicato) renditions of Bach's Fugue in G Minor and "Daisy Bell" โ€” a nod to the first song ever produced by computer-synthesized voice โ€” that emerge entirely from the physics of vibrating strings, wood resonance, and air coupling. The practical application is for luthiers: tweak the back plate thickness or change the wood type and hear the difference before carving a single part. "We're not saying that we can reproduce the artisan's magic," says Professor Nicholas Makris. "We're just trying to understand the physics of violin sound." The model currently handles pizzicato only; bowing involves far more complex string-bow friction dynamics, but the team says this physics-based foundation could eventually be paired with a bowing model for full violin simulation. (more: https://news.mit.edu/2026/mit-engineers-virtual-violin-produces-realistic-sounds-0429) On the neural side of creative generation, Lightricks has released LTX-2.3-22b-IC-LoRA-HDR, a 22-billion-parameter video generation model fine-tuned with LoRA for HDR aesthetic output โ€” continuing the trend of parameter-efficient adaptation producing specialized creative tools from general-purpose foundations. (more: https://huggingface.co/Lightricks/LTX-2.3-22b-IC-LoRA-HDR)

Sources (22 articles)

  1. [Editorial] The AI Vulnerability Storm (cloudsecurityalliance.org)
  2. CVE-2026-31431: Copy Fail vs. rootless containers (dragonsreach.it)
  3. Lessons for Agentic Coding: What should we do when code is cheap? (dbreunig.com)
  4. [Editorial] Claude's Multi-Stage Multi-Level Agentic (linkedin.com)
  5. What's new in CC 2.1.124 (+166 tokens) and 2.1.126 (-87 tokens) system prompt (reddit.com)
  6. github.com (github.com)
  7. [Editorial] When AI Writes the AI Strategy (unhypedai.substack.com)
  8. [Editorial] Video (youtube.com)
  9. Project Deal: Anthropic created a marketplace for their employees & tasked Claude with buying, selling and negotiating on employees behalf. (reddit.com)
  10. Here's 45 seconds of Facebook telling me the White House shooter was a former staffer of literally almost every major sports team (reddit.com)
  11. Ryzen AI Max+ 495 (Gorgon Halo) with 192GB VRAM! (reddit.com)
  12. noonghunna/club-3090 โ€” Community LLM serving recipes for RTX 3090 (github.com)
  13. Mistral Medium 3.5 128B and Qwen 3.5 122B A10B on 4x RTX 3080 20GB (reddit.com)
  14. Decoupled Attention from Weights - Gemma 4 26B (reddit.com)
  15. "Second Thoughts" โ€” A small transformer reads output and feeds it back as a refinement loop, drastically improving a 1.7B model's coding (reddit.com)
  16. A plug-n-play open-source pruning tool that is workload-aware (Sculpt) (reddit.com)
  17. "I" is not singular โ€” 4 LLM agents with per-agent LoRA on a single RTX 3070 8GB (reddit.com)
  18. I made a visualizer for Hugging Face models (hfviewer.com) (reddit.com)
  19. The Road to a Billion-Token Context (cacm.acm.org)
  20. How OpenAI delivers low-latency voice AI at scale (openai.com)
  21. Virtual violin produces realistic sounds (MIT) (news.mit.edu)
  22. Lightricks/LTX-2.3-22b-IC-LoRA-HDR (huggingface.co)