Local GPUs hit real limits: Multimodal speech: promise potholes
Published on
Running bigger local LLMs is still more about bandwidth than bragging rights. One user’s rig—1× 5070 Ti on PCIe 5 x16 and 3× 5060 Ti 16 GB on PCIe 4 x4—drives 20–24 tokens/sec on 30k context...
Local GPUs hit real limits
Running bigger local LLMs is still more about bandwidth than bragging rights. One user’s rig—1× 5070 Ti on PCIe 5 x16 and 3× 5060 Ti 16 GB on PCIe 4 x4—drives 20–24 tokens/sec on 30k context but strains when pushing gpt-oss-120B and GLM-4.5 air at 40k context without spilling to system RAM. The community’s diagnosis: VRAM matters, but interconnect bandwidth and PCIe topology are the first cliffs you’ll fall off. Keeping entire weights plus KV cache in GPU memory avoids the catastrophic slowdown of system RAM; once you spill, PCIe becomes the bottleneck. (more: https://www.reddit.com/r/LocalLLaMA/comments/1o7x3x1/should_i_add_another_5060_ti_16gb_or_two_already/)
Even quantized models differ wildly in footprint and context costs. One commenter argued GLM-4.5 air at 40k context needs roughly 75–80 GB for weights+KV, pushing past a 64 GB pool; others countered that gpt-oss-120B in mixed formats (MXFP4+FP16) can be around 65 GB with lower context needs. Either way, “barely fits” configurations leave no headroom for simultaneous RAG pipelines competing for bandwidth. If you need it now, adding two more 16 GB cards gives breathing room; if you can wait, next-gen parts may collapse this workload onto fewer GPUs with fewer PCIe compromises. (more: https://www.reddit.com/r/LocalLLaMA/comments/1o7x3x1/should_i_add_another_5060_ti_16gb_or_two_already/)
Server-class alternatives are in flux. Fresh DGX Spark benchmarks surfaced this week—useful datapoints if you’re weighing NVLink-class systems over DIY PCIe builds—while separate commentary dubbed Spark “great hardware” but early in its ecosystem maturity. Translation: don’t expect a turnkey advantage without software that fully exploits the fabric. (more: https://www.reddit.com/r/LocalLLaMA/comments/1o6t90n/nvidia_dgx_spark_benchmarks/) (more: https://simonwillison.net/2025/Oct/16/claude-skills/)
Multimodal speech: promise, potholes
Llama.cpp’s new audio hooks work—but model choice and prompting matter. A tester got qwen2.5-omni (3B) ingesting WAV (and MP3) via llama-server, while a smaller Voxtral-1B struggled. Ultravox-0.5-8B transcribed a simple MP3, but failed Simon Willison’s “pelican joke” test: instead of transcribing, it followed instructions embedded in the audio. The fix isn’t magic: pick models trained for speech tasks, route through server modes that force “transcription-only,” and assume smaller generalist models will drift. (more: https://www.reddit.com/r/LocalLLaMA/comments/1o93ad1/audio_transcription_with_llamacpp_multimodal/)
On the generative side, a fully automated podcast pipeline landed: Ollama for script generation, Piper for TTS, push-button from topic to audio. It’s open source and simple to extend, though feedback flagged quality and a GitHub typo. As with most “AI radio” experiments, the bottleneck is less plumbing and more editorial quality control. (more: https://www.reddit.com/r/ollama/comments/1o6c81r/i_built_a_fully_automated_ai_podcast_generator/)
Agent architectures meet reality
Two useful mental models for agent design: flowcharts vs handoffs. Flowcharts prewire a fixed decision graph; handoff-based systems let any agent pass control (and full conversation history) to any other, constructing a dynamic call graph on the fly. The latter avoids combinatorial diagram sprawl at the cost of more runtime orchestration, but it better matches messy, evolving tasks. (more: https://www.reddit.com/r/LocalLLaMA/comments/1o5f6md/flowchart_vs_handoff_two_paradigms_for_building/)
On the platform side, pairing Claude Agent SDK with Cloudflare’s Containers + Workers + Durable Objects divides labor intelligently: do fast context building in Workers, decide early if a request even needs an agent, and spin up the container only when it does. You get cheaper triage and lower cold-start pain. A public repo shows the plumbing. (more: https://www.reddit.com/r/ClaudeAI/comments/1o9le8o/claude_agent_sdk_cloudflare_containers_is_the/)
Agentic plumbing is also getting opinionated. “Turbo Flow” provides a one-script dev environment for Claude Code atop Claude Flow and Agentic Flow, prepacking 610+ sub-agents, upgraded Playwright as a Model Context Protocol (MCP) server, ReasoningBank support, aliases, and a wizard that generates project-tuned configuration—all labeled alpha-quality. The adjacent “Agentic Flow” stacks agents on QUIC streams for concurrent “cognitive threads” with 0-RTT resumptions—ambitious for distributed reasoning, though adoption will hinge on real-world stability and security. (more: https://www.linkedin.com/posts/marcuspatman_agenticops-agenticai-ai-activity-7384809986796789761-yIYJ) (more: https://www.linkedin.com/posts/reuvencohen_what-if-the-internet-could-think-embedding-activity-7384935972330741760-SxCE)
Practical tactics still win. A developer released “Crystal” to run Claude Code and Codex side by side on the same prompt, surfacing different solution paths for the human to choose. Meanwhile, a tool claiming “100% Autonomous (Complex) Coding” drew the expected pushback: UI testing and end-to-end validation are what make autonomy real, not the agent count. The broader enterprise view echoes that: production agents require planning, memory hierarchies, tool orchestration, robust infra, and rigorous monitoring—far beyond “ChatGPT + vector DB.” (more: https://www.reddit.com/r/ChatGPTCoding/comments/1o85mk3/compare_claude_code_and_codex_from_one_prompt/) (more: https://www.reddit.com/r/LocalLLaMA/comments/1o4vnsw/claudiomiro_how_to_achieve_100_autonomous_complex/) (more: https://www.linkedin.com/posts/armand-ruiz_over-the-last-18-months-ai-agents-and-activity-7384906047557230592-irrf)
Claude Skills: simple, powerful, risky
Anthropic’s new “Skills” are folders—SKILL.md plus optional scripts and resources—that Claude can auto-detect and load on demand. They’re token-efficient: a harness scans minimal YAML summaries at session start and only loads full details if relevant. Skills stack, work across Claude apps, Claude Code, and the API, and can include executable code for tasks better done deterministically than via token generation. Anthropic ships a “skill-creator,” versioning APIs, a marketplace, and first-party skills for Excel, PowerPoint, Word, and PDFs. Availability spans Pro, Max, Team, and Enterprise, gated by a secure beta environment. (more: https://www.anthropic.com/news/skills)
Simon Willison’s deep dive argues Skills might be “a bigger deal than MCP” (Model Context Protocol) because they pack guidance in plain text and scripts without MCP’s orchestration overhead and token tax—some MCP stacks burn tens of thousands of tokens just to describe tools. He shows a Slack GIF creator skill that validates file size to fit Slack’s 2 MB cap, and notes Skills’ portability: nothing stops other coding agents from reading the folder and following instructions. The catch is the same as any code-enabled agent: you need a safe execution environment and sandboxing against prompt injection and supply chain risks. Expect a Cambrian explosion of shareable Skills—and the corresponding need for guardrails. (more: https://simonwillison.net/2025/Oct/16/claude-skills/)
New models and datasets matter
A team released what they call the largest open-source five-modality dataset: over 100M automatically matched quintuples spanning caption, image, video, audio, and point clouds; a 1M-pair human-rated subset; and a 3.5K consensus-based eval set for zero-shot audio↔point cloud retrieval. Data assembly used nearest-neighbor retrieval in modality-specific embedding spaces, followed by clustering and greedy sampling to ensure diversity and reduce overlap. Cleaning includes corruption checks, NSFW filtering for annotators, license reporting, and benchmark protection by excluding known eval items from training. A baseline joint embedding model shows strong cross-modal retrieval with gains from fine-tuning on the rated subset, and the authors enumerate headroom: full-token attention, quality-weighted objectives, and data augmentations. (more: https://e-mm1.github.io/)
On identity-preserving image generation, ByteDance’s FaceCLIP learns a shared embedding that fuses facial identity and textual semantics, then guides SDXL and FLUX-based generators. Rather than bolting on adapters, the joint ID–text representation plus multi-modal alignment loss improves photorealism and identity retention relative to prior approaches, according to their evaluations. The models ship under a non-commercial research license. (more: https://huggingface.co/ByteDance/FaceCLIP)
Open weights also advanced beyond English. KORMo-10B is a 10.8B bilingual Korean–English model trained from scratch on roughly 3.7T tokens, with all code, data, and checkpoints open under Apache 2.0. The team publishes detailed bilingual benchmarks—competitive English scores for the size class and notably strong Korean results across CLICK, Haerae, KoBEST, and KMMLU. It includes a “thinking mode,” but the authors caution it’s not safety-tuned or preference-aligned yet. (more: https://huggingface.co/KORMo-Team/KORMo-10B-sft)
Trust the stack, verify everything
ReliaQuest details a year-long ArcGIS server compromise by the China-linked Flax Typhoon group. The actors modified a legitimate ArcGIS Server Object Extension into a covert web shell that accepted base64-encoded commands via REST parameters, gated by a hardcoded secret. Deployed using valid admin credentials, the implant blended with routine operations—and restoring from backups reintroduced the malicious component. It’s a classic “living off the land” play: no signature-based malware, deep persistence, and trust subversion in a ubiquitous enterprise tool. Backups aren’t a failsafe if they contain the backdoor. (more: https://www.theregister.com/2025/10/14/chinese_hackers_arcgis_backdoor/)
For defenders and red teamers mapping internal risk, TaskHound enumerates Windows scheduled tasks that run with privileged accounts, highlights Tier-0 context (e.g., Domain/Enterprise Admins), and exports to Legacy and Community BloodHound formats. It analyzes password age relative to task creation for DPAPI dump viability, resolves SIDs, supports “offline XML” parsing, and even includes a BOF for AdaptixC2. It’s powerful—and explicitly warns about OPSEC and legal use. (more: https://github.com/1r0BIT/TaskHound)
On the services side, go-authkit provides a practical authn/z toolkit for Go: JWT middleware that caches claims via LRU, automatic JWKS discovery and caching, OAuth 2.0 token introspection (HTTP or gRPC), pluggable scope decoders (including Acronis-specific formats), and token providers to securely call introspection endpoints. It centralizes common patterns and reduces bespoke, error-prone security code. (more: https://github.com/armai92/goauth)
LinkedIn’s feed goes semantic
A widely shared editorial breaks down LinkedIn’s own research: replacing a patchwork of systems with a single LLM-powered retrieval model. A fine-tuned LLaMA‑3 maps members and posts into the same embedding space, using your profile, industry, skills, and past engagement as prompts. Numerical performance signals are quantized and embedded too. The system—dubbed GPU‑RAR, “retrieval‑as‑ranking”—indexes 60M items, refreshes every 30 minutes, serves nearest‑neighbor retrieval in under 50 ms, and shows measurable engagement and revenue lifts. Notably absent: knowledge graphs or ontologies—the “Economic Graph” gets subsumed into learned semantics. The practical advice: optimize for genuine engagement, not posting rituals. (more: https://www.linkedin.com/posts/stuart-winter-tear_linkedin-feed-using-causal-language-models-activity-7384874061597782016-aLKU)
When radio was the app store
Before “write once, run anywhere” had a JVM, the Netherlands had BASICODE: a standardized subset of BASIC plus machine-specific runtimes that normalized hardware differences across 8-bit computers. Programs and text “journals” were broadcast over FM radio and recorded to cassette, then loaded through the runtime—free, mass software distribution without disks or modems. Early versions forbade graphics, sound, or direct storage to guarantee portability; later revisions added monochrome graphics, sound, and storage handling. Adoption was strongest in the Netherlands, with some use in Germany’s WDR TV. The conceptual echo is clear: a strict, documented interface plus per-platform shims to bridge heterogeneity—prosaic, effective, and oddly modern. (more: https://hackaday.com/2025/10/14/basicode-a-bit-like-java-but-from-the-1980s/)
Sources (21 articles)
- [Editorial] Claude Skills are awesome, maybe a bigger deal than MCP (simonwillison.net)
- [Editorial] Claude Skills (www.anthropic.com)
- [Editorial] Chart a path (www.linkedin.com)
- [Editorial] LinkedIn Alogrithm (www.linkedin.com)
- [Editorial] Agentic Flow (www.linkedin.com)
- [Editorial] Turbo Flow (www.linkedin.com)
- Claudiomiro: How to Achieve 100% Autonomous (Complex) Coding (www.reddit.com)
- Flowchart vs handoff: two paradigms for building AI agents (www.reddit.com)
- NVIDIA DGX Spark Benchmarks (www.reddit.com)
- Should I add another 5060 Ti 16GB or two? Already had 1 x 5070 Ti and 3 x 5060 Ti 16G (www.reddit.com)
- Audio transcription with llama.cpp multimodal (www.reddit.com)
- I built a fully automated AI podcast generator that connects to ollama (www.reddit.com)
- Compare Claude Code and Codex from one prompt (www.reddit.com)
- Claude Agent SDK + Cloudflare Containers is the perfect agent platform (www.reddit.com)
- 1r0BIT/TaskHound (github.com)
- armai92/goauth (github.com)
- Chinese gang used ArcGIS as a backdoor for a year – and no one noticed (www.theregister.com)
- Show HN: Largest open-source multimodal AI dataset (e-mm1.github.io)
- KORMo-Team/KORMo-10B-sft (huggingface.co)
- ByteDance/FaceCLIP (huggingface.co)
- BASICODE: A Bit Like Java, But From The 1980s (hackaday.com)