Local models nail structure: Agents without the mystery box

Published on

Structured outputs under local pressure A simple but telling LM Studio test of JSON Schema adherence across local models showed Qwen3 variants consistently hitting 98–100% pass rates over 50 runs, ...

Local models nail structure

Structured outputs under local pressure

A simple but telling LM Studio test of JSON Schema adherence across local models showed Qwen3 variants consistently hitting 98–100% pass rates over 50 runs, while some popular names stumbled badly. Google’s Gemma 3 27B and Mistral’s magistral-small also scored 100%, but OpenAI’s GPT-OSS 20B and 120B cratered (2% and 0%, respectively) with incomplete responses, schema violations, and timeouts; Nvidia’s Nemotron Nano 9B v2s similarly failed with 50 incomplete responses. The author used a basic “rate these jokes” schema and published the tester script. Practitioners in the thread stressed that if you truly require valid structure, use an engine that enforces it at decode time (guided JSON/regex grammars in vLLM or llama.cpp), and note LM Studio’s idiosyncratic handling of GPT‑OSS via OpenAI’s Harmony library, which can misalign response formats unless the schema is described up front. A llama.cpp example with json_schema shows GPT‑OSS can comply when the inference stack supports it. (more: https://www.reddit.com/r/LocalLLaMA/comments/1of3r61/test_results_for_various_models_ability_to_give/)

Performance tuning matters as much as adherence. One practitioner shared Mixture-of-Experts configs for vLLM that are specific to the model’s “shape” and GPU, turning Qwen3 Coder REAP 25B from 10+ minute stalls into concurrent 25k‑token requests at ~45 tok/s on an RTX Pro 6000 Blackwell—by supplying the proper E=, N=, device_name JSON files and trimming vLLM’s benchmarking sweep. Repo included. (more: https://www.reddit.com/r/LocalLLaMA/comments/1oi16jj/vllm_moe_benchmark_configs_for_qwen3_coder_reap/)

Meanwhile, Ollama v0.12.7-rc0 added local Qwen3‑VL support from 2B to 32B, drawing some community ire over perceived lack of credit to upstream llama.cpp contributors—a reminder that the open ecosystem’s velocity still leans heavily on upstream merges and attribution norms. (more: https://www.reddit.com/r/LocalLLaMA/comments/1oj2ut1/ollama_supports_qwen3vl_locally/) On-device efficiency also advanced: Meta introduced ExecuTorch 1.0 with KV cache quantization and custom SDPA/KV optimizations—important dials for fitting bigger context windows and faster decoding on constrained hardware. (more: https://www.reddit.com/r/LocalLLaMA/comments/1odg1wm/introducing_executorch_10/)

Agents without the mystery box

Agent frameworks chase control and clarity

Fed up with “too abstract” or “too bare” agent frameworks, a team released Flo AI, a Python framework with OpenTelemetry tracing, multi-agent collaboration (“Arium”), YAML-driven customization, and vendor-agnostic backends (OpenAI, Anthropic, Google, Ollama, vLLM, Vertex AI). The value proposition is observability and composability without lock-in. Feedback was frank: documentation needs work, the space is saturated, and integration to end-user software (e.g., an OpenAI-compatible streaming chat UI) is what ultimately determines adoption. The team says they’ll focus on integrations and an open-source platform on top. (more: https://www.reddit.com/r/LocalLLaMA/comments/1ofq2g3/open_source_we_deployed_numerous_agents_in/)

Skills ecosystems are spreading beyond single vendors. OpenSkills is a CLI that syncs Claude Code “skills” into other coding agents by injecting them as prompts—the author notes Claude Code itself simply lists tools in the system prompt and offers a “load skills” tool; there’s no hidden logic. Skeptics asked about missing intent detection, and an alternative project, Skillz, was cited as a Model Context Protocol (MCP) alternative that loads skills on demand. The point stands: skills are prompt assets and can travel, but orchestration guards still matter. (more: https://www.reddit.com/r/ChatGPTCoding/comments/1oh7d6t/openskills_cli_use_claude_code_skills_with_any/) In the same vein, a community thread shows “Claude Skills” running locally in Apple’s container environment, underscoring portability trends. (more: https://www.reddit.com/r/LocalLLaMA/comments/1ofnm3p/claude_skills_but_running_locally_in_apple/) For prompt hygiene, a Claude Code “prompt improver” hook intercepts vague prompts, builds a research plan, and asks 1–6 grounded clarifying questions via AskUserQuestion before proceeding—minimal overhead, manual install due to a plugin bug, and bypass prefixes if you want raw prompts. (more: https://github.com/severity1/claude-code-prompt-improver) Anecdotally, Sonnet 4.5 users are using “rules” to tame yes‑man behavior during debugging; mileage varies, but codifying expectations helps. (more: https://www.reddit.com/r/ClaudeAI/comments/1oggr03/prompts_avoiding_yes_men_moments/)

One more agentic thread: Coyote, pitched as an agent that “feels like texting a friend” with native async tools, popped up then was removed; a commenter noted VRAM limits on an 8GB laptop—a reminder that UX ideas still live or die on deployment constraints. (more: https://www.reddit.com/r/LocalLLaMA/comments/1oeidck/built_coyote_an_ai_agent_that_feels_like_texting/)

Dev environments are the new perimeter

AI coding assistants widen the attack surface

Security leaders are increasingly blunt: the developer workstation is now part of the attack chain. The recent disclosure of five more RCE CVEs in the Cursor AI editor within a week is a canary for a broader shift. AI coding assistants and agent-enabled IDEs sit inside CI/CD trust boundaries, mediate network access through MCP (Model Context Protocol) servers and other extensions, and can both write and execute code. A malicious prompt, compromised MCP server, or lax OAuth/extension can pivot into a full breach. If assistants are “junior engineers,” they’re junior engineers with shell access—and that changes program risk, not just developer ergonomics. (more: https://www.linkedin.com/posts/gadievron_five-more-cursor-cves-in-a-week-ai-coding-activity-7383397198547480577-2DmY)

Operational takeaways align with the agent news: default to least privilege for agent tools, aggressively log and trace (OpenTelemetry in frameworks like Flo AI can help), isolate API keys, and treat local MCP servers as untrusted network services. Guided outputs and deterministic tool interfaces reduce surprising behaviors; explicit allowlists beat clever prompting.

When hardware trust breaks

Breaking TEEs and listening from orbit

A new attack, TEE.fail, shows that a cheap DDR5 interposer on the memory bus can undermine the security of major Trusted Execution Environments. Because Intel and AMD TEE memory encryption as deployed is deterministic, an attacker observing DRAM traffic can exploit ciphertext equality to extract keys. The researchers demonstrate key extraction against Intel TDX and AMD SEV‑SNP with Ciphertext Hiding, including recovery of ECDSA attestation keys from Intel’s Provisioning Certification Enclave from a single signing operation—collapsing SGX/TDX attestation. Since Nvidia GPU confidential computing chains trust CPU CVMs for attestation, compromising CPU keys cascades to GPUs too. The interposer uses a single DDR5 channel and off‑the‑shelf parts, with cost under $1,000. Attestation—long the bedrock of “confidential computing”—needs a rethink. (more: https://tee.fail/)

At the comms layer, researchers scanned IP traffic downlinks on 39 geosynchronous satellites with a consumer dish and found significant unencrypted flows: cellular backhaul (including T‑Mobile, which enabled encryption after disclosure), airline internet traffic, VoIP, even corporate VPN data with credentials visible. The team published tools for DVB‑S2(X) captures; the broader lesson is boring but vital—assume interception, encrypt everything, and remember integrity protections matter as much as confidentiality. (more: https://hackaday.com/2025/10/27/satellite-snooping-reveals-sensitive-unencrypted-data/)

OTA meets safety reality

Jeep’s OTA misstep triggers emergency recall

An over‑the‑air update intended for the telematics system on 2023–2025 Jeep Wrangler 4xe plug‑in hybrids caused incomplete communication between the Telematics Box Module and Hybrid Control Processor, triggering HCP resets while driving and sudden loss of motive power—no warning. Jeep issued what may be a record-fast safety recall and rolled back the update, restoring drivability while engineers work on a permanent fix. OTA is now a standard patch path for vehicles, but this incident spotlights the operational risks when safety‑critical modules depend on correctly orchestrated software in the field. Rollback capabilities and staged rollouts aren’t nice-to-haves; they’re life-safety controls. (more: https://www.thedrive.com/news/jeep-issues-emergency-recall-for-ota-bricked-wrangler-4xes)

Multimodal and simulated users advance

Better retrieval, richer audio, smarter NPCs

A new arXiv paper brings test‑time scaling to generative retrieval for multimodal, multi‑turn conversational product search. The authors pair an MLLM retriever with test‑time reranking that dynamically adjusts candidate scores as user intent evolves across turns—addressing the ambiguity and catalog grounding that make standard TTS techniques brittle here. Across curated multi‑turn multimodal datasets (including refined Multi‑turn Fashion IQ, MMD, and a synthetic MUSE set), they report average gains of 14.5 MRR points and 10.6 nDCG@1 over strong baselines. The result is not just better first hits, but better adaptation during dialogue. (more: https://arxiv.org/abs/2508.18132v1)

On the audio frontier, Nvidia’s Audio Flamingo 3 is a fully open large audio‑language model (non‑commercial license) built on AF‑Whisper, an MLP adaptor, and a Qwen2.5‑7B backbone, with long‑context comprehension up to 10 minutes, flexible chain‑of‑thought, multi‑turn voice chat, and streaming TTS. It reports state‑of‑the‑art results across 20+ audio understanding and reasoning tasks, targeting A100/H100 inference. (more: https://huggingface.co/nvidia/audio-flamingo-3-hf) In games, Distil‑Labs fine‑tuned Gemma 3 at 270M and 1B into “Distil NPC” models that respond in character as non‑playable characters, moving past rigid dialogue trees to natural language interactions—with qualitative examples showing stylistic, lore‑consistent answers versus base models’ literal echoes. (more: https://www.reddit.com/r/ollama/comments/1oe96b6/distil_npc_family_of_slms_responsing_as_npcs/)

Microsoft Research flipped the script with UserLM‑8b, trained to simulate the “user” side of a conversation from WildChat data. Given a task intent, it generates first turns, follow‑ups, and an end‑conversation token, enabling more realistic assistant evaluation. Reported results show lower perplexity than prior simulators and better adherence to “user” behaviors across six intrinsic metrics; guardrails such as filtering first tokens and avoiding premature termination were necessary. It’s research‑only, not an assistant, and can hallucinate additional constraints—useful for diversity, risky for controlled evaluation. (more: https://huggingface.co/microsoft/UserLM-8b)

Infrastructure: hubs, serialization, APIs

Foundations for open ML and fast data

Hugging Face announced huggingface_hub v1.0, marking five years of building the backbone of open machine learning distribution and versioning. The hub’s evolution into 1.0 signals API stability for the ecosystem that underpins countless local and cloud workflows. (more: https://huggingface.co/blog/huggingface-hub-v1)

At the systems layer, Apache Fory Rust touts 10–20x faster serialization than JSON/Protobuf with cross‑language binary compatibility (Rust, Java, Python, C++, Go), automatic handling of circular references and trait objects, and schema evolution without IDL or codegen steps beyond Rust derive macros. It supports both object-graph and row formats, reference deduplication, and compile‑time codegen for speed and safety—positioning it for polyglot microservices that shuttle rich graphs at high throughput. Licensed under Apache‑2.0. (more: https://fory.apache.org/blog/2025/10/29/fory_rust_versatile_serialization_framework/) On the application edge, a “Vision‑Detection‑API” repo surfaced—details are sparse, but it’s another sign of turnkey vision stacks being packaged for quick integration. (more: https://github.com/snowyfizz/Vision-Detection-API)

These infrastructural moves matter because agentic apps and multimodal models increasingly hit serialization bottlenecks and model‑distribution complexities long before they hit model‑quality ceilings.

Crime, connectivity, and control

Myanmar raids KK Park cybercrime hub

Myanmar’s military says it dismantled a major scam complex at KK Park near the Thai border, detaining 2,198 people and seizing equipment including 30 Starlink terminals. State media cited over 260 unregistered buildings; the compound sits near Myawaddy, an area under mixed control. Officials framed the raid as part of a campaign against online fraud and cross‑border cybercrime ongoing since early September and referenced local armed actors in the area’s control dynamics. Photos published showed Starlink equipment and troops during the operation. (more: https://apnews.com/article/scam-centers-cybercrime-myanmar-a2c9fda85187121e51bd0efdf29c81da)

The seizure underscores a recurring theme: connectivity tools—from satellite internet to LLM‑powered agents—are dual‑use. Governance, attribution, and robust attestation become prerequisites, not afterthoughts, when infrastructure is both borderless and potent.

Sources (22 articles)

  1. [Editorial] https://tee.fail/ (tee.fail)
  2. [Editorial] Developer machine part of attack chain (www.linkedin.com)
  3. vLLM MoE Benchmark Configs for Qwen3 Coder REAP 25B & RTX Pro 6000 (www.reddit.com)
  4. Ollama supports Qwen3-VL locally! (www.reddit.com)
  5. [Open Source] We deployed numerous agents in production and ended up building our own GenAI framework (www.reddit.com)
  6. Test results for various models' ability to give structured responses via LM Studio. Spoiler: Qwen3 won (www.reddit.com)
  7. Claude Skills but running locally in Apple container (www.reddit.com)
  8. Distil NPC: Family of SLMs responsing as NPCs (www.reddit.com)
  9. OpenSkills CLI - Use Claude Code Skills with ANY coding agent (www.reddit.com)
  10. Prompts avoiding Yes Men moments? (www.reddit.com)
  11. severity1/claude-code-prompt-improver (github.com)
  12. snowyfizz/Vision-Detection-API (github.com)
  13. Show HN: Apache Fory Rust – 10-20x faster serialization than JSON/Protobuf (fory.apache.org)
  14. Jeep Issues Emergency Recall for OTA-Bricked Wrangler 4xes (www.thedrive.com)
  15. Myanmar military shuts down a major cybercrime center, detains over 2k people (apnews.com)
  16. nvidia/audio-flamingo-3-hf (huggingface.co)
  17. microsoft/UserLM-8b (huggingface.co)
  18. Satellite Snooping Reveals Sensitive Unencrypted Data (hackaday.com)
  19. Test-Time Scaling Strategies for Generative Retrieval in Multimodal Conversational Recommendations (arxiv.org)
  20. huggingface_hub v1.0: Five Years of Building the Foundation of Open Machine Learning (huggingface.co)
  21. Built Coyote — An AI Agent That Feels Like Texting a Friend and released first model supporting native Async Tools (www.reddit.com)
  22. Introducing ExecuTorch 1.0 (www.reddit.com)