AMD-first LLM inference push: Tiny models big retrieval gains
Published on
Open-weight models meet AMD-first engineering. OpenAI’s GPT-OSS 20B and 120B landed with immediate support across popular stacks, but one team went much further: “gpt-oss-amd,” a pure C++ HIP im...
AMD-first LLM inference push
Open-weight models meet AMD-first engineering. OpenAI’s GPT-OSS 20B and 120B landed with immediate support across popular stacks, but one team went much further: “gpt-oss-amd,” a pure C++ HIP implementation that avoids rocBLAS/hipBLAS/RCCL/MPI and optimizes everything from FlashAttention to matrix-core GEMM, multi-streaming, and MoE routing. On a single node with 8× MI250s, they report 30k tokens/sec on 20B and nearly 10k on 120B in custom benchmarks—evidence AMD hardware can compete on large-scale inference when software is tuned end-to-end. Caveat: current kernels depend on AMD Matrix Cores (MI250 and similar). There’s a legacy path for older GPUs (e.g., MI50) without Matrix Cores, but it runs slower; consumer cards like 7900 XTX may handle 20B but likely not 120B in full precision. The code aims for “llama2.c”-style readability, and users report interest in quantized paths and API layers next. (more: https://www.reddit.com/r/LocalLLaMA/comments/1o3dfib/gptoss_from_scratch_on_amd_gpus/)
The timing aligns with AMD’s larger AI tailwinds. Bloomberg notes OpenAI agreed to buy tens of billions of dollars of AMD chips; AMD’s stock jumped as investors priced in the partnership’s scale. If AMD’s supply comes online as promised, a broader ROCm software push could follow—exactly what frustrated practitioners ask for when trying to run mixed fleets across MI100/MI250 and RDNA cards. (more: https://www.bloomberg.com/opinion/newsletters/2025-10-06/openai-is-good-at-deals)
Serving tactics matter as much as silicon. Dynamic batching layers like batchi show how to cap latency while packing requests over Unix/TCP, isolating invalid jobs so they don’t poison a whole batch. It’s a proof of concept rather than production, but the pattern is what you want for bursty traffic and batch inference economics. (more: https://github.com/adb1274/batchi) And the cost math is still “it depends.” For bursty, intermittent usage, serverless APIs often win; for steady, high-throughput workloads, provisioned hardware plus an engine like vLLM tends to be cheaper. Frequency thresholds in the “every 30–60 seconds” range are one team’s rule-of-thumb for when to switch. (more: https://www.reddit.com/r/LocalLLaMA/comments/1o3ttlo/how_do_i_compare_cost_per_token_for_serverless_vs/)
Model support is steadily normalizing. GPT-OSS runs across vLLM and SGLang; Meituan’s LongCat-Flash-Chat ships deploy adapters for both; llama.cpp remains the fallback on integrated GPUs. The momentum is toward engines that can exploit AMD’s matrix cores when available, degrade gracefully when not, and still deliver decent context and throughput. (more: https://huggingface.co/meituan-longcat/LongCat-Flash-Chat)
Tiny models, big retrieval gains
Small can be surprisingly strong—especially for retrieval. NeuML’s ColBERT Nano series shrinks late-interaction retrieval into sub‑1M parameter models (≈250K, 450K, 950K). By generating multi‑vector embeddings, they enable CPU‑only or on‑device search with limited compute, and can be specialized easily using datasets like FineFineWeb. They’re not chasing leaderboard SOTA like BGE; they’re targeting accuracy-per-watt and deployability. They integrate with txtai via trust_remote_code, and fit edge scenarios where bandwidth, latency, or privacy constrain heavier models. (more: https://www.reddit.com/r/LocalLLaMA/comments/1o1mpt5/introducing_the_colbert_nano_series_of_models_all/)
On the generation side, Liquid AI’s LFM2-8B-A1B uses a hybrid MoE with 8.3B total and just 1.5B active parameters to deliver speed on phones, tablets, and laptops once quantized. Architecture mixes 18 short-range convolution blocks with 6 attention blocks, long context (32k), and BF16/FP8 training. The authors recommend fine-tuning for narrow use cases—agentic tasks, RAG, extraction, creative writing—while cautioning against knowledge- or code-heavy tasks. Benchmarks show instruction following and math competitive with larger dense peers, and the repo includes HF transformers, vLLM, and llama.cpp instructions. (more: https://huggingface.co/LiquidAI/LFM2-8B-A1B)
The broader lesson from practitioners: tiny, specialized models are underrated. For tightly scoped tasks, a 1M‑parameter retriever or a 1.5B‑active MoE can beat a 70B generalist by being close to the data and cheap enough to deploy everywhere. The trick is disciplined scoping and evaluation against the actual job—not generic leaderboards. (more: https://www.reddit.com/r/LocalLLaMA/comments/1o1mpt5/introducing_the_colbert_nano_series_of_models_all/)
MCP apps reshape enterprise workflows
Apps are moving into the model, not the other way around. A widely shared analysis argues OpenAI is turning ChatGPT into an OS‑like distribution channel: an Apps SDK built on the Model Context Protocol (MCP) makes apps first‑class citizens inside ChatGPT, with secure connections to external data/APIs, a pilot app store, and enterprise integrations (e.g., Spotify, Zillow). The piece claims 800M weekly active users and frames the core question for banks: will AI live inside the bank, or will the bank live inside the AI? The same architecture could power internal copilots—compliance, audit, HR—through MCP servers. Risks include misconfigured MCP servers leaking data, platform control over app visibility, and explainability of decisions. (more: https://www.linkedin.com/pulse/from-chatbot-operating-system-what-openais-next-move-means-leimer-ju18c)
Hands-on tooling is catching up. Claude Flow’s v2.5.0‑alpha.141 adds MCP tool integration, an in-process MCP server, a checkpoint/state manager with create/list/restore/delete via CLI, and a hooks system for automating bash and git workflows—useful for workflow recovery, reproducibility, and building robust assistants. (more: https://github.com/ruvnet/claude-flow/issues/793)
On the ground, teams are still kicking the tires. One discussion asks for real-world “agentic AI” office workflows—automation, document handling, task management—and the responses reflect early exploration. Another thread shows why glueing systems via APIs matters more than copy‑pasting: the right answer to “how to automate Claude in the loop” is “call the Claude API or the Agents SDK,” though budget concerns nudge some toward cheaper models and minimal usage. Be wary of low-signal content—the “11 AI Agent Projects” post was removed as self-promo. (more: https://www.reddit.com/r/ollama/comments/1o3waak/anyone_here_building_agentic_ai_into_their_office/) (more: https://www.reddit.com/r/ClaudeAI/comments/1o3tvah/how_would_you_address_it_free_alternatives/) (more: https://www.reddit.com/r/LocalLLaMA/comments/1nzdhif/11_ai_agent_projects_you_can_build_today_with/)
Model-wise, Meituan’s LongCat‑Flash‑Chat is explicitly tuned for agentic tasks. It’s a 560B‑parameter MoE but dynamically activates ≈18.6–31.3B parameters (≈27B on average), uses a shortcut‑connected MoE design to overlap compute and communication, and reports over 100 tokens/sec throughput. The team outlines stability measures for massive training, a data curriculum for reasoning/coding, 128k context, and multi‑agent synthesis for post‑training. Adapters exist for SGLang and vLLM, making it deployable in the same stacks teams already use for GPT‑OSS. (more: https://huggingface.co/meituan-longcat/LongCat-Flash-Chat)
Reliability beats lucky seeds
A useful challenge to benchmark habits: reporting Pass@N without Pass‑all‑N hides instability. For coding tasks like SWE, what matters to an agent user is “Does it work every time?” not “Did it sometimes work across five seeds?” Evaluations on SWE‑rebench show Pass‑all‑5 is consistently lower than mean resolved rate, and the gap varies by model—exactly the reliability signal teams need. Keep Pass@N, but add Pass‑all‑N to standardize reliability reporting for agentic use. (more: https://www.reddit.com/r/LocalLLaMA/comments/1o1dqiy/stop_flexing_passn_show_passalln/)
Model scorecards increasingly include agentic and coding tasks, but interpreting them through reliability helps. LongCat‑Flash reports competitive results across SWE‑Bench‑Verified, LiveCodeBench, MBPP+, and tool‑use benchmarks. These numbers are informative, but stability across runs is what operational teams should demand, especially when plugging models into CI/CD or automated remediation pipelines. (more: https://huggingface.co/meituan-longcat/LongCat-Flash-Chat)
Engineering choices can help models help you. Developers report better AI coding outcomes with modular architectures that keep files under ~1K lines and isolate responsibilities (e.g., MVVM), reducing the context needed for a safe code increment. Frameworks like BMAD and SpecKit are cited as helpful scaffolding. The meta‑lesson: structure your codebase for small, precise edits, and your AI assistant will reward you with fewer regressions. (more: https://www.reddit.com/r/ChatGPTCoding/comments/1o28gge/architecting_a_project_for_optimal_ai_coding_any/)
Also new on the radar: ServiceNow‑AI’s Apriel‑1.5‑15b‑Thinker model card on Hugging Face and Basekick‑Labs’ arc repository on GitHub—both worth a look as the toolbox for building reliable agents keeps expanding. (more: https://huggingface.co/ServiceNow-AI/Apriel-1.5-15b-Thinker) (more: https://github.com/Basekick-Labs/arc)
Small poisons, big risks
A joint Anthropic, UK AI Security Institute, and Alan Turing Institute study finds that a small, fixed number of poisoned documents can reliably implant a backdoor during pretraining, independent of model size or total training volume. In their setup, as few as 250 poisoned samples sufficed to cause a denial‑of‑service backdoor that outputs gibberish on a trigger phrase, across 0.6B–13B‑parameter models. The key takeaway challenges a common assumption: attack success scales with absolute count, not percentage of the corpus, implying practicality even as datasets grow. While the studied behavior is low stakes, the mechanism raises the urgency of scalable defenses and data governance. (more: https://www.anthropic.com/research/small-samples-poison)
Security lapses aren’t confined to model data. Ruby Central’s post‑incident review of a September 2025 event documents AWS root‑level account access tied to unrevoked shared credentials. Although no production harm was found and control was re‑established via MFA and password reset, the episode underscores the need for tight credential lifecycle management, clear ownership, and transparent incident handling—especially for critical package infrastructure. (more: https://rubycentral.org/news/rubygems-org-aws-root-access-event-september-2025/)
These themes converge in the new MCP world. The banking editorial flags risks of poorly configured MCP servers leaking secrets, opaque AI decisioning, and platform distribution power. In short: treat data pipelines, credentials, and app connectors as part of your “model attack surface,” not peripheral plumbing. (more: https://www.linkedin.com/pulse/from-chatbot-operating-system-what-openais-next-move-means-leimer-ju18c)
Few-step video generation arrives
NVLabs’ rCM pushes diffusion distillation to scale, delivering 2–4‑step video generation that preserves both quality and diversity, even when distilled from 10B+ teacher models. The team introduces a score‑regularized, continuous‑time consistency framework, identifies sCM’s quality bottleneck, and overcomes it via a forward–reverse divergence joint distillation. They open‑source a FlashAttention‑2 Jacobian‑vector product kernel with FSDP/checkpoint parallel support and show results on Wan2.1 T2V (1.3B and 14B). Remaining gaps—like physical consistency—are flagged as targets for reward‑based post‑training. The direction is clear: drastically fewer steps for real‑time video generation without collapsing diversity. (more: https://github.com/NVlabs/rcm)
Your light bulbs are sensors now
Smart home ecosystems are turning existing radios into room‑scale motion detectors. Philips Hue’s MotionAware uses Zigbee link fluctuations—signal strength, latency, bit error rates—between multiple bulbs and the new Hue Bridge Pro to infer motion via on‑bridge AI. No camera, no PIR; just RF sensing. Setup includes an empty‑room calibration and works whether lights are on or off; recommended spacing is 1–7 meters across 3–4 bulbs, but reflections can cause cross‑room detections that require tuning. The bridge upgrade is framed as necessary for on‑prem processing, though the monetization angle is fair to ask about. Underlying IP comes from Ivani’s Sensify; Qualcomm holds similar patents. (more: https://hackaday.com/2025/10/07/smart-bulbs-are-turning-into-motion-sensors/)
WiZ beat Hue to market in 2022 with SpaceSense using Wi‑Fi instead of Zigbee and works with as few as two devices. The broader pattern is “ambient AI” that leverages radios you already own for presence and automation—lights on when you enter, HVAC off when unoccupied—without adding hardware or cameras. As always, calibrate carefully and mind the privacy tradeoffs. (more: https://hackaday.com/2025/10/07/smart-bulbs-are-turning-into-motion-sensors/)
Sources (20 articles)
- [Editorial] https://www.anthropic.com/research/small-samples-poison (www.anthropic.com)
- [Editorial] Claude Flow updates (github.com)
- [Editorial] https://www.linkedin.com/pulse/from-chatbot-operating-system-what-openais-next-move-means-leimer-ju18c (www.linkedin.com)
- GPT-OSS from Scratch on AMD GPUs (www.reddit.com)
- Introducing the ColBERT Nano series of models. All 3 of these models come in at less than 1 million parameters (250K, 450K, 950K) (www.reddit.com)
- Stop flexing Pass@N — show Pass-all-N (www.reddit.com)
- How do I compare cost per token for serverless vs provisioned hardware? (www.reddit.com)
- 11 AI Agent Projects You Can Build Today (With Guides) (www.reddit.com)
- Anyone here building Agentic AI into their office workflow? How’s it going so far? (www.reddit.com)
- Architecting a project for optimal AI coding, any tips? (www.reddit.com)
- How would you address it (free alternatives) (www.reddit.com)
- NVlabs/rcm (github.com)
- Basekick-Labs/arc (github.com)
- Rubygems.org AWS Root Access Event – September 2025 (rubycentral.org)
- OpenAI is good at deals (www.bloomberg.com)
- ServiceNow-AI/Apriel-1.5-15b-Thinker (huggingface.co)
- LiquidAI/LFM2-8B-A1B (huggingface.co)
- Smart Bulbs Are Turning Into Motion Sensors (hackaday.com)
- meituan-longcat/LongCat-Flash-Chat (huggingface.co)
- adb1274/batchi (github.com)