Memory hints and retrieval help small models reason

Published on

Small language models can improve at math without changing their weights if you teach them to remember and retrieve their own strategies. A local experiment inspired by Google’s ReasoningBank shows ...

Memory, hints, and retrieval help small models reason

Small language models can improve at math without changing their weights if you teach them to remember and retrieve their own strategies. A local experiment inspired by Google’s ReasoningBank shows Qwen3-1.7B-Instruct jumping from 40% to 48% accuracy on MATH Level 3–4 with a memory bank of “strategy” snippets extracted from its own successful solutions and inserted as hints at inference time. The setup emphasized rigor: filtering out any hint that could leak answers, deterministic seeding, and Wilson intervals; the author notes overlapping 95% CIs and calls for larger test sets. Interestingly, the 1.7B model benefited more (but with some regressions) than Qwen3-4B, which saw a smaller +4% gain without regressions, consistent with the idea that tiny models have more to gain from missing strategies but are more fragile to prompt noise. Code is open-sourced and Phase 2 will try fine-tuning (LoRA) on the model’s own successful traces to “bake in” strategies and iterate. (more: https://www.reddit.com/r/LocalLLaMA/comments/1o623qi/i_tested_if_tiny_llms_can_selfimprove_through/)

A broader, production-ready take on persistent, local memory comes from ReasoningBank, presented as a self-learning, local-first system that stores and retrieves reasoning patterns with confidence weights. The author claims 34% task-effectiveness gains, 16% fewer interaction steps, and 87–95% semantic retrieval accuracy, with low-latency SQLite-backed queries and no retraining required. The v2.7.0-alpha.10 release ships pre-trained databases and integration docs, emphasizing local durability and token savings via shorter prompts. These are ambitious claims, and the assets are available for scrutiny. (more: https://www.linkedin.com/posts/reuvencohen_ive-long-said-that-intelligence-without-ugcPost-7384262981305434113-K7AF) (more: https://github.com/ruvnet/claude-flow/tree/main/docs/reasoningbank/)

On the retrieval side, query transformation may be the bigger lever than embeddings for many RAG systems. A practitioner compared nine techniques and found three consistent winners: HydE (generate a hypothetical answer and search against it), RAG-Fusion (multi-query plus reranking), and Step-Back (ask more abstract questions first). The thread’s best observation: many teams over-tune embeddings while under-investing in query formulation. One commenter notes that HydE-like benefits can be shifted to ingestion time by storing extra question embeddings per chunk—trading disk for latency in interactive settings. (more: https://www.reddit.com/r/LocalLLaMA/comments/1o6s89n/tested_9_rag_query_transformation_techniques_hyde/)

Attention limits, robustness, and long context

A new arXiv paper dissects a quiet failure mode of softmax attention: as more tokens are selected, attention becomes less selective, trending toward uniformity, with training also becoming touchy at low temperatures. The authors provide a theoretical framework with separation bounds and validate on GPT-2, arguing for more robust normalization and selection strategies. The takeaway is not a crisis, but an increasingly relevant constraint as contexts and candidate sets grow. (more: https://arxiv.org/abs/2508.17821)

Security researchers are connecting dots between attention behavior and prompt-injection risk. A CISO-in-residence surveys work showing instructions can “distract” models away from original goals—one paper even tracks attention shifting from primary tasks to injected text during successful attacks. The piece is careful: evidence suggests attention and obedience interact, especially with long contexts, but causality is unproven. Other factors (instruction-tuning/RLHF over-obedience, recency bias, RAG/agent pipelines mixing trust boundaries, decoding choices) also plausibly drive vulnerabilities. Teaching “instruction hierarchies” helps, yet it doesn’t settle whether longer context directly enables attacks. It’s a useful research agenda, not a closed case. (more: https://www.linkedin.com/posts/gadievron_i-love-papers-that-explain-the-voodoo-behind-activity-7383995838865260545-SN99)

Meanwhile, training methods are adapting to long reasoning traces without quadratic cost. McGill’s “Markovian Thinking” reframes the RL environment so the model reasons in fixed-size chunks, periodically resetting the prompt and carrying forward a short textual state. Their Delethink approach keeps state bounded (e.g., 8K), matching or beating LongCoT-RL with far less compute and continuing to improve beyond the trained budget, reportedly enabling up to 128K thinking tokens. They also observe signs that some frontier models already exhibit “Markovian thinking” zero-shot. It’s an elegant way to decouple long reasoning from runaway context costs. (more: https://github.com/McGill-NLP/the-markovian-thinker)

Local and edge AI get faster, lighter

The local/edge scene was unusually dense this week. Highlights: NVIDIA’s Fast-dLLM v2 hits 2.5x decoding speed over standard autoregressive with only ~1B tokens of fine-tuning; RND1 releases as an open diffusion language model; MM-HELIX debuts a 7B multimodal “thinking” model sized for local deployment; StreamDiffusionV2 runs real-time interactive video on consumer hardware (16.6 FPS on 2x RTX 4090); Paris demonstrates decentralized training of an open-weight diffusion model; Meta’s SSDD accelerates image tokenization 3.8x, a drop-in KL-VAE replacement; kani-tts-370M delivers lightweight TTS; and VLM-Lens ships a toolkit to interpret local vision-language models. It’s a reminder that “local” now spans text, vision, and audio at usable speeds. (more: https://www.reddit.com/r/LocalLLaMA/comments/1o5pvo2/last_week_in_multimodal_ai_local_edition/)

Liquid AI’s LFM2-8B-A1B takes a different angle: a hybrid MoE with 8.3B total parameters but only 1.5B active, claiming on-device quality near 3–4B dense models and speed faster than Qwen3-1.7B. Quantized variants are meant to fit high-end phones, tablets, and laptops and run with llama.cpp. If these claims hold across tasks, MoE-on-device may become the default for edge assistants. (more: https://huggingface.co/LiquidAI/LFM2-8B-A1B-GGUF)

On the heavy side, a community AWQ quant of BosonAI’s Higgs-Llama-3-70B compresses a ~140GB model to ~37GB, small enough for 40GB consumer GPUs. The thread is a good reminder to evaluate with more than perplexity—community members point to metrics like KL divergence for fidelity after quantization. (more: https://www.reddit.com/r/LocalLLaMA/comments/1o6rshi/bosonais_higgsllama370b_awq_quantized_140gb_37gb/)

Is CPU-only local worth it? It depends. One dev reports 10–15 tokens/sec on sub-20B models with a 13th-gen i7 and 64 GB RAM, while others caution that agentic coding will feel slow and quality won’t match GPT-4-class systems. AMD dGPUs can help (e.g., 6900XT), but the consensus is to temper expectations and pick models carefully for the workload. (more: https://www.reddit.com/r/ollama/comments/1o4ea93/worthwhile_using_ollama_without_nvidia/)

MCP-powered agent integrations grow up

If agent apps are the new web apps, the Model Context Protocol (MCP) is shaping up as their HTTP. Metorial pitches itself as “Vercel for MCP,” a developer platform to connect any model to thousands of tools and data sources in one function call. It abstracts MCP complexity behind SDKs (TS/JS, Python), handles OAuth session management, logs every session with detailed error reports, and lets you test any MCP server in an embedded dashboard. It’s open source and self-hostable, though the hosted path is the fastest on-ramp. (more: https://github.com/metorial/metorial)

Beyond tooling, the marketplace around Claude Code Plugins is exploding—one tracker jumped from 24 to 115 listings in days. Whether these consolidate or fragment is anyone’s guess, but the speed signals both demand and low barriers to entry for MCP-aligned tools. (more: https://www.reddit.com/r/ClaudeAI/comments/1o5yml5/holy_marketplaces_batman/)

DIY LLMs, from $100 chats to RL coders

Andrej Karpathy’s “nanochat” is a compact, hackable end-to-end LLM stack—training, inference, and a web UI—in about 8K lines of code. The pitch: rent an 8x H100 node for ~$24/hour; after ~4 hours (~$100) you get a conversational model; ~12 hours slightly nudges past GPT-2. The default is a ~561M-parameter model trained first on ~24GB of curated web text, then on 568K instruction examples, and finally some preference-style data—small enough to run almost anywhere. A community member even coaxed it to run on CPU on macOS. It’s not state of the art, but it’s transparent, end-to-end, and cheap enough to learn with. (more: https://simonwillison.net/2025/Oct/13/nanochat/) (more: https://github.com/karpathy/nanochat/discussions/1)

For a slower-burn pathway, a practitioner’s series on building an LLM from scratch hits the milestone where the stitched-together architecture finally trains and “talks back,” even if only on 20K characters of Edith Wharton. It later loads GPT-2 weights onto the hand-built model. The post is short, but the lesson stands: carefully assembling the pieces pays off in understanding. (more: https://www.gilesthomas.com/2025/10/llm-from-scratch-22-finally-training-our-llm)

On the other end of the spectrum, Kwaipilot’s KAT-Dev-72B-Exp opens a large RL-tuned coder model with strong SWE-Bench Verified results (74.6% under the SWE-agent scaffold, with specified temperature/turn caps). Under the hood, they rewrote the attention kernel and redesigned training around shared-prefix trajectories for efficiency, and reshaped advantages to prevent exploration collapse. It’s a window into how big-coder RL is evolving, with practical scaffolding and inference configs provided. (more: https://huggingface.co/Kwaipilot/KAT-Dev-72B-Exp)

Science, security, and scrappy engineering

Google’s C2S-Scale 27B—a Gemma-based foundation model for single-cell analysis reportedly helped generate a novel hypothesis about cancer cell behavior, with open weights, code, and a bioRxiv preprint available. The work highlights LLMs as scientific reasoning partners, not just text predictors. A cancer researcher in the thread adds healthy skepticism: the “novel” finding amounts to a particular combo of immunostimulatory compounds working in cell culture, which may not be surprising and may not translate in vivo. The open release invites independent evaluation, which is exactly what’s needed. (more: https://www.reddit.com/r/LocalLLaMA/comments/1o81rvs/google_c2sscale_27b_based_on_gemma_built_with/)

Security-wise, Apple’s iOS architecture keeps inching toward microkernel-like compartmentalization. A new analysis explains SPTM as the authority for memory retyping, creating trust domains that gap functionality like TXM (code signing/entitlements) from XNU. It also details Exclaves and their communication via xnuproxy and the Tightbeam IPC framework. The upshot: more sensitive components move out of the kernel’s direct reach, raising the bar even if XNU is compromised. (more: https://arxiv.org/abs/2510.09272)

Two small but useful engineering notes. First, a tiny Go/Fiber service that returns your public IP accurately, even behind CDNs, demonstrates impressive concurrency with ~21 ms average latency under heavy load, and ships a systemd setup for production use. Simple tools, well-executed, still matter. (more: https://github.com/MrDevAnony/MyIP) Second, Hackaday’s Component Abuse Challenge features a delightful blast from the BEAM-robot past: driving a tiny motor using a 74ACT139 demultiplexer. It works because the motor’s draw is minuscule—but the point is creative constraint-bending, not a new motor-driver standard. (more: https://hackaday.com/2025/10/14/2025-component-abuse-challenge-making-a-ttl-demultiplexer-sweat/)

Sources (21 articles)

  1. [Editorial] The best ChatGPT that $100 can buy. (github.com)
  2. [Editorial] Train your own LLM (simonwillison.net)
  3. [Editorial] ReasoningBank is a self-learning, local-first memory system (www.linkedin.com)
  4. [Editorial] ReasoningBank is a self-learning, local-first memory system (github.com)
  5. [Editorial] Limitations of Normalization in Attention Mechanism (arxiv.org)
  6. [Editorial] Explaining the voodoo behind how AI works (www.linkedin.com)
  7. Last week in Multimodal AI - Local Edition (www.reddit.com)
  8. I tested if tiny LLMs can self-improve through memory: Qwen3-1.7B gained +8% accuracy on MATH problems (www.reddit.com)
  9. Tested 9 RAG query transformation techniques – HydE is absurdly underrated (www.reddit.com)
  10. Google C2S-Scale 27B (based on Gemma) built with Yale generated a novel hypothesis about cancer cellular behavior - Model + resources are now on Hugging Face and GitHub (www.reddit.com)
  11. BosonAI's Higgs-Llama-3-70B AWQ Quantized (140GB → 37GB) (www.reddit.com)
  12. Worthwhile using Ollama without nVidia? (www.reddit.com)
  13. Holy Marketplaces, Batman! (www.reddit.com)
  14. MrDevAnony/MyIP (github.com)
  15. McGill-NLP/the-markovian-thinker (github.com)
  16. Show HN: Metorial (YC F25) – Vercel for MCP (github.com)
  17. Modern iOS Security Features – A Deep Dive into SPTM, TXM, and Exclaves (arxiv.org)
  18. Writing an LLM from scratch, part 22 – training our LLM (www.gilesthomas.com)
  19. LiquidAI/LFM2-8B-A1B-GGUF (huggingface.co)
  20. Kwaipilot/KAT-Dev-72B-Exp (huggingface.co)
  21. 2025 Component Abuse Challenge: Making A TTL Demultiplexer Sweat (hackaday.com)