r/LocalLLaMA Daily Update (24h, 2026-03-23 JST)
Top concrete r/LocalLLaMA updates from the last 24 hours: model/open-weight announcements, tooling releases, and useful resources.
Window: last 24 hours (reported on 2026-03-23 JST)
Models
- MiniMax M2.7 will be open weights — notable upcoming open-weights model announcement.
- Alibaba confirms continued open-sourcing of Qwen and Wan models — roadmap-level signal for future model drops.
- Qwen3.5-122B-A10B Uncensored (Aggressive) GGUF release + new K_P quants — fresh quantized local deployment option.
- Qwen3.5-9B-Claude-4.6-Opus-Uncensored-v2 Q4_K_M GGUF — updated small-footprint GGUF release.
Tools/Frameworks
- Kreuzberg v4.5.0 released — Docling model integration with a faster engine.
- FeatherOps: fast fp8 matmul on RDNA3 (without native fp8) — low-level performance work relevant to AMD local inference stacks.
- Docker vLLM config for Qwen3.5-122B-A10B-NVFP4 — practical deployment config for large-model serving.
Resources
- A Collection of Nice Datasets — curated dataset links for training/eval workflows.
- One-command local AI stack for AMD Strix Halo — setup shortcut for local stack bootstrap on Strix Halo systems.