MiniMax released M2.5 open weights on HuggingFace, held an AMA, and appeared on SWE-rebench January 2026 leaderboard alongside GLM-5, Opus 4.6, and Qwen3-Coder-Next.
MiniMax M2.5 appeared on SWE-rebench alongside top proprietary models — next bottleneck is quantized inference support and community tooling for its MoE architecture.
4 sources
- reddit MiniMaxAI/MiniMax-M2.5 · Hugging Face 390pts
- reddit MiniMax-M2.5 Checkpoints on huggingface will be in 8 hours 182pts
- reddit AMA with MiniMax — Ask Us Anything! 237pts
- reddit SWE-rebench Jan 2026: GLM-5, MiniMax M2.5,... 277pts
GPT-OSS 120B (128 experts, top-4 routing, ~5.1B active params) released in native MXFP4, while GPT-OSS 20B runs 100% in-browser via WebGPU with ONNX Runtime and Transformers.js v4.
GPT-OSS 20B runs in-browser via WebGPU and 120B ships in MXFP4 — next bottleneck is WebGPU memory limits preventing larger MoE active parameter counts in browser.
2 sources
- reddit GPT-OSS 120b Uncensored Aggressive Release (MXFP4 GGUF) 342pts
- reddit GPT-OSS (20B) running 100% locally in your browser on WebGPU 141pts
Two independent 5B-parameter multimodal models (Dhi-5B trained for $1200, DeepGen 1.0) released, both emphasizing compute-optimal training at small scale.
Dhi-5B trained from scratch for $1200 at 5B params — next bottleneck is evaluation rigor, as neither model has third-party benchmark verification.
2 sources
- reddit UG student launches Dhi-5B (Trained from Scratch) 272pts
- reddit DeepGen 1.0: A 5B parameter "Lightweight" unified... 231pts
Community discussion and SWE-rebench results show GLM-5 and other open-weight models approaching Claude Opus 4.6 on coding benchmarks, with the gap described as the smallest ever.
SWE-rebench Jan 2026 shows open-weight models competitive with proprietary on coding tasks — next bottleneck is whether this holds on harder agentic benchmarks beyond single-PR resolution.
2 sources
- reddit The gap between open-weight and proprietary model... 660pts
- reddit SWE-rebench Jan 2026: GLM-5, MiniMax M2.5,... 277pts
ICML reviewer reports every paper in their batch contains hidden prompt-injection text in the PDF, targeting LLM-based reviewers despite Policy A prohibiting LLM use.
Prompt injection found in every paper in an ICML review batch — next step is whether conferences adopt PDF sanitization or automated detection before reviewer assignment.
2 sources
- reddit [D] ICML: every paper in my review batch contains... 401pts
- reddit [D] Has anyone received their ICML papers to review yet? 12pts
Multiple community workflows use Flux.2 Klein (4B and 9B variants) for all-in-one image editing (inpaint, replace, remove), historical photo restoration, game screenshot remastering, and LoRA fine-tuning for UV maps.
Flux.2 Klein 9B is becoming the default community backbone for image editing workflows — next bottleneck is LoRA training data requirements (38 images reported for UV maps) limiting domain-specific quality.
4 sources
- reddit Flux.2 Klein / Ultimate AIO Pro (t2i, i2i, Inpaint,... 51pts
- reddit DOA is back (!) so I used Klein 9b to remaster it 323pts
- reddit I restored a few historical figures, using Flux.2 Klein 9B. 649pts
- reddit Flux 2 Klein 4b trained on LoRa for UV maps 79pts
Nvidia's Dynamic Memory Sparsification (DMS) retrofits existing LLMs to cut reasoning costs by 8x by dynamically pruning the KV cache during inference without accuracy loss.
1 sources
- reddit Nvidia’s new technique cuts LLM reasoning costs by 8x... 217pts
LTX-2 inpaint tested for lip sync, and SCAIL+VACE+SVI combined for consistent high-quality video shot compositing in diffusion pipelines.
2 sources
- reddit LTX-2 Inpaint test for lip sync 176pts
- reddit Combining SCAIL, VACE & SVI for consistent, very... 45pts
A 2.5-year study on 'undictionary' words — nonexistent tokens that produce consistent effects in CLIP-based diffusion models — published with systematic analysis.
1 sources
Evaluation of 22 model configurations on 169 web research tasks shows higher effort/thinking settings reduce deep research accuracy for GPT-5 and Gemini Flash 3.