Who is your favourite quant publisher and why?
Mirrored from r/LocalLLaMA for archival readability. Support the source by reading on the original site.
Hey everyone,
I’ve been a big fan of Unsloth for several reasons:
- They publish models ASAP after release.
- They usually offer the lowest PPL.
- Their website has tons of helpful tutorials and documentation.
Recently, I stumbled upon this Reddit thread suggesting to try out an Apex MoE quant of Mudler instead:
👉 https://www.reddit.com/r/LocalLLaMA/comments/1t3n6jo/apex_moe_quants_update_25_new_models_since_the/
So I decided to test it myself. I tried running Qwen3.5 122B IQuality, which is roughly the same size as Qwen3.5 122B Q4_K_XL. So far, I haven’t noticed a difference in real world tasks between these two models in terms of output quality so i decided to run one gsm8k benchmark and unsloth was slightly better.
So im asking you now, who is your fav publisher and why?
[link] [comments]
More from r/LocalLLaMA
-
24+ tok/s from ~30B MoE models on an old GTX 1080 (8 GB VRAM, 128k context)
May 13
-
Web-Search is coming to a screeching performance halt as Google shuts down their free search index, and traffic defenders like Cloudflare challenge AI at every gateway. What are our options?
May 13
-
Side Projects.
May 13
-
MI50s Qwen 3.6 27B @52.8 tps TG @1569 tps PP (no MTP, no Quant)
May 13
Discussion (0)
Sign in to join the discussion. Free account, 30 seconds — email code or GitHub.
Sign in →No comments yet. Sign in and be the first to say something.