Who is your favourite quant publisher and why?
Summary
A user shares their preference for Unsloth quantized models due to fast releases and low perplexity, compares them with Apex MoE quants, and asks the community for their favorite quant publisher.
Similar Articles
MagicQuant (v2.0) - Hybrid Mixed GGUF Models + Unsloth Dynamic Learned Quant Configurations + Benchmark table with collapsed winners and more
MagicQuant v2.0 is a pipeline for creating hybrid mixed GGUF quant models, learning from Unsloth and other methods to find optimal quant configurations based on KLD benchmarks, with a focus on nonlinear wins and anomaly detection.
@Italianclownz: Tested MTP, TriAttention, TurboQuant on @UnslothAI @Alibaba_Qwen Qwen 3.6 35B A3B MTP MXFP4_MoE on @huggingface @no_stp…
A user benchmarks MTP, TriAttention, and TurboQuant optimizations on Qwen 3.6 35B using Unsloth on consumer hardware, finding TurboQuant to be the most effective.
@_EldarKurtic: TurboQuant has drawn a lot of attention recently, but the accompanying evals didn't tell the full story. So we ran what…
Eldar Kurtic presents a comprehensive study on TurboQuant, revealing its real-world effects on accuracy, latency, and throughput beyond initial evaluations.
Need Info on quality benchmarks to run on DeepSeek V3.2 different quant levels [D]
Developer seeks quality benchmarks to evaluate runtime quantization impact on DeepSeek V3.2 model performance.
Models and Quants quality test results - the chessboard svg (Qwen3.6 27B/35B-A3B/Zaya1)
Community testers evaluate quantized versions of Qwen3.6, ZAYA1, and other models for SVG chessboard generation accuracy using local inference frameworks like MLX.