Qwen3.6-35B-A3B-Abliterated-Heretic-MLX-4bit
Summary
The user reviews a quantized and fine-tuned version of the Qwen3.6-35B model optimized for Apple Silicon via MLX, praising its speed, intelligence, and lack of safety disclaimers.
Similar Articles
Qwen/Qwen3.6-35B-A3B-FP8
Alibaba releases Qwen3.6-35B-A3B-FP8, an open-weight quantized variant of Qwen3.6 with 35B parameters and 3B activated via MoE, featuring improved agentic coding capabilities and thinking preservation for iterative development.
@songjunkr: SuperQwen3.6-35B-DFlash-MLX is ready. Benchmark: Comparison of original vs. tuned versions on 100 actual items from com…
A fine-tuned 35B-parameter Qwen model optimized for MLX shows benchmark gains on GPQA Diamond, MMLU-Pro, IFEval, HumanEval+ and MBPP+ and ships without censorship.
Running Qwen3.6-35B-A3B Locally for Coding Agent: My Setup & Working Config
A detailed guide for running the 35B-parameter Qwen3.6 model locally on Apple Silicon with llama.cpp to power the pi coding agent, including optimized configuration flags and sampling parameters.
Qwen 3.6 35B A3B vs Qwen 3.5 122B A10B
User reports Qwen 3.5 122B significantly outperforms Qwen 3.6 35B on multi-step tasks despite benchmark claims, questioning if quantization or setup issues are to blame.
I benchmarked 21 local LLMs on a MacBook Air M5 for code quality AND speed
A developer benchmarked 21 local LLMs on MacBook Air M5 using HumanEval+ and found Qwen 3.6 35B-A3B (MoE) leads at 89.6% with 16.9 tok/s, while Qwen 2.5 Coder 7B offers the best RAM-to-performance ratio at 84.2% in 4.5 GB. Notably, Gemma 4 models significantly underperformed expectations (31.1% for 31B), possibly due to Q4_K_M quantization effects.