@Daniel_Farinax: Qwen3.6-27B on MacBook Pro M5 128GB MLX with custom coding CLI optimized for it. Should also work on M1, M2, M3, M4 Mac…
Summary
Daniel Farinax announces a custom CLI for running Qwen3.6-27B on MacBooks via MLX, seeking beta testers and moving to TypeScript for faster iteration.
View Cached Full Text
Cached at: 05/13/26, 10:25 PM
Qwen3.6-27B on MacBook Pro M5 128GB MLX with custom coding CLI optimized for it. Should also work on M1, M2, M3, M4 Macs.
Created this in 8 prompts. I need beta testers for the CLI. Pivoting to TypeScript for faster iteration. https://t.co/DHESMMbcp3
Similar Articles
@Daniel_Farinax: Qwen3.6-27B on MacBook Pro M5 128GB. Third version of the game. This time a low-poly GTA, built overnight using a custo…
Daniel Farinax demonstrates running Qwen3.6-27B on a MacBook Pro M5 128GB, using a custom Rust CLI (MPTLX) to build a low-poly GTA game overnight, claiming blazing fast performance comparable to Claude 4.6 running locally.
Qwen3.6-35B-A3B-Abliterated-Heretic-MLX-4bit
The user reviews a quantized and fine-tuned version of the Qwen3.6-35B model optimized for Apple Silicon via MLX, praising its speed, intelligence, and lack of safety disclaimers.
@rohanpaul_ai: Qwen 3.6 27B on a MacBook Pro M5 Max 64GB hitting 34tokens per sec, locally with atomic[.]chat 90% acceptance rate, i.e…
Qwen 3.6 27B achieves 34 tokens/sec on a MacBook Pro M5 Max 64GB locally with 90% draft acceptance, enabled by TurboQuant, GGUF, and llama.cpp, showcasing a major advancement in laptop-based AI inference.
I benchmarked 21 local LLMs on a MacBook Air M5 for code quality AND speed
A developer benchmarked 21 local LLMs on MacBook Air M5 using HumanEval+ and found Qwen 3.6 35B-A3B (MoE) leads at 89.6% with 16.9 tok/s, while Qwen 2.5 Coder 7B offers the best RAM-to-performance ratio at 84.2% in 4.5 GB. Notably, Gemma 4 models significantly underperformed expectations (31.1% for 31B), possibly due to Q4_K_M quantization effects.
@remilouf: Following @julien_c’s tweet I bought a MacBook Pro with 128B unified memory, and started running Qwen3.6 as my daily dr…
The author shares their experience running the Qwen3.6 model on a MacBook Pro with 128GB of unified memory, praising Apple's hardware efficiency for local AI inference.