Note the new recommended sampling parameters for Qwen3.6 27B
Summary
Qwen team updated recommended temperature, top_p and presence_penalty values for their new 27B model to optimize both thinking and instruct modes.
Similar Articles
Optimizing Qwen 3.6 35B A3B sampling parameters.
A researcher seeks faster, lower-variance benchmarks to tune temperature, top_p, top_k and min_p for Qwen 3.6 35B A3B, estimating months of 3090-time with current setups.
Qwen/Qwen3.6-35B-A3B
Qwen releases Qwen3.6-35B-A3B, an open-weight Mixture-of-Experts model with 35B total parameters and 3B active parameters, featuring significant improvements in agentic coding and reasoning preservation.
Qwen/Qwen3.6-35B-A3B-FP8
Alibaba releases Qwen3.6-35B-A3B-FP8, an open-weight quantized variant of Qwen3.6 with 35B parameters and 3B activated via MoE, featuring improved agentic coding capabilities and thinking preservation for iterative development.
Need a second pair of eyes, this Qwen3.6 27B quant recipe consistently thinks less and is correct
The author shares a quantization recipe for Qwen3.6 27B that makes the model use significantly fewer thinking tokens while still producing correct answers, leading to faster inference on math benchmarks.
Qwen3.6-27B
Alibaba's Qwen team released Qwen3.6-27B, a new 27-billion-parameter language model, accompanied by benchmark results.