Note the new recommended sampling parameters for Qwen3.6 27B

Reddit r/LocalLLaMA Models

Summary

Qwen team updated recommended temperature, top_p and presence_penalty values for their new 27B model to optimize both thinking and instruct modes.

Taken from their [Huggingface Page:](https://huggingface.co/Qwen/Qwen3.6-27B) *We recommend using the following set of sampling parameters for generation* Thinking mode for general tasks: temperature=1.0, top_p=0.95, top_k=20, min_p=0.0, presence_penalty=0.0, repetition_penalty=1.0 Thinking mode for precise coding tasks (e.g. WebDev): temperature=0.6, top_p=0.95, top_k=20, min_p=0.0, presence_penalty=0.0, repetition_penalty=1.0 Instruct (or non-thinking) mode: temperature=0.7, top_p=0.80, top_k=20, min_p=0.0, presence_penalty=1.5, repetition_penalty=1.0 These are different from 3.5 so I thought I would draw your attention to them.
Original Article

Similar Articles

Optimizing Qwen 3.6 35B A3B sampling parameters.

Reddit r/LocalLLaMA

A researcher seeks faster, lower-variance benchmarks to tune temperature, top_p, top_k and min_p for Qwen 3.6 35B A3B, estimating months of 3090-time with current setups.

Qwen/Qwen3.6-35B-A3B

Hugging Face Models Trending

Qwen releases Qwen3.6-35B-A3B, an open-weight Mixture-of-Experts model with 35B total parameters and 3B active parameters, featuring significant improvements in agentic coding and reasoning preservation.

Qwen/Qwen3.6-35B-A3B-FP8

Hugging Face Models Trending

Alibaba releases Qwen3.6-35B-A3B-FP8, an open-weight quantized variant of Qwen3.6 with 35B parameters and 3B activated via MoE, featuring improved agentic coding capabilities and thinking preservation for iterative development.

Qwen3.6-27B

Reddit r/LocalLLaMA

Alibaba's Qwen team released Qwen3.6-27B, a new 27-billion-parameter language model, accompanied by benchmark results.