Tag
Zyphra released ZAYA1-8B, an 8.4B parameter Mixture-of-Experts model with 760M active parameters, demonstrating high efficiency and strong performance in mathematical and coding reasoning tasks.
Qwen releases Qwen3.6-27B, a 27B dense model claiming flagship-level coding performance surpassing the larger Qwen3.5-397B-A17B MoE, with impressive SVG generation demos.
Xiaomi released Mimo-V2.5, an open-weight AI model, adding to today’s string of open model drops alongside Qwen-27B.
Kimi K2.6 is released as an open-weight model with strong agentic capabilities, accessible via FireworksAI’s fast inference APIs.
Kimi 2.6 is the first open-weight model that demonstrates top-tier agentic capabilities, successfully handling complex chained tool tasks.
Alibaba releases Qwen3.6-35B-A3B-FP8, an open-weight quantized variant of Qwen3.6 with 35B parameters and 3B activated via MoE, featuring improved agentic coding capabilities and thinking preservation for iterative development.
MiniMaxAI releases MiniMax-M2.7, an open-weight model featuring self-evolution capabilities, advanced agent team support, and strong performance on software engineering benchmarks (56.22% on SWE-Pro, 66.6% medal rate on MLE Bench Lite), with notable applications in production incident recovery and professional work tasks.
This is a Hugging Face release for an abliterated version of the Gemma-4-31B model, designed to bypass safety filters for security and harm benchmark testing while maintaining multimodal capabilities.
IBM releases Granite 4.0 3B Vision, a compact vision-language model designed for enterprise document understanding, featuring specialized capabilities for table extraction, chart interpretation via ChartNet, and key-value pair grounding.
Mistral AI has released Mistral Medium 3.5, a dense 128B multimodal model featuring a 256k context window, configurable reasoning capabilities, and improved performance in instruction following, reasoning, and coding tasks.