Tag
This paper introduces Continuous-Time Distribution Matching (CDM), a method for few-step diffusion distillation that migrates from discrete to continuous optimization to improve visual fidelity and preserve fine details.
Stream-R1 introduces a reliability-perplexity aware reward distillation framework for streaming video generation that adaptively weights supervision to improve visual and motion quality without additional computational overhead.
This entry describes Qwen3.5-9B-DeepSeek-V4-Flash, a distilled AI model that transfers reasoning capabilities from DeepSeek-V4 into a smaller 9B parameter space for efficient inference.
SCURank introduces Summary Content Units to rank candidate summaries, enabling small models distilled from multiple LLMs to outperform traditional metrics and single-LLM distillates.
TIPSv2 introduces enhanced vision-language pretraining techniques including patch-level distillation, an upgraded masked image objective (iBOT++), and improved caption sampling strategies to achieve superior dense patch-text alignment. The resulting family of image-text encoder models demonstrates strong performance across 9 tasks and 20 datasets.
Jackrong releases Qwen3.5-27B-Claude-4.6-Opus-Reasoning-Distilled, a fine-tuned 27B parameter model with improved reasoning capabilities and stability, along with comprehensive training guides and code on GitHub using the Unsloth framework.
OpenAI introduces Consistency Models, a new family of generative models that enable fast one-step image generation by directly mapping noise to data, while supporting multi-step sampling and zero-shot editing tasks like inpainting and super-resolution. The approach achieves state-of-the-art FID scores on CIFAR-10 and ImageNet 64x64 for one-step generation.
Alibaba’s 6B-parameter Z-Image-Turbo text-to-image model, further compressed by PrunaAI, generates 1024×1024 photorealistic images with bilingual text in <1s on 8 diffusion steps.