distillation

Tag

Cards List
#distillation

Continuous-Time Distribution Matching for Few-Step Diffusion Distillation

Hugging Face Daily Papers · 2d ago Cached

This paper introduces Continuous-Time Distribution Matching (CDM), a method for few-step diffusion distillation that migrates from discrete to continuous optimization to improve visual fidelity and preserve fine details.

0 favorites 0 likes
#distillation

Stream-R1: Reliability-Perplexity Aware Reward Distillation for Streaming Video Generation

Hugging Face Daily Papers · 4d ago Cached

Stream-R1 introduces a reliability-perplexity aware reward distillation framework for streaming video generation that adaptively weights supervision to improve visual and motion quality without additional computational overhead.

0 favorites 0 likes
#distillation

Jackrong/Qwen3.5-9B-DeepSeek-V4-Flash-GGUF

Hugging Face Models Trending · 2026-04-29 Cached

This entry describes Qwen3.5-9B-DeepSeek-V4-Flash, a distilled AI model that transfers reasoning capabilities from DeepSeek-V4 into a smaller 9B parameter space for efficient inference.

0 favorites 0 likes
#distillation

SCURank: Ranking Multiple Candidate Summaries with Summary Content Units for Enhanced Summarization

arXiv cs.CL · 2026-04-22 Cached

SCURank introduces Summary Content Units to rank candidate summaries, enabling small models distilled from multiple LLMs to outperform traditional metrics and single-LLM distillates.

0 favorites 0 likes
#distillation

TIPSv2: Advancing Vision-Language Pretraining with Enhanced Patch-Text Alignment

Hugging Face Daily Papers · 2026-04-13 Cached

TIPSv2 introduces enhanced vision-language pretraining techniques including patch-level distillation, an upgraded masked image objective (iBOT++), and improved caption sampling strategies to achieve superior dense patch-text alignment. The resulting family of image-text encoder models demonstrates strong performance across 9 tasks and 20 datasets.

0 favorites 0 likes
#distillation

Jackrong/Qwen3.5-27B-Claude-4.6-Opus-Reasoning-Distilled

Hugging Face Models Trending · 2026-02-27 Cached

Jackrong releases Qwen3.5-27B-Claude-4.6-Opus-Reasoning-Distilled, a fine-tuned 27B parameter model with improved reasoning capabilities and stability, along with comprehensive training guides and code on GitHub using the Unsloth framework.

0 favorites 0 likes
#distillation

Consistency Models

OpenAI Blog · 2024-06-20 Cached

OpenAI introduces Consistency Models, a new family of generative models that enable fast one-step image generation by directly mapping noise to data, while supporting multi-step sampling and zero-shot editing tasks like inpainting and super-resolution. The approach achieves state-of-the-art FID scores on CIFAR-10 and ImageNet 64x64 for one-step generation.

0 favorites 0 likes
#distillation

prunaai/z-image-turbo

Replicate Explore · 2026-04-21 Cached

Alibaba’s 6B-parameter Z-Image-Turbo text-to-image model, further compressed by PrunaAI, generates 1024×1024 photorealistic images with bilingual text in <1s on 8 diffusion steps.

0 favorites 0 likes
← Back to home

Submit Feedback