diffusion-language-models

Tag

Cards List
#diffusion-language-models

LEAP: Unlocking dLLM Parallelism via Lookahead Early-Convergence Token Detection

arXiv cs.LG · 17h ago Cached

This paper introduces LEAP, a training-free method to accelerate inference in Diffusion Language Models (dLLMs) by detecting early-converging tokens, reducing denoising steps by 30% without losing accuracy.

0 favorites 0 likes
#diffusion-language-models

Self-Distilled Trajectory-Aware Boltzmann Modeling: Bridging the Training-Inference Discrepancy in Diffusion Language Models

arXiv cs.CL · 17h ago Cached

This paper introduces TABOM, a self-distilled trajectory-based post-training framework for Diffusion Language Models that aligns training with inference trajectories using Boltzmann modeling to mitigate the training-inference discrepancy and reduce catastrophic forgetting.

0 favorites 0 likes
#diffusion-language-models

Don't Retrain, Align: Adapting Autoregressive LMs to Diffusion LMs via Representation Alignment

arXiv cs.LG · 2d ago Cached

This paper introduces Repr-Align, a method to adapt autoregressive language models into diffusion language models via representation alignment, achieving up to 4x training acceleration without retraining representations from scratch.

0 favorites 0 likes
#diffusion-language-models

DiffRetriever: Parallel Representative Tokens for Retrieval with Diffusion Language Models

Hugging Face Daily Papers · 5d ago Cached

This paper introduces DiffRetriever, a method that uses diffusion language models to generate multiple representative tokens in parallel for efficient information retrieval, outperforming autoregressive baselines in speed and accuracy.

0 favorites 0 likes
#diffusion-language-models

DALM: A Domain-Algebraic Language Model via Three-Phase Structured Generation

arXiv cs.CL · 2026-04-20 Cached

DALM proposes a domain-algebraic language model that generates text under exact structural constraints derived from a domain lattice, addressing hallucination by organizing knowledge into separate domain fibers with algebraic guarantees. The model uses three-phase structured denoising (domain → relation → concept) with domain-annotated training data to prevent cross-domain contamination.

0 favorites 0 likes
← Back to home

Submit Feedback