@HuggingPapers: Stable-GFlowNet: Toward Diverse and Robust LLM Red-Teaming via Contrastive Trajectory Balance Naver AI eliminates unsta…
Summary
Naver AI introduces Stable-GFlowNet, a method to improve LLM red-teaming by eliminating unstable partition function estimation in Generative Flow Networks through contrastive trajectory balance.
View Cached Full Text
Cached at: 05/09/26, 08:16 PM
Stable-GFlowNet: Toward Diverse and Robust LLM Red-Teaming via Contrastive Trajectory Balance
Naver AI eliminates unstable partition function estimation in Generative Flow Networks via pairwise comparisons and robust masking, preventing mode collapse while maintaining diverse https://t.co/xRXREBVzmu
Similar Articles
LeapAlign: Post-Training Flow Matching Models at Any Generation Step by Building Two-Step Trajectories
LeapAlign is a post-training method that improves flow matching model alignment with human preferences by reducing computational costs through two-step trajectory shortcuts while enabling stable gradient propagation to early generation steps. The method outperforms state-of-the-art approaches when fine-tuning Flux models across various image quality and text-alignment metrics.
Flow Map Learning via Nongradient Vector Flow [pdf]
Proposes a nongradient vector flow method for learning flow maps, likely aimed at improving optical flow or motion estimation tasks.
Reinforcement Learning via Value Gradient Flow
Value Gradient Flow (VGF) presents a scalable approach to behavior-regularized reinforcement learning by formulating it as an optimal transport problem solved through discrete gradient flow, achieving state-of-the-art results on offline RL and LLM RL benchmarks. The method eliminates explicit policy parameterization while enabling adaptive test-time scaling by controlling transport budget.
@__JohnNguyen__: Today we released the code for our CVPR 2026 paper, Flowception. Flowception bridges fully bidirectional sequence model…
Meta's FAIR team released the code for Flowception, a CVPR 2026 paper presenting a non-autoregressive video generation framework that interleaves frame insertion with continuous denoising to reduce error accumulation and computational cost.
SDFlow: Similarity-Driven Flow Matching for Time Series Generation
This paper introduces SDFlow, a similarity-driven flow matching framework for time series generation that addresses exposure bias in autoregressive models. It achieves state-of-the-art performance and inference speedups by operating in the frozen VQ latent space with low-rank manifold decomposition.