[2602.16796] Efficient Tail-Aware Generative Optimization via Flow Model Fine-Tuning
Summary
This article presents Tail-aware Flow Fine-Tuning (TFFT), a novel algorithm that optimizes generative models by controlling tail behavior, enhancing reliability and discovery in machine learning applications.
Why It Matters
The ability to manage tail behavior in generative models is crucial for applications requiring reliability and innovation. This research addresses a gap in existing methods by providing a principled approach to optimize both low-reward and high-reward outcomes, which can significantly impact fields like AI-driven design and content generation.
Key Takeaways
- TFFT optimizes generative models by focusing on tail behavior.
- The method uses Conditional Value-at-Risk (CVaR) for effective fine-tuning.
- It balances the need for reliability (low-reward) and innovation (high-reward).
- The algorithm is computationally efficient compared to traditional methods.
- Demonstrated effectiveness in diverse applications like text-to-image generation.
Computer Science > Machine Learning arXiv:2602.16796 (cs) [Submitted on 18 Feb 2026] Title:Efficient Tail-Aware Generative Optimization via Flow Model Fine-Tuning Authors:Zifan Wang, Riccardo De Santi, Xiaoyu Mo, Michael M. Zavlanos, Andreas Krause, Karl H. Johansson View a PDF of the paper titled Efficient Tail-Aware Generative Optimization via Flow Model Fine-Tuning, by Zifan Wang and 5 other authors View PDF HTML (experimental) Abstract:Fine-tuning pre-trained diffusion and flow models to optimize downstream utilities is central to real-world deployment. Existing entropy-regularized methods primarily maximize expected reward, providing no mechanism to shape tail behavior. However, tail control is often essential: the lower tail determines reliability by limiting low-reward failures, while the upper tail enables discovery by prioritizing rare, high-reward outcomes. In this work, we present Tail-aware Flow Fine-Tuning (TFFT), a principled and efficient distributional fine-tuning algorithm based on the Conditional Value-at-Risk (CVaR). We address two distinct tail-shaping goals: right-CVaR for seeking novel samples in the high-reward tail and left-CVaR for controlling worst-case samples in the low-reward tail. Unlike prior approaches that rely on non-linear optimization, we leverage the variational dual formulation of CVaR to decompose it into a decoupled two-stage procedure: a lightweight one-dimensional threshold optimization step, and a single entropy-regularized fine-t...