[2602.16198] Training-Free Adaptation of Diffusion Models via Doob's $h$-Transform

[2602.16198] Training-Free Adaptation of Diffusion Models via Doob's $h$-Transform

arXiv - Machine Learning 4 min read Article

Summary

This paper presents a novel training-free adaptation method for diffusion models, leveraging Doob's $h$-transform to enhance sampling efficiency without additional training overhead.

Why It Matters

The research addresses the limitations of existing adaptation methods for diffusion models, which often require extensive training and specific reward assumptions. By proposing a computationally efficient approach, this work has implications for various applications in machine learning, particularly in reinforcement learning and generative modeling.

Key Takeaways

  • DOIT (Doob-Oriented Inference-time Transformation) offers a training-free adaptation method for diffusion models.
  • The method efficiently transports pre-trained generative distributions to high-reward targets without modifying the original model.
  • Theoretical guarantees of convergence to target distributions are established, enhancing the reliability of the approach.
  • Empirical results show that DOIT outperforms existing state-of-the-art methods on D4RL offline RL benchmarks.
  • This research opens avenues for further exploration in non-differentiable reward scenarios.

Computer Science > Machine Learning arXiv:2602.16198 (cs) [Submitted on 18 Feb 2026] Title:Training-Free Adaptation of Diffusion Models via Doob's $h$-Transform Authors:Qijie Zhu, Zeqi Ye, Han Liu, Zhaoran Wang, Minshuo Chen View a PDF of the paper titled Training-Free Adaptation of Diffusion Models via Doob's $h$-Transform, by Qijie Zhu and 4 other authors View PDF HTML (experimental) Abstract:Adaptation methods have been a workhorse for unlocking the transformative power of pre-trained diffusion models in diverse applications. Existing approaches often abstract adaptation objectives as a reward function and steer diffusion models to generate high-reward samples. However, these approaches can incur high computational overhead due to additional training, or rely on stringent assumptions on the reward such as differentiability. Moreover, despite their empirical success, theoretical justification and guarantees are seldom established. In this paper, we propose DOIT (Doob-Oriented Inference-time Transformation), a training-free and computationally efficient adaptation method that applies to generic, non-differentiable rewards. The key framework underlying our method is a measure transport formulation that seeks to transport the pre-trained generative distribution to a high-reward target distribution. We leverage Doob's $h$-transform to realize this transport, which induces a dynamic correction to the diffusion sampling process and enables efficient simulation-based computatio...

Related Articles

Machine Learning

[HIRING] Machine Learning Evaluation Specialist | Remote | $50/hr

​ We are onboarding domain experts with strong machine learning knowledge to design advanced evaluation tasks for AI systems. About the R...

Reddit - ML Jobs · 1 min ·
Machine Learning

Japan is adopting robotics and physical AI, with a model where startups innovate and corporations provide scale

Physical AI is emerging as one of the next major industrial battlegrounds, with Japan’s push driven more by necessity than anything else....

Reddit - Artificial Intelligence · 1 min ·
Machine Learning

mining hardware doing AI training - is the output actually useful

there's this network that launched recently routing crypto mining hardware toward AI training workloads. miners seem happy with the econo...

Reddit - Artificial Intelligence · 1 min ·
AI is changing how small online sellers decide what to make | MIT Technology Review
Machine Learning

AI is changing how small online sellers decide what to make | MIT Technology Review

Entrepreneurs based in the US are using tools like Alibaba’s Accio to compress weeks of product research and supplier hunting into a sing...

MIT Technology Review · 8 min ·
More in Machine Learning: This Week Guide Trending

No comments

No comments yet. Be the first to comment!

Stay updated with AI News

Get the latest news, tools, and insights delivered to your inbox.

Daily or weekly digest • Unsubscribe anytime