Improving and Generalizing Flow-Based Generative Models with Minibatch Optimal Transport
Valence Labs via YouTube
Learn Generative AI, Prompt Engineering, and LLMs for Free
Build with Azure OpenAI, Copilot Studio & Agentic Frameworks — Microsoft Certified
Overview
Google, IBM & Meta Certificates — All 10,000+ Courses at 40% Off
One annual plan covers every course and certificate on Coursera. 40% off for a limited time.
Get Full Access
Explore the advancements in flow-based generative models through this comprehensive conference talk. Dive into the world of continuous normalizing flows (CNFs) and discover the innovative generalized conditional flow matching (CFM) technique. Learn how CFM overcomes limitations in simulation-based maximum likelihood training, offering a stable regression objective similar to diffusion models while maintaining efficient inference. Understand the advantages of optimal transport CFM (OT-CFM) in creating simpler, more stable flows for faster inference. Examine the application of these techniques in various conditional and unconditional generation tasks, including single cell dynamics inference, unsupervised image translation, and Schrödinger bridge inference. Gain insights into the comparison between diffusion models and CNFs, and explore the potential of training CNFs like diffusion models. Follow along as the speaker covers topics such as flow matching, conditional flow matching, the impact of probability path choice on flow properties, and the relationship between score and flow matching. Conclude with key takeaways and participate in a Q&A session to deepen your understanding of these cutting-edge concepts in AI and generative modeling.
Syllabus
- Intro
- Background on diffusion + flow models
- Why do diffusion models beat CNFs?
- Main idea: how can we train a CNF like a diffusion model?
- Flow matching
- Conditional flow matching
- Properties of flow depend on the choice of the probability path
- Score and flow matching
- Main takeaways
- Q+A
Taught by
Valence Labs