Skip to content

ChiehHsinJesseLai/SeriesTalksDiffusionModels2024

Repository files navigation

Evolution of Diffusion Models: From Birth to Enhanced Accuracy and Efficiency

This repository contains my (Chieh-Hsin (Jesse) Lai) slides for a series of talks on Diffusion Models in 2024.

The content covers the origin of Diffusion Models, the challenges they face (including training inaccuracies, slow sampling, and scalability issues), and our efforts to address these problems. It also explores applications in solving inverse problems and achieving controllable generation with Diffusion Models. Overall, it includes our following works:

  • FP-Diffusion (ICML'23): Improving the density estimation and training accuracy of diffusion models by incorporating underlying partial differential equations.
  • CTM (ICLR'24): A unified framework for diffusion models and distillation-based methods for diverse samplers and one-step SOTA generation.
  • SoundCTM (Preprint'24): Extension of CTM to Text-to-Sound Generation with Flexible Sampling.
  • PaGoDA (NeurIPS'24): A 64x64 pre-trained diffusion model is all you need for one-step high-resolution state-of-the-art generation.
  • GibbsDDRM (ICML'23): Training-free blind inverse problem solver based on diffusion models.
  • MPGD (ICLR'24): A fast, efficient, training-free, and controllable diffusion-based generation method.

These talks would not have been possible without the tremendous efforts of the authors in the works mentioned above!

How to cite this PDF:

@misc{lai24evolutiondiffusionmodels,
  author = {Lai, Chieh-Hsin and Mitsufuji, Yuki},
  title = {Evolution of Diffusion Models: From Birth to Enhanced Accuracy and Efficiency},
  year = {2024},
  journal = {GitHub repository},
  url = {https://github.com/ChiehHsinJesseLai/SeriesTalksDiffusionModels2024}, 
}

Contact

Chieh-Hsin (Jesse) LAI:

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published