Abstract

  • Support fast one-step generation by design, while still allowing multistep sampling to trade compute for sample quality
  • trained either by distilling pre-trained diffusion models, or as standalone generative models altogether
  • They build on top of the probability flow (PF) ordinary differential equation (ODE) in continuous-time diffusion models, whose trajectories smoothly transition the data distribution into a tractable noise distribution. We propose to learn a model that maps any point at any time step to the trajectory’s starting point.
    • self-consistency property: Points on the same trajectory map to the same initial point

Main idea

  • A trained diffusion model, one way or another, estimates the score function of the probability distribution
    • whether that’s done directly through score matching
    • or through denoising objective, where in practice,
      • where is the source noise i.e.
  • As soon as you have the score function, assuming it’s a Gaussian Diffusion model, you can sample trajectories from pure noise to new “clean” samples , using an ODE solver (e.g. Euler).
  • Then you can enforce the objective that any samples on a trajectory should be mapped back to .