Support fast one-step generation by design, while still allowing multistep sampling to trade compute for sample quality
trained either by distilling pre-trained diffusion models, or as standalone generative models altogether
They build on top of the probability flow (PF) ordinary differential equation (ODE) in continuous-time diffusion models, whose trajectories smoothly transition the data distribution into a tractable noise distribution. We propose to learn a model that maps any point at any time step to the trajectory’s starting point.
self-consistency property: Points on the same trajectory map to the same initial point
Main idea
A trained diffusion model, one way or another, estimates the score function of the probability distribution
whether that’s done directly through score matching
or through denoising objective, where in practice, ∇p(xt)=−αϵ0
where ϵ0 is the source noise i.e. xt=x0+ϵ0
As soon as you have the score function, assuming it’s a Gaussian Diffusion model, you can sample trajectories from pure noise xT to new “clean” samples x^0, using an ODE solver (e.g. Euler).
Then you can enforce the objective that any samples on a trajectory should be mapped back to x^0.