A Closer Look at Time Steps is Worthy of
Triple Speed-Up for Diffusion Model Training


Kai Wang 2
Yukun Zhou 1,2
Mingjia Shi 2
Zhihang Yuan 3
Yuzhang Shang 4
Xiaojiang Peng 1
Hanwang Zhang 5
Yang You 2
1 2 3
4 5




SpeeD : Closer Look at Time Steps

$$ \text{Take a closer look at process increments }\delta_{t}:=x_{t+1}-x_{t} \text{ over time steps } t \text{.} $$


Abstract

Training diffusion models is always a computation-intensive task. In this paper, we introduce a novel speed-up method for diffusion model training, called SpeeD, which is based on a closer look at time steps. Our key findings are: i) Time steps can be empirically divided into acceleration, deceleration, and convergence areas based on the process increment. ii) These time steps are imbalanced, with many concentrated in the convergence area. iii) The concentrated steps provide limited benefits for diffusion training. To address this, we design an asymmetric sampling strategy that reduces the frequency of steps from the convergence area while increasing the sampling probability for steps from other areas. Additionally, we propose a weighting strategy to emphasize the importance of time steps with rapid-change process increments. As a plug-and-play and architecture-agnostic approach, SpeeD consistently achieves 3× acceleration across various diffusion architectures, datasets, and tasks. Notably, due to its simple design, our approach significantly reduces the cost of diffusion model training with minimal overhead. Our research enables more researchers to train diffusion models at a lower cost


Method

Illustration of the proposed SpeeD framework. Our approach consists of two proceses, asymmetric sampling and change-aware weighting. Asymmetric Sampling: Suppress the attendance of the time step in convergence areas. Change-Aware Weighting: The faster changing time steps in the diffusion process are given more weight.


Results

Our method obtains similar or even higher performance than baselines. The results of SpeeD are averaged in three runs. Bold entries are best results.

BibTeX

@article{wang2024closer,
title={A Closer Look at Time Steps is Worthy of Triple Speed-Up for Diffusion Model Training},
author={Kai Wang, Yukun Zhou, Mingjia Shi, Zhihang Yuan, Yuzhang Shang, Xiaojiang Peng, Hanwang Zhang and Yang You},
year={2024},
journal={arXiv preprint arXiv:2405.17403},
}



Please feel free to contact us via email at Kai Wang, Yukun Zhou or Mingjia Shi for any inquiries or assistance.