Abstract
Diffusion models are a powerful class of techniques in ML for generating realistic data, but they are highly prone to overfitting, especially with limited training data. While data augmentation such as image rotation can mitigate this issue, it often causes leakage, where augmented content appears in generated samples. In this paper, we propose a novel regularization framework, called shifted data-augmentation (SDA), for training unconditional diffusion models. SDA introduces an auxiliary diffusion path using transformed data and the noise-shift technique alongside the standard path with original data. This dual-path structure enables effective regularization while suppressing leakage through a trajectory shift in the diffusion process. We implement SDA with image rotation as its most basic and interpretable configuration. We also conduct synthetic and empirical analyses demonstrating that SDA effectively leverages the regularization benefit of image rotation. In particular, SDA yielded notable performance in training with limited data.
Original language | English |
---|---|
Journal | IEEE Access |
Volume | 13 |
DOIs | |
State | Accepted/In press - 2025 |
Bibliographical note
Publisher Copyright:© 2013 IEEE.
Keywords
- data augmentation
- diffusion model
- noise-shift
- regularization
ASJC Scopus subject areas
- General Computer Science
- General Materials Science
- General Engineering