Improved Contrastive Divergence Training of Energy Based Models

Yilun Du, Shuang Li, Joshua Tenenbaum, Igor Mordatch

We propose several different techniques to improve contrastive divergence training of energy-based models (EBMs). We first show that a gradient term neglected in the popular contrastive divergence formulation is both tractable to estimate and is important to avoid training instabilities in previous models. We further highlight how data augmentation, multi-scale processing, and reservoir sampling can be used to improve model robustness and generation quality. Thirdly, we empirically evaluate stability of model architectures and show improved performance on a host of benchmarks and use cases, such as image generation, OOD detection, and compositional generation.

Knowledge Graph

arrow_drop_up

Comments

Sign up or login to leave a comment