“W.A.L.T: Photorealistic Video Generation With Diffusion Models”, 2023-12-11 ():
We present W.A.L.T, a transformer-based approach for photorealistic video generation via diffusion modeling. Our approach has two key design decisions. First, we use a causal encoder to jointly compress images and videos within a unified latent space, enabling training and generation across modalities.
Second, for memory and training efficiency, we use a window attention architecture tailored for joint spatial and spatiotemporal generative modeling. Taken together these design decisions enable us to achieve state-of-the-art performance on established video (UCF101 and Kinetics-600) and image generation benchmarks such as ImageNet without using classifier free guidance.
Finally, we also train a cascade of 3 models for the task of text-to-video generation consisting of a base latent video diffusion model, and two video super-resolution diffusion models to generate videos of 512×896 resolution at 8 FPS.