Papers
Topics
Authors
Recent
Search
2000 character limit reached

JVID: Joint Video-Image Diffusion for Visual-Quality and Temporal-Consistency in Video Generation

Published 21 Sep 2024 in cs.CV | (2409.14149v2)

Abstract: We introduce the Joint Video-Image Diffusion model (JVID), a novel approach to generating high-quality and temporally coherent videos. We achieve this by integrating two diffusion models: a Latent Image Diffusion Model (LIDM) trained on images and a Latent Video Diffusion Model (LVDM) trained on video data. Our method combines these models in the reverse diffusion process, where the LIDM enhances image quality and the LVDM ensures temporal consistency. This unique combination allows us to effectively handle the complex spatio-temporal dynamics in video generation. Our results demonstrate quantitative and qualitative improvements in producing realistic and coherent videos.

Summary

  • The paper introduces the JVID model that combines latent image and video diffusion for superior frame quality and temporal coherence.
  • It employs a novel timestep sampling strategy that alternates between image and video denoising models to optimize both quality and consistency.
  • Experimental results on UCF-101 demonstrate enhanced performance with improved FID, FVD, and IS scores compared to baseline methods.

JVID: Joint Video-Image Diffusion for Visual-Quality and Temporal-Consistency in Video Generation

Introduction

The paper "JVID: Joint Video-Image Diffusion for Visual-Quality and Temporal-Consistency in Video Generation" introduces the Joint Video-Image Diffusion (JVID) model, targeting the enhancement of video generation in terms of visual quality and temporal consistency. This approach addresses the challenges in video generation posed by high dimensionality and complex spatio-temporal dynamics. It leverages two diffusion models: a Latent Image Diffusion Model (LIDM) and a Latent Video Diffusion Model (LVDM), trained respectively on images and video data, and strategically combines them during the reverse diffusion process.

Methodology

The JVID model integrates LIDM and LVDM via a novel sampling strategy. The LIDM is responsible for refining image quality, while the LVDM maintains temporal consistency across frames. The key innovation lies in mixing these models during the denoising process of the diffusion models, where each timestep involves selecting one model to contribute to the denoising operation, thus merging their capabilities effectively. Figure 1

Figure 1: Our proposed mixture of denoising model sampling approach. At each step, we select one model, which is used to denoise our noise sample st\bm{s}_t. When the sample is fully denoised, it is decoded with the VAE decoder to reconstruct the generated video frames.

The paper outlines a method in which noise samples are progressively denoised through different stages, using either the LIDM or LVDM to achieve an optimal balance between image quality and temporal coherence. Another contribution is modeling video and image distributions in a latent space, aligning with methods like latent diffusion models, significantly reducing computational overhead during both training and inference.

Experimental Setup

The experiments are conducted on the UCF-101 dataset, selected for its comprehensive action classes captured in diverse conditions. By encoding video samples using a pre-trained VAE from Stable Diffusion, JVID capitalizes on reduced computational costs and increased efficiency in handling high-dimensional video data. The models are rigorously evaluated based on Fréchet Inception Distance (FID), Fréchet Video Distance (FVD), and Inception Score (IS) across varying resolutions, confirming improvements afforded by the JVID framework. Figure 2

Figure 2: Video samples generated with our JVID model, combining both an image and a video diffusion model during sampling, to produce high quality and temporally coherent videos. Rows 1-4 are generated at 128 × 128, while rows 5-8 are 64 × 64.

Results and Discussion

The results demonstrate that JVID achieves significant visual and temporal improvements in video generation compared to traditional methods. JVID's incorporation of two independently trained models harmonizes the typically conflicting requirements of frame quality and temporal smoothness. Numerical assessments highlight that the JVID model outperforms baseline approaches, though further scaling could improve its competitive stance against state-of-the-art models pre-trained on larger datasets.

This research suggests an efficient solution for synthesizing videos that maintain aesthetic quality across frames, providing a groundwork for further study in conditional and large-scale deployments. Future research could explore its application in domains requiring stringent video quality standards, such as autonomous driving simulations or augmented reality.

Conclusion

The JVID model presents a promising advancement in video generation by integrating image and video diffusion processes, achieving superior temporal coherence and visual quality. This work provides a strategic approach to enhance video synthesis efficiency, inviting future explorations to further optimize and scale the methodology for broader AI applications.

Paper to Video (Beta)

Whiteboard

No one has generated a whiteboard explanation for this paper yet.

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Collections

Sign up for free to add this paper to one or more collections.