Stability AI unveils Stable Video Diffusion, a foundation model for generative video

Stability AI unveils Stable Video Diffusion, a foundation model for generative video

Stability AI, the company responsible for the well-known deep learning model Stable Diffusion, has recently announced the launch of Stable Video Diffusion. This marks the first foundation model for generative video, based on the image model Stable Diffusion.

The newly released video model can be adapted to a range of downstream tasks. Notably, it can be fine-tuned on multi-view datasets to achieve multi-view synthesis from a single image. Stability AI intends to develop a variety of models that build upon and expand this base, aiming to create an ecosystem akin to the one that surrounds Stable Diffusion.

Stability AI has released Stable Video Diffusion in the form of two image-to-video models. These models can generate 14 and 25 frames respectively, with customizable frame rates ranging from 3 to 30 frames per second.

Upon release, external evaluation revealed that these models outperform the leading closed models (like Pika Labs and Runway) in user preference studies, according to Stability AI. However, the company has clarified that this model is not intended for real-world or commercial applications at the current stage.

The code for Stable Video Diffusion is accessible on GitHub. The necessary weights for running the model locally are available on the Hugging Face page.

by Paul

Maoholguin
ni
Ola
Maoholguin found this interesting
  • ...

Stable Video Diffusion is a versatile tool designed for extensive video applications across multiple sectors including Media, Entertainment, Education, and Marketing. It enables users to convert text and image inputs into dynamic scenes, transforming ideas into cinematic productions. Notable alternatives to Stable Video Diffusion include Wonder Studio, Kaiber, and Pika Labs.

Comments

niteshsinghmrl
1

OHH that can save a time in editing ! intresting

Gu