Google Makes It Easy to Turn Photos into Short Videos - Meet VDIM

Google Makes It Easy to Turn Photos into Short Videos - Meet VDIM

Google has unveiled a new AI model that can take two images and fill in the gaps between them to create a seamless animation that closely resembles live action.

VDIM (Video Interpolation With Diffusion Models) was created by Google's research division DeepMind and uses one image as the first frame and the other as the last frame. All shots in between are then filled in using AI to create the video.

This could be great for bringing to life a photo of a child playing in the park or during an event where you forgot to capture the action.

While currently only a preview of the research phase, the underlying technology may one day become an everyday part of photography with smartphones.

VDIM converts still images into video by creating missing frames using a diffusion model similar to those found in Midjourney, DALL-E, and Google's own Imagen 2.

It begins by creating a low-resolution version of the complete final video. This is done by running a cascade of diffusion models in sequence and continuously refining the video. This first step allows VDIM to capture the motion and dynamics of the final output.

This information is then passed to a higher resolution step, where it is upscaled and improved to be closer to the input image and to make the motion more fluid.

One potential use case for VDIM that the team examined in their research paper is video restoration; AI has been used to improve old images, which could be useful for cleaning up old family films or restoring films with broken frames.

Older films may have burned out frames in the middle of a sequence, making them difficult to see. Or there are several frames with scratches.

VDIM is given a clean frame at the beginning and end and used to recreate the motion between those two points.

Since VDIM is a research project, no one outside of the Google DeepMind research team has actually used it yet, but the example of the video clip is a good start for a new type of AI video.

Examples of videos shared by Google DeepMind include the start of a box-cart race with only two still images.

Another video showing a woman on a swing transformed two still images into a flowing swing movement.

Personally, I think this is one research project that Google should pursue and find a way to implement in live software, especially in video restoration. Especially if it can be extended beyond a few seconds or a few dozen frames.

Categories