Google DeepMind launched multi-view video diffusion model CAT4D, single-view video can also convert multi-view

#News ·2025-01-07

This article is reprinted with the authorization of AIGC Studio public account, please contact the source for reprinting.

Monocular vision 4D reconstruction breakthrough! Google DeepMind and other teams have launched a multi-view video diffusion model CAT4D, which supports the input of a single view of the video, which can be dragged by itself after conversion.

The model can fix the viewing Angle and time respectively, and then output the effects of the static viewing Angle (fixed position) time change, the time constant viewing Angle change, and the viewing Angle time change respectively. As shown in the picture below:

图片图片

Related link

Thesis: https://arxiv.org/pdf/2411.18613

• Home: https://cat-4d.github.io/

Paper introduction

图片

CAT4D: Create any content in 4D form using a multi-view video diffusion model

Working principle

How does it work

Given an input monocular video, we use a multi-view video diffusion model to generate a multi-view video with a new viewpoint. These generated videos are then used to reconstruct the dynamic 3D scene into anamorphic 3D Gauss.

图片

Separate camera and time control

At the heart of CAT4D is the multi-view video diffusion model, which unlocks the control of camera and scene motion. We demonstrate this by giving 3 input images (with camera pose) to generate three types of output sequences: 1) fixed view and change time, 2) changing view and change time, and 3) changing view and change time.

图片

Compare

Compare our approach to baselines for different tasks. Try choosing different tasks and scenarios!

Given 3 input images, we generate three types of output sequences:

1. Fixed viewpoint and change time

2. Change viewpoint and fix time

3. Change viewpoint and change time.

图片

Given just a few pose images of a dynamic scene, we can create a "bullet time" effect by reconstructing a static 3D scene that corresponds to the time of an input view. Three input images are displayed on the left, the first of which is the target bullet time frame.

图片

Comparison of monocular video dynamic scene reconstruction on DyCheck dataset.

图片

TAGS:

  • 400-000-0000

  • No. xx, xxx Street, Suzhou City, Jiangsu Province

  • 123123@163.com

  • wechat

  • WeChat official account

Copyright © 2011-2024 苏州竹子网络科技有限公司 版权所有 ICP:苏ICP备88888888号

friend link