3D motion decomposition for RGBD future dynamic scene synthesis

Xiaojuan QI, Zhengzhe LIU, Qifeng CHEN, Jiaya JIA

Research output: Book Chapters | Papers in Conference ProceedingsConference paper (refereed)Researchpeer-review

16 Citations (Scopus)

Abstract

A future video is the 2D projection of a 3D scene with predicted camera and object motion. Accurate future video prediction inherently requires understanding of 3D motion and geometry of a scene. In this paper, we propose a RGBD scene forecasting model with 3D motion decomposition. We predict ego-motion and foreground motion that are combined to generate a future 3D dynamic scene, which is then projected into a 2D image plane to synthesize future motion, RGB images and depth maps. Optional semantic maps can be integrated. Experimental results on KITTI and Driving datasets show that our model outperforms other state-of-the-arts in forecasting future RGBD dynamic scenes.
Original languageEnglish
Title of host publicationProceedings : 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition, CVPR 2019
PublisherIEEE Computer Society
Pages7665-7674
Number of pages10
ISBN (Electronic)9781728132938
DOIs
Publication statusPublished - Jun 2019
Externally publishedYes

Bibliographical note

Publisher Copyright:
© 2019 IEEE.

Keywords

  • Image and Video Synthesis
  • RGBD sensors and analytics

Fingerprint

Dive into the research topics of '3D motion decomposition for RGBD future dynamic scene synthesis'. Together they form a unique fingerprint.

Cite this