TrajectoryCrafter: Redirecting Camera Trajectory for Monocular Videos via Diffusion Models
YU, Mark, Hu, Wenbo, Xing, Jinbo, Shan, Ying
–arXiv.org Artificial Intelligence
We present TrajectoryCrafter, a novel approach to redirect camera trajectories for monocular videos. By disentangling deterministic view transformations from stochastic content generation, our method achieves precise control over user-specified camera trajectories. We propose a novel dual-stream conditional video diffusion model that concurrently integrates point cloud renders and source videos as conditions, ensuring accurate view transformations and coherent 4D content generation. Instead of leveraging scarce multi-view videos, we curate a hybrid training dataset combining web-scale monocular videos with static multi-view datasets, by our innovative double-reprojection strategy, significantly fostering robust generalization across diverse scenes. Extensive evaluations on multi-view and large-scale monocular videos demonstrate the superior performance of our method.
arXiv.org Artificial Intelligence
Mar-7-2025
- Country:
- Asia (0.28)
- Genre:
- Overview > Innovation (0.34)
- Research Report > Promising Solution (0.48)
- Technology: