See and Think: Disentangling Semantic Scene Completion
Liu, Shice, HU, YU, Zeng, Yiming, Tang, Qiankun, Jin, Beibei, Han, Yinhe, Li, Xiaowei
–Neural Information Processing Systems
Semantic scene completion predicts volumetric occupancy and object category of a 3D scene, which helps intelligent agents to understand and interact with the surroundings. In this work, we propose a disentangled framework, sequentially carrying out 2D semantic segmentation, 2D-3D reprojection and 3D semantic scene completion. This three-stage framework has three advantages: (1) explicit semantic segmentation significantly boosts performance; (2) flexible fusion ways of sensor data bring good extensibility; (3) progress in any subtask will promote the holistic performance. Experimental results show that regardless of inputing a single depth or RGB-D, our framework can generate high-quality semantic scene completion, and outperforms state-of-the-art approaches on both synthetic and real datasets.
Neural Information Processing Systems
Dec-31-2018