Goto

Collaborating Authors

 scene representation network



Scene Representation Networks: Continuous 3D-Structure-Aware Neural Scene Representations

Neural Information Processing Systems

Unsupervised learning with generative models has the potential of discovering rich representations of 3D scenes. While geometric deep learning has explored 3D-structure-aware representations of scene geometry, these models typically require explicit 3D supervision. Emerging neural scene representations can be trained only with posed 2D images, but existing methods ignore the three-dimensional structure of scenes. We propose Scene Representation Networks (SRNs), a continuous, 3D-structure-aware scene representation that encodes both geometry and appearance. SRNs represent scenes as continuous functions that map world coordinates to a feature representation of local scene properties. By formulating the image formation as a differentiable ray-marching algorithm, SRNs can be trained end-to-end from only 2D images and their camera poses, without access to depth or shape.




Reviews: Scene Representation Networks: Continuous 3D-Structure-Aware Neural Scene Representations

Neural Information Processing Systems

The proposed scene representation is a map phi from the 3D physical space to a feature space encoding properties such as color, distance from closest scene surface, in practice implemented with an MLP. This choice of parametrization results in a natural way of controlling the level of spatial detail the map can achieve with the chosen network capacity, without using a fixed/discrete spatial resolution (as in voxel grids). Images are generated from the scene representation phi, conditioned on a given camara (inclusive of intrinsic and extrinsic parameters) via a differentiable ray marching algorithm. Ray marching is performed using a fixed length unroll of an RNN which can operate on phi effectively decoding distance from the closest surface and learning to correctly how to update the marcher step length. This formulation has the nice bi-product of producing depth maps'for free'.


Scene Representation Networks: Continuous 3D-Structure-Aware Neural Scene Representations

Neural Information Processing Systems

Unsupervised learning with generative models has the potential of discovering rich representations of 3D scenes. While geometric deep learning has explored 3D-structure-aware representations of scene geometry, these models typically require explicit 3D supervision. Emerging neural scene representations can be trained only with posed 2D images, but existing methods ignore the three-dimensional structure of scenes. We propose Scene Representation Networks (SRNs), a continuous, 3D-structure-aware scene representation that encodes both geometry and appearance. SRNs represent scenes as continuous functions that map world coordinates to a feature representation of local scene properties.


Scene Representation Networks: Continuous 3D-Structure-Aware Neural Scene Representations

Sitzmann, Vincent, Zollhoefer, Michael, Wetzstein, Gordon

Neural Information Processing Systems

Unsupervised learning with generative models has the potential of discovering rich representations of 3D scenes. While geometric deep learning has explored 3D-structure-aware representations of scene geometry, these models typically require explicit 3D supervision. Emerging neural scene representations can be trained only with posed 2D images, but existing methods ignore the three-dimensional structure of scenes. We propose Scene Representation Networks (SRNs), a continuous, 3D-structure-aware scene representation that encodes both geometry and appearance. SRNs represent scenes as continuous functions that map world coordinates to a feature representation of local scene properties.