Goto

Collaborating Authors

 pifuhd



Review for NeurIPS paper: Geo-PIFu: Geometry and Pixel Aligned Implicit Functions for Single-view Human Reconstruction

Neural Information Processing Systems

Weaknesses: 1. Evaluation setup is not completely clear. Authors do not mention whether they re-trained the baselines (PIFu and DeepHuman) with their images or directly used the pre-trained models. This information is critical as the training data (number of samples and quality of scans) for PIFu was very different than the DeepHuman dataset. In line 290, authors write that they add several FC layers to the fusion network. It is not clear whether the number of parameters were kept the same for Exp-a,b.


Occupancy Planes for Single-view RGB-D Human Reconstruction

Zhao, Xiaoming, Hu, Yuan-Ting, Ren, Zhongzheng, Schwing, Alexander G.

arXiv.org Artificial Intelligence

Single-view RGB-D human reconstruction with implicit functions is often formulated as per-point classification. Specifically, a set of 3D locations within the view-frustum of the camera are first projected independently onto the image and a corresponding feature is subsequently extracted for each 3D location. The feature of each 3D location is then used to classify independently whether the corresponding 3D point is inside or outside the observed object. This procedure leads to sub-optimal results because correlations between predictions for neighboring locations are only taken into account implicitly via the extracted features. For more accurate results we propose the occupancy planes (OPlanes) representation, which enables to formulate single-view RGB-D human reconstruction as occupancy prediction on planes which slice through the camera's view frustum. Such a representation provides more flexibility than voxel grids and enables to better leverage correlations than per-point classification. On the challenging S3D data we observe a simple classifier based on the OPlanes representation to yield compelling results, especially in difficult situations with partial occlusions due to other objects and partial visibility, which haven't been addressed by prior work.


IntegratedPIFu: Integrated Pixel Aligned Implicit Function for Single-view Human Reconstruction

Chan, Kennard Yanting, Lin, Guosheng, Zhao, Haiyu, Lin, Weisi

arXiv.org Artificial Intelligence

We propose IntegratedPIFu, a new pixel-aligned implicit model that builds on the foundation set by PIFuHD. IntegratedPIFu shows how depth and human parsing information can be predicted and capitalized upon in a pixel-aligned implicit model. In addition, IntegratedPIFu introduces depth-oriented sampling, a novel training scheme that improve any pixel-aligned implicit model's ability to reconstruct important human features without noisy artefacts. Lastly, IntegratedPIFu presents a new architecture that, despite using less model parameters than PIFuHD, is able to improves the structural correctness of reconstructed meshes. Our results show that IntegratedPIFu significantly outperforms existing state-of-the-arts methods on single-view human reconstruction. We provide the code in our supplementary materials.


PIFuHD: High-Resolution 3D Human Digitization (CVPR2020 Oral, Video Results)

#artificialintelligence

Although current approaches have demonstrated the potential in real world settings, they still fail to produce reconstructions with the level of detail often present in the input images. We argue that this limitation stems primarily form two conflicting requirements; accurate predictions require large context, but precise predictions require high resolution. Due to memory limitations in current hardware, previous approaches tend to take low resolution images as input to cover large spatial context, and produce less precise (or low resolution) 3D estimates as a result. We address this limitation by formulating a multi-level architecture that is end-to-end trainable. A coarse level observes the whole image at lower resolution and focuses on holistic reasoning.