Timezone: »

DISN: Deep Implicit Surface Network for High-quality Single-view 3D Reconstruction
Qiangeng Xu · Weiyue Wang · Duygu Ceylan · Radomir Mech · Ulrich Neumann

Wed Dec 11 10:45 AM -- 12:45 PM (PST) @ East Exhibition Hall B + C #62

Reconstructing 3D shapes from single-view images has been a long-standing research problem. In this paper, we present DISN, a Deep Implicit Surface Net- work which can generate a high-quality detail-rich 3D mesh from a 2D image by predicting the underlying signed distance fields. In addition to utilizing global image features, DISN predicts the projected location for each 3D point on the 2D image and extracts local features from the image feature maps. Combin- ing global and local features significantly improves the accuracy of the signed distance field prediction, especially for the detail-rich areas. To the best of our knowledge, DISN is the first method that constantly captures details such as holes and thin structures present in 3D shapes from single-view images. DISN achieves the state-of-the-art single-view reconstruction performance on a variety of shape categories reconstructed from both synthetic and real images. Code is available at https://github.com/laughtervv/DISN. The supplemen- tary can be found at https://xharlie.github.io/images/neurips_ 2019_supp.pdf

Author Information

Qiangeng Xu (USC)
Weiyue Wang (Waymo)
Duygu Ceylan (Adobe Research)
Radomir Mech (Adobe Systems Incorporated)
Ulrich Neumann (USC)

More from the Same Authors