(presented at i3D 2018)
3D Sketching using Multi-View Deep Volumetric Prediction
Abstract
Sketch-based modeling strives to bring the ease and immediacy of drawing to the 3D world. However, while drawings are easy for humans to create, they are very challenging for computers to interpret due to their sparsity and ambiguity. We propose a data-driven approach that tackles this challenge by learning to reconstruct 3D shapes from one or more drawings. At the core of our approach is a deep convolutional neural network (CNN) that predicts occupancy of a voxel grid from a line drawing. This CNN provides an initial 3D reconstruction as soon as the user completes a single drawing of the desired shape. We complement this single-view network with an updater CNN that refines an existing prediction given a new drawing of the shape created from a novel viewpoint. A key advantage of our approach is that we can apply the updater iteratively to fuse information from an arbitrary number of viewpoints, without requiring explicit stroke correspondences between the drawings. We train both CNNs by rendering synthetic contour drawings from hand-modeled shape collections as well as from procedurally-generated abstract shapes. Finally, we integrate our CNNs in an interactive modeling system that allows users to seamlessly draw an object, rotate it to see its 3D reconstruction, and refine it by re-drawing from another vantage point using the 3D reconstruction as guidance.
Video
Downloads and code
Procedural database
- Procedural meshes (Archive, 1 GB) : contains 20K objects in .off format along with contour renderings and viewpoints. Also contains the chairs and vases from ShapeCOSEG.
- Code for procedural generation (Archive, C++ code) : if you want to generate new procedural shapes.
- Code for database generation (Archive, C++ code) : needed to voxelize objects and pack everything together so that you can plug the data easily into our network models.
Networks
- Custom caffe version : all our networks are implemented using Caffe framework. To use our models, you will need this fork from the main framework which contains input layers to load and rotate voxel grids.
- Networks and weights (Zip, 400 MB)
BibTex
@article{delanoy20183d, title={3d sketching using multi-view deep volumetric prediction}, author={Delanoy, Johanna and Aubry, Mathieu and Isola, Phillip and Efros, Alexei A and Bousseau, Adrien}, journal={Proceedings of the ACM on Computer Graphics and Interactive Techniques}, volume={1}, number={1}, pages={1--22}, year={2018}, publisher = {Association for Computing Machinery}, url = {https://dl.acm.org/doi/10.1145/3203197}, doi = {https://doi.org/10.1145/3203197} }