10
Learning Hierarchical Shape Segmentation and Labeling from Online Repositories LI YI, Stanford University LEONIDAS GUIBAS, Stanford University AARON HERTZMANN, Adobe Research VLADIMIR G. KIM, Adobe Research HAO SU, Stanford University ERSIN YUMER, Adobe Research Car Body Cluster76 Door Wheel Rim Tire Cluster71 Cluster72 SteeringWheel Trunk Dashboard Window Headlight Hood Mirror Roof Floor Wiper Cluster30 Cluster41 Bumper Car Car Car Spoiler RightWheel Car Wheel Wheel Wheel Wheel (a) Input Noisy Scene Graphs Learned Model (c) Hierarchical Shape Segmentation (b) Novel Instance Fig. 1. Large online model repositories contain abundant additional data beyond 3D geometry, such as part labels and artist’s part decompositions, flat or hierarchical. We tap into this trove of sparse and noisy noisy data to train a network for simultaneous hierarchical shape structure decomposition and labeling. Our method learns to take new geometry, and segment it into parts, label the parts, and place them in a hierarchy. In this paper, we visualize scene graphs with a circular visualization, in which the root node is near the center. Blue lines indicate parent-child relationships, and red dashed arcs connect siblings. The input geometry in online databases are broken as connected components, visualized in the input as random colors. We propose a method for converting geometric shapes into hierarchically segmented parts with part labels. Our key idea is to train category-specic models from the scene graphs and part names that accompany 3D shapes in public repositories. ese freely-available annotations represent an enor- mous, untapped source of information on geometry. However, because the models and corresponding scene graphs are created by a wide range of mod- elers with dierent levels of expertise, modeling tools, and objectives, these models have very inconsistent segmentations and hierarchies with sparse and noisy textual tags. Our method involves two analysis steps. First, we perform a joint optimization to simultaneously cluster and label parts in the Li Yi co-developed and implemented the method; the other authors are in alphabetical order. is work started during Li Yi’s internship at Adobe Research. is work is supported by NSF grants DMS-1521608 and DMS-1546206, ONR grant MURI N00014- 13-1-0341, and Adobe Systems Inc. Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for prot or commercial advantage and that copies bear this notice and the full citation on the rst page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permied. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specic permission and/or a fee. Request permissions from [email protected]. © 2017 ACM. 0730-0301/2017/7-ART70 $15.00 DOI: hp://dx.doi.org/10.1145/3072959.3073652 database while also inferring a canonical tag dictionary and part hierarchy. We then use this labeled data to train a method for hierarchical segmentation and labeling of new 3D shapes. We demonstrate that our method can mine complex information, detecting hierarchies in man-made objects and their constituent parts, obtaining ner scale details than existing alternatives. We also show that, by performing domain transfer using a few supervised ex- amples, our technique outperforms fully-supervised techniques that require hundreds of manually-labeled models. CCS Concepts: Computing methodologies Machine learning ap- proaches; Shape analysis; Additional Key Words and Phrases: hierarchical shape structure, shape labeling, learning, Siamese networks ACM Reference format: Li Yi, Leonidas Guibas, Aaron Hertzmann, Vladimir G. Kim, Hao Su, and Ersin Yumer. 2017. Learning Hierarchical Shape Segmentation and Labeling from Online Repositories. ACM Trans. Graph. 36, 4, Article 70 (July 2017), 12 pages. DOI: hp://dx.doi.org/10.1145/3072959.3073652 ACM Transactions on Graphics, Vol. 36, No. 4, Article 70. Publication date: July 2017.

Learning Hierarchical Shape Segmentation and Labeling from …s2017.siggraph.org/sites/default/files/firstpages_part... · 2017-07-25 · Learning Hierarchical Shape Segmentation

  • Upload
    others

  • View
    2

  • Download
    0

Embed Size (px)

Citation preview

Page 1: Learning Hierarchical Shape Segmentation and Labeling from …s2017.siggraph.org/sites/default/files/firstpages_part... · 2017-07-25 · Learning Hierarchical Shape Segmentation

Learning Hierarchical Shape Segmentation and Labelingfrom Online Repositories

LI YI, Stanford UniversityLEONIDAS GUIBAS, Stanford UniversityAARON HERTZMANN, Adobe ResearchVLADIMIR G. KIM, Adobe ResearchHAO SU, Stanford UniversityERSIN YUMER, Adobe Research

Car

Body

Cluster76

Door

Wheel

Rim

Tire

Cluster71Cluster72SteeringWheel

TrunkDashboard

Window

Headlight

Hood

Mirror

RoofFloor Wiper

Cluster30Cluster41

BumperCar

Car Car

SpoilerRightWheel

Car

Wheel

Wheel Wheel

Wheel

… …(a) Input Noisy Scene Graphs

Learned Model

(c) Hierarchical Shape Segmentation(b) Novel Instance

Fig. 1. Large online model repositories contain abundant additional data beyond 3D geometry, such as part labels and artist’s part decompositions, flat orhierarchical. We tap into this trove of sparse and noisy noisy data to train a network for simultaneous hierarchical shape structure decomposition and labeling.Our method learns to take new geometry, and segment it into parts, label the parts, and place them in a hierarchy. In this paper, we visualize scene graphswith a circular visualization, in which the root node is near the center. Blue lines indicate parent-child relationships, and red dashed arcs connect siblings. Theinput geometry in online databases are broken as connected components, visualized in the input as random colors.

We propose a method for converting geometric shapes into hierarchicallysegmented parts with part labels. Our key idea is to train category-speci�cmodels from the scene graphs and part names that accompany 3D shapes inpublic repositories. �ese freely-available annotations represent an enor-mous, untapped source of information on geometry. However, because themodels and corresponding scene graphs are created by a wide range of mod-elers with di�erent levels of expertise, modeling tools, and objectives, thesemodels have very inconsistent segmentations and hierarchies with sparseand noisy textual tags. Our method involves two analysis steps. First, weperform a joint optimization to simultaneously cluster and label parts in the

Li Yi co-developed and implemented the method; the other authors are in alphabeticalorder. �is work started during Li Yi’s internship at Adobe Research. �is work issupported by NSF grants DMS-1521608 and DMS-1546206, ONR grant MURI N00014-13-1-0341, and Adobe Systems Inc.Permission to make digital or hard copies of all or part of this work for personal orclassroom use is granted without fee provided that copies are not made or distributedfor pro�t or commercial advantage and that copies bear this notice and the full citationon the �rst page. Copyrights for components of this work owned by others than ACMmust be honored. Abstracting with credit is permi�ed. To copy otherwise, or republish,to post on servers or to redistribute to lists, requires prior speci�c permission and/or afee. Request permissions from [email protected].© 2017 ACM. 0730-0301/2017/7-ART70 $15.00DOI: h�p://dx.doi.org/10.1145/3072959.3073652

database while also inferring a canonical tag dictionary and part hierarchy.We then use this labeled data to train a method for hierarchical segmentationand labeling of new 3D shapes. We demonstrate that our method can minecomplex information, detecting hierarchies in man-made objects and theirconstituent parts, obtaining �ner scale details than existing alternatives. Wealso show that, by performing domain transfer using a few supervised ex-amples, our technique outperforms fully-supervised techniques that requirehundreds of manually-labeled models.

CCS Concepts: •Computing methodologies → Machine learning ap-proaches; Shape analysis;

Additional Key Words and Phrases: hierarchical shape structure, shapelabeling, learning, Siamese networks

ACM Reference format:Li Yi, Leonidas Guibas, Aaron Hertzmann, Vladimir G. Kim, Hao Su, and ErsinYumer. 2017. Learning Hierarchical Shape Segmentation and Labelingfrom Online Repositories. ACM Trans. Graph. 36, 4, Article 70 (July 2017),12 pages.DOI: h�p://dx.doi.org/10.1145/3072959.3073652

ACM Transactions on Graphics, Vol. 36, No. 4, Article 70. Publication date: July 2017.

Page 2: Learning Hierarchical Shape Segmentation and Labeling from …s2017.siggraph.org/sites/default/files/firstpages_part... · 2017-07-25 · Learning Hierarchical Shape Segmentation

Convolutional Neural Networks on Surfaces via Seamless Toric Covers

HAGGAI MARON, Weizmann Institute of ScienceMEIRAV GALUN, NOAM AIGERMAN, MIRI TROPE, NADAV DYM, Weizmann Institute of ScienceERSIN YUMER, VLADIMIR G. KIM, Adobe ResearchYARON LIPMAN, Weizmann Institute of Science

The recent success of convolutional neural networks (CNNs) for imageprocessing tasks is inspiring research efforts attempting to achieve similarsuccess for geometric tasks. One of the main challenges in applying CNNsto surfaces is defining a natural convolution operator on surfaces.

In this paper we present a method for applying deep learning to sphere-type shapes using a global seamless parameterization to a planar flat-torus,for which the convolution operator is well defined. As a result, the standarddeep learning framework can be readily applied for learning semantic, high-level properties of the shape. An indication of our success in bridging thegap between images and surfaces is the fact that our algorithm succeedsin learning semantic information from an input of raw low-dimensionalfeature vectors.

We demonstrate the usefulness of our approach by presenting two ap-plications: human body segmentation, and automatic landmark detectionon anatomical surfaces. We show that our algorithm compares favorablywith competing geometric deep-learning algorithms for segmentation tasks,and is able to produce meaningful correspondences on anatomical surfaceswhere hand-crafted features are bound to fail.

CCSConcepts: •Computingmethodologies→Neural networks; Shapeanalysis;

Additional Key Words and Phrases: Geometric deep learning, Convolutionalneural network, Shape analysis, shape segmentation

ACM Reference format:Haggai Maron, Meirav Galun, Noam Aigerman, Miri Trope, Nadav Dym,Ersin Yumer, Vladimir G. Kim, and Yaron Lipman. 2017. Convolutional NeuralNetworks on Surfaces via Seamless Toric Covers. ACM Trans. Graph. 36, 4,Article 71 (July 2017), 10 pages.DOI: http://dx.doi.org/10.1145/3072959.3073616

1 INTRODUCTIONA recent research effort in the geometry processing and vision com-munities is to translate the incredible success of deep convolutionalneural networks (CNN) to geometric settings. One particularly in-teresting scenario is applying CNNs for supervised learning of func-tions or labels over curved two dimensional sphere-like surfaces.This is a common problem in analyzing human bodies, anatomical,and medical data.

Applying CNNs to extrinsic surface representation such as volu-metric grids [Qi et al. 2016b] or depth map projections on extrinsic2D cameras [Wei et al. 2016] requires working with 3D grids, ordealing with many camera and lighting parameters, and is very sen-sitive to deformations (e.g., human pose changes). While it might be

Permission to make digital or hard copies of all or part of this work for personal orclassroom use is granted without fee provided that copies are not made or distributedfor profit or commercial advantage and that copies bear this notice and the full citationon the first page. Copyrights for components of this work owned by others than ACMmust be honored. Abstracting with credit is permitted. To copy otherwise, or republish,to post on servers or to redistribute to lists, requires prior specific permission and/or afee. Request permissions from [email protected].© 2017 ACM. 0730-0301/2017/7-ART71 $15.00DOI: http://dx.doi.org/10.1145/3072959.3073616

a

a

b

ba

a

b

ba

a

b

ba

a

b

b

a

a

b

b

a

a b

b

a

a

b

b

a

ab

b

Fig. 1. Defining convolutional neural networks on sphere-like surfaces: weconstruct a torus (top-right) from four exact copies of the surface (top-left)and map it to the flat-torus (bottom-right) to define a local, translationinvariant convolution (bottom-left). This construction is unique up to achoice of three points on the surface (colored disks).

possible to learn a representation that is invariant to these deforma-tions, this requires substantial amount of training data. In contrast,the goal of our paper is to provide an intrinsic representation thatwould enable applying CNNs directly to the surfaces.

One of the main challenges in applying CNN to curved surfacesis that there is no clear generalization of the convolution opera-tor. In particular, two properties of the convolution operator thatare considered pivotal in the success of the CNN architectures arelocality and translation invariance. It is possible to parameterize asurface locally on a geodesic patch around a point [Masci et al. 2015],however, this representation lacks global context. Sinha et al. [2016]proposed geometry images to globally parameterize sphere-likesurface into an image, however, although continuous, their repre-sentation is not seamless (the parameterization is dependent on thecuts made for its computation), their space of parameterizations,namely area-preserving maps has a very high number of degrees offreedom (i.e., it requires an infinite number of point constraints touniquely define a mapping) and therefore can represent the sameshape in many different arbitrary ways in the image (see Figure 2for an example). Lastly, the convolution on the geometry image isnot translation invariant.

ACM Transactions on Graphics, Vol. 36, No. 4, Article 71. Publication date: July 2017.

Page 3: Learning Hierarchical Shape Segmentation and Labeling from …s2017.siggraph.org/sites/default/files/firstpages_part... · 2017-07-25 · Learning Hierarchical Shape Segmentation

O-CNN: Octree-based Convolutional Neural Networks for 3D ShapeAnalysis

PENG-SHUAI WANG, Tsinghua University and Microsoft Research AsiaYANG LIU, Microsoft Research AsiaYU-XIAO GUO, University of Electronic Science and Technology of China and Microsoft Research AsiaCHUN-YU SUN, Tsinghua University and Microsoft Research AsiaXIN TONG, Microsoft Research Asia

normal field

octree input (d-depth)

...

convolution pooling convolution pooling

d d-1 3 2

...

... ...

Fig. 1. An illustration of our octree-based convolutional neural network (O-CNN). Our method represents the input shape with an octree and feeds theaveraged normal vectors stored in the finest leaf octants to the CNN as input. All the CNN operations are efficiently executed on the GPU and the resultingfeatures are stored in the octree structure. Numbers inside the blue dashed square denote the depth of the octants involved in computation.

We present O-CNN, an Octree-based Convolutional Neural Network (CNN)for 3D shape analysis. Built upon the octree representation of 3D shapes,our method takes the average normal vectors of a 3D model sampled in thefinest leaf octants as input and performs 3D CNN operations on the octantsoccupied by the 3D shape surface. We design a novel octree data structure toefficiently store the octant information and CNN features into the graphicsmemory and execute the entire O-CNN training and evaluation on theGPU. O-CNN supports various CNN structures and works for 3D shapes indifferent representations. By restraining the computations on the octantsoccupied by 3D surfaces, the memory and computational costs of the O-CNNgrow quadratically as the depth of the octree increases, which makes the 3DCNN feasible for high-resolution 3D models. We compare the performanceof the O-CNN with other existing 3D CNN solutions and demonstrate theefficiency and efficacy of O-CNN in three shape analysis tasks, includingobject classification, shape retrieval, and shape segmentation.

CCS Concepts: • Computing methodologies → Mesh models; Point-based models; Neural networks;

Additional Key Words and Phrases: octree, convolutional neural network,object classification, shape retrieval, shape segmentation

Permission to make digital or hard copies of all or part of this work for personal orclassroom use is granted without fee provided that copies are not made or distributedfor profit or commercial advantage and that copies bear this notice and the full citationon the first page. Copyrights for components of this work owned by others than ACMmust be honored. Abstracting with credit is permitted. To copy otherwise, or republish,to post on servers or to redistribute to lists, requires prior specific permission and/or afee. Request permissions from [email protected].© 2017 Association for Computing Machinery.0730-0301/2017/7-ART72 $15.00https://doi.org/10.1145/3072959.3073608

ACM Reference format:Peng-Shuai Wang, Yang Liu, Yu-Xiao Guo, Chun-Yu Sun, and Xin Tong.2017. O-CNN: Octree-based Convolutional Neural Networks for 3D ShapeAnalysis. ACM Trans. Graph. 36, 4, Article 72 (July 2017), 11 pages.https://doi.org/10.1145/3072959.3073608

1 INTRODUCTIONWith recent advances in low-cost 3D acquisition devices and 3Dmodeling tools, the amount of 3D models created by end usershas been increasing quickly. Analyzing and understanding these3D shapes, as for classification, segmentation, and retrieval, havebecome more and more important for many graphics and visionapplications. A key technique for these shape analysis tasks is toextract features of 3D models that can sufficiently characterize theirshapes and parts.

In the computer vision field, convolutional neural networks (CNNs)are widely used for image feature extraction and have demonstratedtheir advantages over manually-crafted solutions in most imageanalysis and understanding tasks. However, it is a non-trivial taskto adapt a CNN designed for regularly sampled 2D images to 3Dshapes modeled by irregular triangle meshes or point clouds. A setof methods convert the 3D shapes to regularly sampled representa-tions and apply a CNN to them. Voxel-based methods [Maturanaand Scherer 2015; Wu et al. 2015] rasterize a 3D shape as an indi-cator function or distance function sampled over dense voxels andapply a 3D CNN over the entire 3D volume. Since the memory andcomputation cost grow cubically as the voxel resolution increases,

ACM Transactions on Graphics, Vol. 36, No. 4, Article 72. Publication date: July 2017.

Page 4: Learning Hierarchical Shape Segmentation and Labeling from …s2017.siggraph.org/sites/default/files/firstpages_part... · 2017-07-25 · Learning Hierarchical Shape Segmentation

ClothCap: Seamless 4D Clothing Capture and Retargeting

GERARD PONS-MOLL∗, Max Planck Institute for Intelligent Systems, Tübingen, GermanySERGI PUJADES∗, Max Planck Institute for Intelligent Systems, Tübingen, GermanySONNY HU, Body Labs, New York, NY, USAMICHAEL J. BLACK, Max Planck Institute for Intelligent Systems, Tübingen, Germany

Fig. 1. ClothCap. From left to right: (1) An example 3D textured scan that is part of a 4D sequence. (2) Our multi-part aligned mesh model, layered over thebody. (3) The estimated minimally clothed shape (MCS) under the clothing. (4) The body made fatter and dressed in the same clothing. Note that the clothingadapts in a natural way to the new body shape. (5) This new body shape posed in a new, never seen, pose. This illustrates how ClothCap supports a range ofapplications related to clothing capture, modeling, retargeting, reposing, and try-on.

Designing and simulating realistic clothing is challenging. Previous methodsaddressing the capture of clothing from 3D scans have been limited to singlegarments and simple motions, lack detail, or require specialized texturepatterns. Here we address the problem of capturing regular clothing onfully dressed people in motion. People typically wear multiple pieces ofclothing at a time. To estimate the shape of such clothing, track it over time,and render it believably, each garment must be segmented from the othersand the body. Our ClothCap approach uses a new multi-part 3D model ofclothed bodies, automatically segments each piece of clothing, estimates theminimally clothed body shape and pose under the clothing, and tracks the3D deformations of the clothing over time. We estimate the garments andtheir motion from 4D scans; that is, high-resolution 3D scans of the subjectin motion at 60 fps. ClothCap is able to capture a clothed person in motion,extract their clothing, and retarget the clothing to new body shapes; thisprovides a step towards virtual try-on.

CCS Concepts: • Computing methodologies → Computer graphics;Shape modeling; Animation; Mesh models;

Additional Key Words and Phrases: Clothing, animation, 3D shape, 3D seg-mentation, try-on, performance capture.

ACM Reference format:Gerard Pons-Moll, Sergi Pujades, Sonny Hu, and Michael J. Black. 2017.ClothCap: Seamless 4DClothing Capture and Retargeting.ACMTrans. Graph.36, 4, Article 73 (July 2017), 15 pages.DOI: http://dx.doi.org/10.1145/3072959.3073711

∗Equal contribution

Permission to make digital or hard copies of part or all of this work for personal orclassroom use is granted without fee provided that copies are not made or distributedfor profit or commercial advantage and that copies bear this notice and the full citationon the first page. Copyrights for third-party components of this work must be honored.For all other uses, contact the owner/author(s).© 2017 Copyright held by the owner/author(s). 0730-0301/2017/7-ART73 $15.00DOI: http://dx.doi.org/10.1145/3072959.3073711

1 INTRODUCTIONDressing virtual avatars and animating them with high quality, visu-ally plausible, results is a challenging task. Highly realistic physicalsimulation of clothing on human bodies in motion is complex: cloth-ing models are laborious to construct, patterns must be graded sothat they can be sized to different characters, and the physical param-eters of the cloth must be known. Instead, we propose a data-drivenclothing capture (ClothCap) approach; we capture dynamic clothingon humans from 4D scans and transform it to more easily dressvirtual avatars.

We proceed by capturing the garment geometry in motion on abody, estimate the body shape and pose under clothing, and seg-ment and extract the clothing pieces. We then retarget the capturedclothing to new body shapes and poses. To that end, we develop anovel multi-part mesh model and show how to segment, track, andrecover garment shape from sequences of 3D scans (see Fig. 1).Previous methods for 3D garment capture are not sufficiently

accurate or detailed to compete with physical simulation. Existingcapture methods suffer from low resolution, static shapes, simplebody motions, capture only one clothing piece, or do not segmentthe clothing from the body. The key problems to solve include high-quality capture, segmentation, tracking of surface shape, as well asbody shape and pose estimation.We address these problems by introducing a novel multi-mesh

representation that we fit to sequences of 3D scans captured witha high-resolution 4D scanner (Fig. 1 (1)). Our method requires asinput a point cloud with texture, here we used the same active stereosystem as in (Pons-Moll et al. 2015a). The multi-mesh representationis consistent with how clothing is worn, modeling the changingvisibility of surfaces over time. It further allows us to address the

ACM Transactions on Graphics, Vol. 36, No. 4, Article 73. Publication date: July 2017.

Page 5: Learning Hierarchical Shape Segmentation and Labeling from …s2017.siggraph.org/sites/default/files/firstpages_part... · 2017-07-25 · Learning Hierarchical Shape Segmentation

Fusing State Spaces for Markov Chain Monte Carlo Rendering

HISANARI OTSU, The University of TokyoANTON S. KAPLANYAN, NVIDIAJOHANNES HANIKA, Karlsruhe Institute of TechnologyCARSTEN DACHSBACHER, Karlsruhe Institute of TechnologyTOSHIYA HACHISUKA, The University of Tokyo

Reference MMLT MLT Ours

Door

Necklace

Fig. 1. Equal-time rendering (20 minutes) of two scenes. The door scene contains glossy and specular materials and is illuminated by indirect lights withdifficult visibility. This scene can be effectively rendered with MLT [Veach and Guibas 1997], but exhibits suboptimal performance with multiplexed MLT(MMLT) [Hachisuka et al. 2014]. On the other hand, the necklace scene, which is characterized by glossy interreflections, shows the opposite behavior. Ourframework makes it possible to combine mutation strategies using the two different state spaces of MLT and MMLT, enabling the combination of specializedmutation strategies, and resulting in a general algorithm that works robustly in many cases.

Rendering algorithms using Markov chain Monte Carlo (MCMC) currently

build upon two different state spaces. One of them is the path space, where

the algorithms operate on the vertices of actual transport paths. The other

state space is the primary sample space, where the algorithms operate on

sequences of numbers used for generating transport paths. While the two

state spaces are related by the sampling procedure of transport paths, all

existing MCMC rendering algorithms are designed to work within only

one of the state spaces. We propose a first framework which provides a

comprehensive connection between the path space and the primary sample

Permission to make digital or hard copies of all or part of this work for personal or

classroom use is granted without fee provided that copies are not made or distributed

for profit or commercial advantage and that copies bear this notice and the full citation

on the first page. Copyrights for components of this work owned by others than the

author(s) must be honored. Abstracting with credit is permitted. To copy otherwise, or

republish, to post on servers or to redistribute to lists, requires prior specific permission

and/or a fee. Request permissions from [email protected].

© 2017 Copyright held by the owner/author(s). Publication rights licensed to ACM.

0730-0301/2017/7-ART74 $15.00

DOI: 0.1145/3072959.3073691

space. Using this framework, we can use mutation strategies designed for one

space with mutation strategies in the respective other space. As a practical

example, we take a combination of manifold exploration and multiplexed

Metropolis light transport using our framework. Our results show that the

simultaneous use of the two state spaces improves the robustness of MCMC

rendering. By combining efficient local exploration in the path space with

global jumps in primary sample space, our method achieves more uniform

convergence as compared to using only one space.

CCS Concepts: • Computing methodologies → Ray tracing;

Additional KeyWords and Phrases: global illumination, Markov chain Monte

Carlo

ACM Reference format:Hisanari Otsu, Anton S. Kaplanyan, Johannes Hanika, Carsten Dachsbacher,

and Toshiya Hachisuka. 2017. Fusing State Spaces for Markov Chain Monte

Carlo Rendering. ACM Trans. Graph. 36, 4, Article 74 (July 2017), 10 pages.

DOI: 0.1145/3072959.3073691

ACM Transactions on Graphics, Vol. 36, No. 4, Article 74. Publication date: July 2017.

Page 6: Learning Hierarchical Shape Segmentation and Labeling from …s2017.siggraph.org/sites/default/files/firstpages_part... · 2017-07-25 · Learning Hierarchical Shape Segmentation

Charted Metropolis Light Transport

JACOPO PANTALEONI, NVIDIA

Fig. 1. Charted Metropolis light transport considers path sampling methods and their primary sample space coordinates as charts of the pathspace, allowing to easily jump between them. In particular, it does so without requiring classical invertibility of the sampling methods, making thealgorithm practical even in the presence of complex materials.

In this manuscript, inspired by a simpler reformulation of primary samplespace Metropolis light transport, we derive a novel family of general Markovchain Monte Carlo algorithms called charted Metropolis-Hastings, that in-troduces the notion of sampling charts to extend a given sampling domainand make it easier to sample the desired target distribution and escape fromlocal maxima through coordinate changes. We further apply the novel al-gorithms to light transport simulation, obtaining a new type of algorithmcalled charted Metropolis light transport, that can be seen as a bridge betweenprimary sample space and path space Metropolis light transport. The newalgorithms require to provide only right inverses of the sampling functions,a property that we believe crucial to make them practical in the context oflight transport simulation.

Permission to make digital or hard copies of all or part of this work for personal orclassroom use is granted without fee provided that copies are not made or distributedfor profit or commercial advantage and that copies bear this notice and the full citationon the first page. Copyrights for components of this work owned by others than ACMmust be honored. Abstracting with credit is permitted. To copy otherwise, or republish,to post on servers or to redistribute to lists, requires prior specific permission and/or afee. Request permissions from [email protected].© 2017 ACM. 0730-0301/2017/7-ART75 $15.00DOI: http://dx.doi.org/10.1145/3072959.3073677

CCS Concepts: • Computing methodologies→ Ray tracing; Reflectancemodeling; Massively parallel algorithms; Graphics processors;

Additional Key Words and Phrases: global illumination, light transport sim-ulation, Markov Chain Monte Carlo

ACM Reference format:Jacopo Pantaleoni. 2017. Charted Metropolis Light Transport. ACM Trans.Graph. 36, 4, Article 75 (July 2017), 14 pages.DOI: http://dx.doi.org/10.1145/3072959.3073677

1 INTRODUCTIONLight transport simulation can be notoriously hard. The main prob-lem is that forming an image requires evaluating millions of infinitedimensional integrals, whose integrands, while correlated, may con-tain an infinity of singularities and different modes at disparatefrequencies. Many approaches have been proposed to solve the ren-dering equation, though most of them rely on variants of MonteCarlo integration. One of the most robust algorithms, Metropolislight transport (MLT), has been proposed by Veach and Guibas in

ACM Transactions on Graphics, Vol. 36, No. 4, Article 75. Publication date: July 2017.

Page 7: Learning Hierarchical Shape Segmentation and Labeling from …s2017.siggraph.org/sites/default/files/firstpages_part... · 2017-07-25 · Learning Hierarchical Shape Segmentation

Topology-controlled Reconstruction of Multi-labelled Domains fromCross-sections

ZHIYANG HUANG, Washington University in St. LouisMING ZOU, Washington University in St. LouisNATHAN CARR, Adobe SystemsTAO JU, Washington University in St. Louis

Fig. 1. Given several multi-labeled planes depicting the anatomical regions of a mouse brain (a), reconstruction without topology control (b1) leads toredundant handles for the red and yellow labels (black arrows in c1, d1) and disconnection for the green label (e1). Our method (b2) allows the user to prescribethe topology such that the red label has one tunnel (gray arrow in c2), the yellow label has no tunnels (d2), and the green label is connected (e2). The legendsin (b1,b2) report, for each label in the reconstruction, the genus of each surface component bounding that label (e.g., “0,0” means two surfaces each with genus0). User-specified constraints are colored red.

In this work we present the �rst algorithm for reconstructing multi-labeledmaterial interfaces the allows for explicit topology control. Our algorithmtakes in a set of 2D cross-sectional slices (not necessarily parallel), eachpartitioned by a curve network into labeled regions representing di�erentmaterial types. For each label, the user has the option to constrain the num-ber of connected components and genus. Our algorithm is able to not onlyproduce a material interface that interpolates the curve networks but alsosimultaneously satisfy the topological requirements. Our key innovation isde�ning a space of topology-varying material interfaces, which extends thefamily of level sets in a scalar function, and developing discrete methodsfor sampling distinct topologies in this space. Besides specifying topolog-ical constraints, the user can steer the algorithm interactively, such as byscribbling. We demonstrate, on synthetic and biological shapes, how ouralgorithm opens up new opportunities for topology-aware modeling in themulti-labeled context.

CCS Concepts: • Computing methodologies → Mesh models;

This work is supported by the National Science Foundation, under grants IIS-084607and IIS-1302200, and a gift from Adobe Inc. Ming Zou is now at Waymo.Permission to make digital or hard copies of all or part of this work for personal orclassroom use is granted without fee provided that copies are not made or distributedfor pro�t or commercial advantage and that copies bear this notice and the full citationon the �rst page. Copyrights for components of this work owned by others than ACMmust be honored. Abstracting with credit is permitted. To copy otherwise, or republish,to post on servers or to redistribute to lists, requires prior speci�c permission and/or afee. Request permissions from [email protected].© 2017 ACM. 0730-0301/2017/7-ART76 $15.00DOI: http://dx.doi.org/10.1145/3072959.3073644

Additional Key Words and Phrases: surface reconstruction, material inter-faces, topology, contour interpolation

ACM Reference format:Zhiyang Huang, Ming Zou, Nathan Carr, and Tao Ju. 2017. Topology-controlledReconstruction of Multi-labelled Domains from Cross-sections. ACM Trans.Graph. 36, 4, Article 76 (July 2017), 12 pages.DOI: http://dx.doi.org/10.1145/3072959.3073644

1 INTRODUCTIONComputational modeling of multi-labeled domains arises in manydisciplines, such as biomedicine (e.g., organs made up of multi-ple anatomical regions) and mechanical engineering (e.g., machinepieces made up of blocks of di�erent materials). Such domains areoften represented by the non-manifold network of surfaces that par-tition the domain into labeled sub-domains. This network, knownas the material interface, is widely used in applications includinggeometric processing, physical simulations, and manufacturing.

To be useful for applications, a material interface has to meetcertain correctness criteria. Most importantly, the material interfaceshould be geometrically valid (i.e., free of holes and intersections),so that it de�nes a proper partitioning of the domain into disjointsub-domains. Furthermore, some applications are also sensitive tothe topology of the surface. For example, �uid simulation withinone or more sub-domains can be adversely a�ected if the surfaces

ACM Transactions on Graphics, Vol. 36, No. 4, Article 76. Publication date: July 2017.

Page 8: Learning Hierarchical Shape Segmentation and Labeling from …s2017.siggraph.org/sites/default/files/firstpages_part... · 2017-07-25 · Learning Hierarchical Shape Segmentation

Field-Aligned Online Surface Reconstruction

NICO SCHERTLER, New York University and TU DresdenMARCO TARINI, Università dell’Insubria and ISTI - CNRWENZEL JAKOB, École Polytechnique Fédérale de Lausanne (EPFL)MISHA KAZHDAN, Johns Hopkins UniversitySTEFAN GUMHOLD, TU DresdenDANIELE PANOZZO, New York University

...

Fig. 1. Multiple successively acquired 3D scans (top row) are interactively integrated into a coarse base mesh (bottom row). The user sees the final reconstructedresult at all times, enriched with textures that encode colors and displacement (right-most). The user can decide to reconstruct a triangle or a quad-dominantmesh, which has high isotropy and regularity. [Original Sculpture Courtesy of Michael Defeo, mesh statistics can be found in Table 1]

Today’s 3D scanning pipelines can be classified into two overarching cat-egories: offline, high accuracy methods that rely on global optimizationto reconstruct complex scenes with hundreds of millions of samples, andonline methods that produce real-time but low-quality output, usually fromstructure-from-motion or depth sensors. The method proposed in this paperis the first to combine the benefits of both approaches, supporting onlinereconstruction of scenes with hundreds of millions of samples from high-resolution sensing modalities such as structured light or laser scanners. Thekey property of our algorithm is that it sidesteps the signed-distance com-putation of classical reconstruction techniques in favor of direct filtering,parametrization, and mesh and texture extraction. All of these steps can be

This work was supported by the NSF CAREER award 1652515, NSF award 1422325,MIUR project DSurf, and a fellowship within the FITweltweit program of the GermanAcademic Exchange Service (DAAD).Permission to make digital or hard copies of all or part of this work for personal orclassroom use is granted without fee provided that copies are not made or distributedfor profit or commercial advantage and that copies bear this notice and the full citationon the first page. Copyrights for components of this work owned by others than ACMmust be honored. Abstracting with credit is permitted. To copy otherwise, or republish,to post on servers or to redistribute to lists, requires prior specific permission and/or afee. Request permissions from [email protected].© 2017 ACM. 0730-0301/2017/7-ART77 $15.00DOI: http://dx.doi.org/10.1145/3072959.3073635

realized using only weak notions of spatial neighborhoods, which allowsfor an implementation that scales approximately linearly with the size ofeach dataset that is integrated into a partial reconstruction. Combined, thesealgorithmic differences enable a drastically more efficient output-driveninteractive scanning and reconstruction workflow, where the user is able tosee the final quality field-aligned textured mesh during the entirety of thescanning procedure. Holes or parts with registration problems are displayedin real-time to the user and can be easily resolved by adding further localizedscans, or by adjusting the input point cloud using our interactive editingtools with immediate visual feedback on the output mesh. We demonstratethe effectiveness of our algorithm in conjunction with a state-of-the-artstructured light scanner and optical tracking system and test it on a largevariety of challenging models.

CCS Concepts: • Computing methodologies → Mesh models;

Additional KeyWords and Phrases: Surface Reconstruction, Parameterization

ACM Reference format:Nico Schertler, Marco Tarini, Wenzel Jakob, Misha Kazhdan, Stefan Gumhold,and Daniele Panozzo. 2017. Field-Aligned Online Surface Reconstruction.ACM Trans. Graph. 36, 4, Article 77 (July 2017), 13 pages.DOI: http://dx.doi.org/10.1145/3072959.3073635

ACM Transactions on Graphics, Vol. 36, No. 4, Article 77. Publication date: July 2017.

Page 9: Learning Hierarchical Shape Segmentation and Labeling from …s2017.siggraph.org/sites/default/files/firstpages_part... · 2017-07-25 · Learning Hierarchical Shape Segmentation

Tanks and Temples: Benchmarking Large-Scale Scene Reconstruction

ARNO KNAPITSCH, JAESIK PARK, QIAN-YI ZHOU, and VLADLEN KOLTUN, Intel Labs

Fig. 1. Ground-truth model for the Panther dataset, one of the datasets in the presented benchmark for large-scale scene reconstruction.

We present a benchmark for image-based 3D reconstruction. The benchmarksequences were acquired outside the lab, in realistic conditions. Ground-truth data was captured using an industrial laser scanner. The benchmarkincludes both outdoor scenes and indoor environments. High-resolutionvideo sequences are provided as input, supporting the development of novelpipelines that take advantage of video input to increase reconstruction fi-delity. We report the performance of many image-based 3D reconstructionpipelines on the new benchmark. The results point to exciting challengesand opportunities for future work.

CCS Concepts: • Computing methodologies → Computer graphics;

Additional KeyWords and Phrases: Structure frommotion, multi-view stereo,image-based reconstruction, large-scale scene reconstruction

ACM Reference format:Arno Knapitsch, Jaesik Park, Qian-Yi Zhou, and Vladlen Koltun. 2017. Tanksand Temples: Benchmarking Large-Scale Scene Reconstruction. ACM Trans.Graph. 36, 4, Article 78 (July 2017), 13 pages.http://dx.doi.org/10.1145/3072959.3073599

Permission to make digital or hard copies of all or part of this work for personal orclassroom use is granted without fee provided that copies are not made or distributedfor profit or commercial advantage and that copies bear this notice and the full citationon the first page. Copyrights for components of this work owned by others than ACMmust be honored. Abstracting with credit is permitted. To copy otherwise, or republish,to post on servers or to redistribute to lists, requires prior specific permission and/or afee. Request permissions from [email protected].© 2017 Association for Computing Machinery.0730-0301/2017/7-ART78 $15.00http://dx.doi.org/10.1145/3072959.3073599

1 INTRODUCTIONIn the past decade, structure from motion (SfM) and multi-viewstereo (MVS) techniques have advanced to enable remarkable re-constructions of landmark scenes from community photo collec-tions [Agarwal et al. 2011; Shan et al. 2013; Snavely et al. 2008].Due to these accomplishments, image-based reconstruction is de-servedly considered one of the great successes of visual computing,combining rigorous theory [Hartley and Zisserman 2000; Triggset al. 2000], advanced computational methods [Agarwal et al. 2010;Furukawa and Hernández 2015; Wu et al. 2011], and a culture ofopen software development [Fuhrmann et al. 2015; Furukawa 2011;Moulon et al. 2016; Schönberger 2016; Snavely 2010; Wu 2011].

Nonetheless, existing reconstruction techniques have significantlimitations. Take a nearby camera, walk around a building whilerecording a video, and feed the data into a standard SfM+MVSpipeline. There is a good chance that the result will not be a cleanand accurate reconstruction of the building. For an even greaterchallenge, take a video while walking around the interior of yourresidence and use that. A clean and complete 3D reconstruction ofthe environment is unlikely to emerge.

How can these limitations coexist with the remarkable successesof image-based reconstruction? Reconstruction from communityphoto collections involves dealing with massive and highly redun-dant datasets. Processing such datasets brings up significant chal-lenges in system and algorithm design [Agarwal et al. 2011; Frahmet al. 2010; Heinly et al. 2015; Schönberger and Frahm 2016; Wu2013]. But it also affords significant freedom in automatically se-lecting a subset of the input data that can be successfully recon-structed [Furukawa et al. 2010; Goesele et al. 2007; Li et al. 2008;Schönberger et al. 2016]. Such data selection enables challenginginput to be discarded and can leave limitations in the underlying

ACM Transactions on Graphics, Vol. 36, No. 4, Article 78. Publication date: July 2017.

Page 10: Learning Hierarchical Shape Segmentation and Labeling from …s2017.siggraph.org/sites/default/files/firstpages_part... · 2017-07-25 · Learning Hierarchical Shape Segmentation

Dip Transform for 3D Shape Reconstruction

KFIR ABERMAN∗, AICFVE Beijing Film Academy, Tel-Aviv UniversityOREN KATZIR*, AICFVE Beijing Film Academy, Tel-Aviv UniversityQIANG ZHOU, Shandong UniversityZEGANG LUO, Shandong UniversityANDREI SHARF, Ben-Gurion University of the NEGEV, AICFVE Beijing Film AcademyCHEN GREIF, University of British ColombiaBAOQUAN CHEN†, Shandong UniversityDANIEL COHEN-OR, Tel-Aviv University

Fig. 1. 3D scanning using a dip scanner. The object is dipped using a robot arm in a bath of water (le�), acquiring a dip transform. The quality of thereconstruction is improving as the number of dipping orientations is increased (from le� to right).

�e paper presents a novel three-dimensional shape acquisition and recon-struction method based on the well-known Archimedes equality between�uid displacement and the submerged volume. By repeatedly dipping a shapein liquid in di�erent orientations and measuring its volume displacement,we generate the dip transform: a novel volumetric shape representation thatcharacterizes the object’s surface. �e key feature of our method is that itemploys �uid displacements as the shape sensor. Unlike optical sensors, theliquid has no line-of-sight requirements, it penetrates cavities and hiddenparts of the object, as well as transparent and glossy materials, thus bypass-ing all visibility and optical limitations of conventional scanning devices.Our new scanning approach is implemented using a dipping robot armand a bath of water, via which it measures the water elevation. We showresults of reconstructing complex 3D shapes and evaluate the quality of thereconstruction with respect to the number of dips.

CCS Concepts: •Computing methodologies→ Shape modeling;

Additional Key Words and Phrases: shape reconstruction, volume, dataacquisition

ACM Reference format:K�r Aberman, Oren Katzir*, Qiang Zhou, Zegang Luo, Andrei Sharf, ChenGreif, Baoquan Chen, and Daniel Cohen-Or. 2017. Dip Transform for 3D

∗Joint �rst authors.†Corresponding author.

Permission to make digital or hard copies of all or part of this work for personal orclassroom use is granted without fee provided that copies are not made or distributedfor pro�t or commercial advantage and that copies bear this notice and the full citationon the �rst page. Copyrights for components of this work owned by others than ACMmust be honored. Abstracting with credit is permi�ed. To copy otherwise, or republish,to post on servers or to redistribute to lists, requires prior speci�c permission and/or afee. Request permissions from [email protected].© 2017 ACM. 0730-0301/2017/7-ART79 $15.00DOI: h�p://dx.doi.org/10.1145/3072959.3073693

Shape Reconstruction. ACM Trans. Graph. 36, 79, Article 79 (July 2017),11 pages.DOI: h�p://dx.doi.org/10.1145/3072959.3073693

1 INTRODUCTION3D shape acquisition and reconstruction methods are based on opti-cal devices, most commonly laser scanners, that sample the visibleshape surface. �ese scanners yield point clouds which are o�ennoisy and incomplete. Many techniques have been developed toamend such point clouds, denoise them, and complete their missingparts using various priors [Berger et al. 2014]. All these surface-based reconstruction techniques assume a reasonable sampling rateof the surface. Notably, these techniques fall short in cases wherethe shapes contain highly occluded parts that are inaccessible tothe scanner’s line-of-sight. �us, complex shapes such as the three-elephant statue in Figure 2(a) and Figure 2(b) cannot be properlyacquired nor reconstructed based on conventional (optical) scan-ners. Moreover, some objects, like Figure 2(c), are made of glossy ortransparent materials, which pose another challenge that commonoptics cannot deal with.

In this work, we take a completely di�erent approach to shapereconstruction. �e idea is to cast surface reconstruction into avolumetric problem. Our technique is based on the ancient �uid dis-placement discovery made by Archimedes, stating that: the volumeof �uid displaced is equal to the volume of the part that was submerged.By dipping an object in liquid along an axis, we can measure thedisplacement of the liquid volume displacement and transform itinto a series of thin volume slices of the shape along the dipping axis.By repeatedly dipping the object in various orientations (see Figure

ACM Transactions on Graphics, Vol. 36, No. 79, Article 79. Publication date: July 2017.