25
An Application of Manifold Learning in Global Shape Descriptors Fereshteh S. Bashiri *, 1,4 , Reihaneh Rostami 2 , Peggy Peissig 4 , Roshan M. D’Souza 3 , and Zeyun Yu *, 1,2 1 Department of Electrical Engineering, University of Wisconsin-Milwaukee, WI, USA 2 Department of Computer Science, University of Wisconsin-Milwaukee, WI, USA 3 Department of Mechanical Engineering, University of Wisconsin-Milwaukee, WI, USA 4 Center for Computational and Biomedical Informatics, Marshfield Clinic Research Institute, WI, USA January 10, 2019 Abstract With the rapid expansion of applied 3D computational vision, shape descriptors have become increasingly important for a wide variety of appli- cations and objects from molecules to planets. Appropriate shape descrip- tors are critical for accurate (and efficient) shape retrieval and 3D model classification. Several spectral-based shape descriptors have been intro- duced by solving various physical equations over a 3D surface model. In this paper, for the first time, we incorporate a specific group of techniques in statistics and machine learning, known as manifold learning, to develop a global shape descriptor in the computer graphics domain. The pro- posed descriptor utilizes the Laplacian Eigenmap technique in which the Laplacian eigenvalue problem is discretized using an exponential weight- ing scheme. As a result, our descriptor eliminates the limitations tied to the existing spectral descriptors, namely dependency on triangular mesh representation and high intra-class quality of 3D models. We also present a straightforward normalization method to obtain a scale-invariant de- scriptor. The extensive experiments performed in this study show that the present contribution provides a highly discriminative and robust shape descriptor under the presence of a high level of noise, random scale varia- tions, and low sampling rate, in addition to the known isometric-invariance property of the Laplace-Beltrami operator. The proposed method signifi- cantly outperforms state-of-the-art algorithms on several non-rigid shape retrieval benchmarks. Keywords: Scale-invariant shape descriptor; Shape retrieval; Manifold Learning; Laplacian Eigenmap * Corresponding authors. 1 arXiv:1901.02508v1 [cs.GR] 8 Jan 2019

An Application of Manifold Learning in Global Shape ...variants have been utilized in multiple FEM-based discretization methods. Another approach to constructing the matrix Ais to

  • Upload
    others

  • View
    2

  • Download
    0

Embed Size (px)

Citation preview

Page 1: An Application of Manifold Learning in Global Shape ...variants have been utilized in multiple FEM-based discretization methods. Another approach to constructing the matrix Ais to

An Application of Manifold Learning in

Global Shape Descriptors

Fereshteh S. Bashiri∗,1,4, Reihaneh Rostami2, Peggy Peissig4,Roshan M. D’Souza3, and Zeyun Yu∗,1,2

1Department of Electrical Engineering, University of Wisconsin-Milwaukee, WI, USA2Department of Computer Science, University of Wisconsin-Milwaukee, WI, USA3Department of Mechanical Engineering, University of Wisconsin-Milwaukee, WI,

USA4Center for Computational and Biomedical Informatics, Marshfield Clinic Research

Institute, WI, USA

January 10, 2019

Abstract

With the rapid expansion of applied 3D computational vision, shapedescriptors have become increasingly important for a wide variety of appli-cations and objects from molecules to planets. Appropriate shape descrip-tors are critical for accurate (and efficient) shape retrieval and 3D modelclassification. Several spectral-based shape descriptors have been intro-duced by solving various physical equations over a 3D surface model. Inthis paper, for the first time, we incorporate a specific group of techniquesin statistics and machine learning, known as manifold learning, to developa global shape descriptor in the computer graphics domain. The pro-posed descriptor utilizes the Laplacian Eigenmap technique in which theLaplacian eigenvalue problem is discretized using an exponential weight-ing scheme. As a result, our descriptor eliminates the limitations tied tothe existing spectral descriptors, namely dependency on triangular meshrepresentation and high intra-class quality of 3D models. We also presenta straightforward normalization method to obtain a scale-invariant de-scriptor. The extensive experiments performed in this study show thatthe present contribution provides a highly discriminative and robust shapedescriptor under the presence of a high level of noise, random scale varia-tions, and low sampling rate, in addition to the known isometric-invarianceproperty of the Laplace-Beltrami operator. The proposed method signifi-cantly outperforms state-of-the-art algorithms on several non-rigid shaperetrieval benchmarks.

Keywords: Scale-invariant shape descriptor; Shape retrieval;Manifold Learning; Laplacian Eigenmap

∗Corresponding authors.

1

arX

iv:1

901.

0250

8v1

[cs

.GR

] 8

Jan

201

9

Page 2: An Application of Manifold Learning in Global Shape ...variants have been utilized in multiple FEM-based discretization methods. Another approach to constructing the matrix Ais to

1 Introduction

Three-dimensional models are ubiquitous data in the form of 3D surface meshes,point clouds, volumetric data, etc. in a wide variety of domains such as ma-terial and mechanical engineering [1], genetics [2], molecular biology [3], ento-mology [4], and dentistry [5, 6], to name a few. Processing such large datasets(e.g., shape retrieval, matching, or recognition) is computationally expensiveand memory intensive. For example, to query against a large database of 3Dmodels to find the closest match for a 3D model of interest, one needs to developan appropriate similarity measure as well as an efficient algorithm for search andretrieval. Shape descriptors assist with the example problem by providing dis-criminating feature vectors for shape retrieval [7, 8] and play a fundamental rolewhen dealing with shape analysis problems such as shape matching [9, 10] andclassification [11].

In general, there are two types of shape descriptors: local descriptors, alsocalled point signatures, and global descriptors, referred to as shape fingerprints.A local shape descriptor computes a feature vector for every point of a 3Dmodel. On the other hand, a global shape descriptor represents the whole3D shape model in the form of a low-dimension vector. A descriptor that isinformative and concise captures as much information as possible from the 3Dshape including the geometric and topological features. Such a vector drasticallylowers the shape analysis burdens in terms of both computational intensity andmemory.

While a large number of successful non-spectral shape descriptors have beenproposed in the literature, spectral descriptors have proved to be beneficial inmany applications [12, 13]. The spectral methods take advantage of eigen-decomposition of the Laplace-Beltrami (LB) operator applied on the shapesand construct their informative descriptors using the eigenvalues and eigenvec-tors. These methods have found successful applications in graph processing [14],computational biology [15], and point-to-point correspondence [16].

One of the first spectral descriptors introduced to the computer graphicscommunity is Shape-DNA, developed by Reuter et al. in 2006 [17]. Shape-DNAattracted a great deal of attention for its unique isometric and rotation invariantfeatures [17]. Since then, several local as well as global shape descriptors havebeen introduced in accordance with Shape-DNA such as Heat Kernel Signature(HKS) [18], Wave Kernel Signature (WKS) [19], and Global Point Signature(GPS) [20]. The common ground between these methods is the discretizationapproach used to solve the Laplacian eigenvalue problem, which uses a cotangentweighting scheme along with area normalization.

Although there are many advantages of using variations of the cotangentscheme, there are several limitations. First, by their nature, they are limitedto triangulated meshes. Second, they do not perform well when dealing withdegenerate and non-uniform sampled meshes [21, 22]. Also, their convergenceerror depends on factors such as the linearity of the function on the surface [22].One possible approach to address these limitations is through the use of manifoldlearning.

2

Page 3: An Application of Manifold Learning in Global Shape ...variants have been utilized in multiple FEM-based discretization methods. Another approach to constructing the matrix Ais to

Nonlinear dimensionality reduction techniques, known as manifold learn-ing, assume the existence of a low-dimensional space, which represents a high-dimensional manifold without much loss of information [23]. Similar to globaldescriptors, manifold learning methods attempt to learn the geometry of a man-ifold in order to extract a low dimensional vector of features that is informativeand discriminative. However, unlike shape descriptors, the number of dimen-sions of a space does not confine manifold learning methods. To the best ofour knowledge, the application of manifold learning, an active research topicin statistics and machine learning, has not been investigated in the computergraphics community for extracting global shape descriptors. This motivates theprimary aim of this research, which is to explore the effectiveness of a manifoldlearning method, more specifically Laplacian Eigenmap [24], in representing a3D model with a low-dimensional vector. Our work introduces a novel Lapla-cian Eigenmap-based global shape descriptor and provides a straightforwardnormalization method that significantly outperforms existing state-of-the-artapproaches.

In our first contribution, inspired by the idea of Laplacian Eigenmaps [24],we learn the manifold of a 3D model and then, analogous to the approach takenby Shape-DNA, we make use of the spectrum of the embedded manifold to buildthe global shape descriptor. This approach has two main advantages. First, itrelies on the adjacency of the nodes, disregarding the fine details of the meshstructure. Therefore, it can be used for degenerate or non-uniform sampledmeshes. Second, as manifold learning does not rely on the mesh structure andis not limited to a specific type of meshes, e.g., triangulated meshes, it can beapplied easily to any other mesh types such as quadrilateral meshes.

In our second contribution, we present a simple and straightforward nor-malization technique (motivated by [17, 25, 26]) to obtain a scale-invariantglobal shape descriptor that is more robust to noise. To this end, we pro-pose to subtract the first non-zero eigenvalue from the shape descriptor aftertaking the logarithm of the spectrum. One advantage of our approach over theidea of Bronstein et al. [25] is that we avoid taking the direct derivative; thisadvantage is significant since the differential operator amplifies the noise. Tak-ing the logarithm additionally helps to suppress the effect of the noise that ispresent in higher order elements of the spectrum.

The remainder of this paper is organized as follows. In Section 2, we brieflyoverview spectral shape analysis and manifold learning. Then in Section 3, weintroduce the proposed shape descriptor along with some technical background.In Section 4, the performance of the proposed method, as well as the robustnessof the algorithm are examined and compared with multiple well-known shapedescriptors by performing several qualitative and quantitative experiments usingwidely used 3D model datasets. Section 5 discusses the results in more detailand draws conclusions.

3

Page 4: An Application of Manifold Learning in Global Shape ...variants have been utilized in multiple FEM-based discretization methods. Another approach to constructing the matrix Ais to

2 Background

In this section, we first review spectral shape analysis, more specifically globalshape fingerprints, and different discretization methods of the LB operator.Then, we briefly review manifold learning, more specifically Laplacian Eigen-map, to provide the necessary foundation for developing our proposed LaplacianEigenmap based scale-invariant shape descriptor, which from now on we callLESI.

2.1 Spectral Shape Analysis

The Laplace-Beltrami operator ∆ is a linear differential operator defined on thedifferentiable manifold M as the divergence of the gradient of a function f asthe following form [17, 27]:

∆f = div(grad(f)). (1)

Levy [28] noted that the eigenfunctions φi of the continuous LB operator,which are the solution to the following Laplacian eigenvalue problem:

∆f = −λf (2)

are the orthogonal basis for the space of functions defined on the surface of amanifold. In other words, a function f on the surface can be expressed as a sumover coefficients of these infinite bases:

f = c0φ0 + c1φ1 + ...

Furthermore, the LB operator is positive semi-definite, having non-negativeeigenvalues λi that can be sorted as follows:

0 ≤ λ1 ≤ λ2 ≤ ... ≤ λi ≤ ...

The sequence of eigenvalues of the LB operator is called the spectrum of the LBoperator. As it is computed based on the gradient and divergence that dependon the Riemannian structure of the manifold, it possesses the isometry invariantproperty [17].

These significant features of the LB operator, which include the orthogonalbasis and non-negative spectrum, motivated researchers to develop various localand global shape descriptors. The Shape-DNA and HKS were developed byconsidering the heat distribution as the function f on the surface. The WKSwas obtained by solving the Schrodinger wave equation on the surface of themanifold. Also, it has been shown that the GPS descriptor is in close relationto the Green’s function on the surface [20].

To approximate equation (2), despite the choice of function f , we needa discretization scheme. Different discretization schemas (e.g., Taubin [29],Mayer [30]) of the LB operator on the triangular meshes are discussed in [31].

4

Page 5: An Application of Manifold Learning in Global Shape ...variants have been utilized in multiple FEM-based discretization methods. Another approach to constructing the matrix Ais to

A 3D shape, sampled from the surface of a Riemannian manifold M, isusually presented by a set of vertices V and their connectivity E in the form ofthe graph G = (V,E). For a surface mesh G, according to [21], the equation (2)can be discretized as:

Af = λBf (3)

where A is the stiffness matrix and B is the lumped mass matrix. One popularapproach to constructing the matrix A, is using weights:

wij =cot(αij) + cot(βij)

2(4)

where αij and βij are the two angles facing the edge (i, j). Different massnormalization methods using the triangle area or the Voronoi region area aresuggested to construct the matrix B. The cotangent weighting schema and itsvariants have been utilized in multiple FEM-based discretization methods.

Another approach to constructing the matrix A is to use the heat kernelweight, also known as the exponential weighting scheme, as follows:

wij = e−‖xi−xj‖

2

t (5)

where ‖.‖ denotes the Euclidean distance between two adjacent nodes i and j.In [21], several existing discretization methods, including variants of linear

FEM [32, 33] and heat kernel weighting proposed in [22] are compared. Accord-ing to [21] and from the discussion led by Xu in [31, 34], discrete LB operatorusing cotangent weighting scheme may not converge in all situations, specificallywhen dealing with non-uniform meshes. However, the heat kernel weightingscheme proposed in [22] does not depend on the peculiarities of the triangula-tion and outperforms all linear approaches [18, 21]. In addition, concerning thetype of the function f , the cotangent scheme only converges for linear functions,while the heat kernel scheme converges well for nonlinear functions as well [22].The proposed exponential approximation scheme provides point-wise conver-gence with good stability with respect to noise. It is important to note thatalthough the method was discussed for surfaces without boundary, the resultsare valid for interior points of a surface with boundary [22].

2.2 Manifold Learning

To make the current contribution self-contained, we provide a brief introduc-tion from the data analysis perspective. Dimensionality reduction of high-dimensional data is a critical step in data analysis and processing. Non-linear di-mensionality reduction, also known as manifold learning, is a problem of findinga low-dimensional representation for high-dimensional data. Several local andglobal manifold learning methods have been developed including Isomap [35, 36],LLE [37, 38], Laplacian Eigenmap [24], and Diffusion maps [39]. We review thesenext.

Consider a set x1, ..., xn ∈ M of n points on manifold embedded in Rl.Manifold learning methods look for a set of corresponding points y1, ..., yn in

5

Page 6: An Application of Manifold Learning in Global Shape ...variants have been utilized in multiple FEM-based discretization methods. Another approach to constructing the matrix Ais to

Rm(m� l) as a structural representation, while respecting some local or globalinformation. Each method attempts to minimize a cost function in this mapping.

Laplacian Eigenmap [40], proposed by Belkin and Niyogy in 2002, is a com-putationally efficient and mathematically well-studied manifold learning tech-nique. It is based upon graph Laplacian and Laplace-Beltrami operator on themanifold. Accordingly, it is considered as a spectral analysis method. LaplacianEigenmap deals with sparse, symmetric, and positive semi-definite matrices. Itis in close connection to the heat flow [24, 40].

Briefly speaking, for a given manifold, Laplacian Eigenmap applies the graphLaplacian operator and uses the eigenfunctions of such operator to provide theoptimal embedding. Laplacian Eigenmap preserves local information by mini-mizing the distance between embedded points, which are mapped from adjacentpoints in the original high-dimensional space [24]. Aside from the locality pre-serving property, it provides structural equivalence and discrimination by cap-turing the intrinsic geometric structure of the manifold. The structural equiva-lence property states that two similar manifolds will have similar representationafter projecting into a lower dimension space [41, 42].

Some other manifold learning methods, e.g., Isomap, LLE, and Diffusionmap are also based on spectral analysis of the high-dimensional manifold. Incontrast to these methods that construct the orthogonal basis of their desiredlow-dimensional space using eigenfunctions of an LB operator, we develop ourscale-invariant shape descriptor using the spectrum of the LB operator. Eventhough our primary focus is on the Laplacian Eigenmap, owing to its uniqueproperties, we believe that other spectral manifold learning methods are alsocapable of extracting informative and discriminative shape fingerprints.

3 Method

In this section, we elaborate our proposed LESI global shape descriptor. Aflowchart of the proposed approach is shown in Figure 1.

Construct:Laplacian matrix (L)

Solve:Lf = λDf

Spectrum:0 < λ1 ≤ ... ≤ λm

Normalization:log(λi) − log(λ1)

1 ≤ i ≤ m

3Dmodel

Figure 1: The block diagram of the proposed Laplacian Eigenmaps based scale-invariant (LESI) global shape descriptor.

6

Page 7: An Application of Manifold Learning in Global Shape ...variants have been utilized in multiple FEM-based discretization methods. Another approach to constructing the matrix Ais to

3.1 Laplacian Eigenmap-Based Shape Descriptor

We treat a global descriptor as a dimensionality reduction problem as it squeezesthe latent information of a 3D model into a vector. Due to the fact that theLaplacian Eigenmap has two properties of structural equivalence and localitypreservation [24], we propose a global shape descriptor using the spectrum ofgraph Laplacian.

A graph Laplacian is constructed over an undirected weighted graph G =(V,E) with a set of points xi ∈ V and a set E of edges that connects nearbypoints ((i, j) ∈ E). The theory behind finding the optimal embedding in aLaplacian Eigenmap requires an undirected graph. Every 3D model is givenin bidirectional connection and hence, we need to neither examine nor force itto the graph. However, as we will explain later, we need to remove isolatedpoints. Considering the advantages of the heat kernel weighting scheme, whichare summarized in Section 2.1 and discussed in details in [22], Laplacian Eigen-map suggests constructing the weighted graph as follows:

wij =

{e−‖xi−xj‖

2

t , if (i, j) ∈ E0 , otherwise

(6)

The only parameter in equation (6) is t, which defines the extent to whichdistant neighbors influence the embedding of each point. The choice of parame-ter t is data-dependent, and there exists no unique way in the literature to selectthe proper value, but it can be tuned empirically. As t has neither a very highimpact on the final embedding nor the convergence rate of our final derivations,we empirically recommend:

t = 2d2max where,

dmax = max ‖xi − xj‖ ,∀(i, j) ∈ E.(7)

Here, weights are bounded as 0.6 ≈ e−0.5 ≤ wij ≤ 1.Laplacian Eigenmap attempts to find a low dimensional data set that pre-

serves local information. For this purpose, it assumes two neighboring points xiand xj stay close after being mapped to yi and yj . Therefore, it minimizes thefollowing function [24]:

1

2

∑ij

(yi − yj)2Wij = yTLy. (8)

where L = D−W is the so called Laplacian matrix, Wij is a symmetric weightmatrix, and Dii =

∑jWij , the degree matrix, is a diagonal matrix. The as-

sumption that the graph is undirected yields the symmetric property of W andconsequently, D and L. It plays a critical role in deriving equation (8).

By adding the orthogonality constraint yTD1 = 0 in order to eliminate thetrivial solution and the constraint yTDy = 1 for removing an arbitrary scaling

7

Page 8: An Application of Manifold Learning in Global Shape ...variants have been utilized in multiple FEM-based discretization methods. Another approach to constructing the matrix Ais to

factor in the embedding, the minimization problem (8) simplifies to:

arg miny

yTDy=1

yTD1=0

yTLy. (9)

The matrix L is real, symmetric, and positive semi-definite. Therefore, thesolution vector y (in equation (9)) is obtained by the minimum eigenvalue so-lution to the generalized eigenvalue problem [24]:

Ly = λDy. (10)

At this point, the optimal low dimensional embedding, suggested by theLaplacian Eigenmap, is obtained by utilizing the eigenvectors. However, wefocus on the spectrum of the graph Laplacian and its’ properties. Eigenvaluesobtained from equation (10) are real, non-negative, and sorted in increasingorder as follows:

0 ≤ λ1 ≤ λ2 ≤ ... ≤ λn.As the row (or column) sum of L is equal to zero, eigenvalue λ = 0 and a cor-

responding eigenvector 1 are trivial solutions to equation (10). The multiplicityof eigenvalue zero is associated with the number of connected components of thegraph. Eigen-solvers often obtain very small, though not precisely zero, eigen-values due to the computational approximations. If we may know the numberof connected components of L, we can discard all eigenvalues equal to zero, andform our shape fingerprint using the more informative section of the spectrum.This is easily done by Dulmage-Mendelsohn decomposition [43].

The second smallest eigenvalue, also known as the Fiedler value, is a measureof the connectivity within the graph. If the graph has c connected components,our proposed shape descriptor is a set of d eigenvalues as:

LESI := (λc+1, λc+2, . . . , λc+d) (11)

The LESI descriptor is composed of the spectrum of the LB operator, andhence, it is isometric invariant, independent of the shape location, and informa-tive. The latter, discussed in spectral graph theory, states that the spectrum ofgraph Laplacian contains a considerable amount of geometrical and topologicalinformation of the graph. Moreover, LESI has the similarity property, causedby the structural equivalence property of Laplacian Eigenmap, meaning thattwo 3D models from the same class of models have similar fingerprints. Un-like Shape-DNA and other shape descriptors that are based on the cotangentweighting scheme, LESI is not limited to triangulated mesh structures becausethe Laplacian Eigenmap is capable of dealing with high-dimensional data. Forsome applications in which scale is not a determinant factor, it is favorable tohave a scale-invariant descriptor. A fast and efficient normalization method ispresented in Section 3.2.

One important matter to consider is the convergence and accuracy of theproposed fingerprint, which ultimately depends on the heat kernel-based dis-cretization of the LB operator. The cotangent weighting scheme and its variants

8

Page 9: An Application of Manifold Learning in Global Shape ...variants have been utilized in multiple FEM-based discretization methods. Another approach to constructing the matrix Ais to

(a) Teddy Bearmodels

(b) Spectrum ofLB operator

(c) Logarithm ofspectrum

(d) NormalizedSpectrum

Figure 2: An example showing the proposed normalization method of the shapedescriptor. (a) The Teddy Bear model and its scaled version (scale factor 0.7).(b) The spectrum of original (blue) and scaled (red) Teddy Bear models. Pleasenote that the original spectrum is approximately multiplied by half. (c) TheLogarithm of the spectrum shown in (b). (d) The normalized spectrum of orig-inal and scaled Teddy Bear models after subtracting first element of logarithmof the spectrum .

are sensitive to the peculiarities and quality of the particular triangulation of themesh (refer to Section 2.1). While an exponential weighting scheme has shownaccurate performance in dealing with nonlinear functions over the surface andnon-uniform mesh representations, it is not clear how this method can handlemanifolds with boundaries [21, 22]. It does, however, behave well for interiorpoints of the surface. Therefore, we recommend removing rows and columns ofL and D corresponding to isolated points, before solving equation (10). Thedescriptor obtained from the rest of the connected graph is an informative anddiscriminative descriptor of the graph.

3.2 Scale Normalization

For some applications, the size of an object is not a determinant factor in shapecomparison and identification. Therefore, a scale-invariant shape descriptorwith a solid normalization method is more desirable. For that purpose, someshape descriptors including Shape-DNA, have proposed multiple normalizationmethods. Most normalization methods of Shape-DNA focus on finding an ap-propriate scaling factor, such as the surface area, the volume, or coefficient of afitting curve, which will be multiplied in the descriptor.

Moreover, it is shown that eigenvalues with a higher order are more suscep-tible to noise. For that reason, the original Shape-DNA recommends croppingthe spectrum and using no more than 100 eigenvalues [17]. In this section, wepropose a simple and efficient normalization method that significantly reducesthe effect of scale variations as well as noise. In this approach, we are interestedin taking the scaling factor out from the descriptor in one step, rather than anextra step to find an appropriate neutralizing factor. Although the normaliza-tion seems simple, later in the experiments section, we will show its efficiency.The work presented in [18] and [17] influenced this method.

9

Page 10: An Application of Manifold Learning in Global Shape ...variants have been utilized in multiple FEM-based discretization methods. Another approach to constructing the matrix Ais to

According to Weyl’s law [44], when an arbitrary scale object is scaled withfactor α, the eigenvalues are scaled with factor α2. Let Λs = (α2λ1, . . . , α

2λd)be a LESI descriptor obtained from a scaled 3D model with unknown factor α.To normalize the descriptor and eliminate the effect of the scaling factor, werecommend computing:

Λn(i) = log(Λs(i))− log(Λs(1)) , for 1 ≤ i ≤ d (12)

To achieve a scale invariant shape descriptor, we first take the logarithm ofthe descriptor vector and then compute the difference of the new vector from itssmallest element. This is equivalent to dividing the vector by its first elementand taking the logarithm next. Basically, division takes the factor out, and thelogarithm eliminates the influence of noise.

Figure 2 illustrates the details of the proposed algorithm in an example. InFigure 2(a), two Teddy Bear models are shown. One model is in the originalsize whereas the other model is scaled with factor 0.7. It is clear from Figure2(b) that the spectrum of the scaled model is almost (0.7)2 of the spectrum ofthe original model. Taking the logarithm of the spectrum takes away the scalingfactor from multiplicand and leaves it as augend. Therefore, subtracting oneterm (e.g., the first element) removes the scaling factor from all other terms.The result is a normalized and scale-free spectrum.

3.3 Algorithm

Our proposed descriptor consists of three major steps. For a given 3D polygonalmodel G = (V,E) with a set of vertices V = (x1, . . . , xn) ∈ R3×n and a set ofneighbor connections E, the LESI descriptor is a d-dimensional vector Λ =(λ1, . . . , λd) of real and positive values.

In the first step, we compute the n × n real, symmetric, and sparse weightmatrix (W ) for a 1-ring neighbor of every point as stated in equation (6) usingthe inner scaling factor given in equation (7). Next, we form the generalizedeigenvalue problem equation (10) by constructing Laplacian and degree matri-ces (L and D respectively) without difficulty. The matrix L is sparse, real,symmetric and semi-positive. Utilizing the Dulmage-Mendelsohn decomposi-tion, we find the number of connected components of L. The objective of thesecond step is to find the spectrum of the LB operator. For that purpose, wesolve the generalized eigenvalue problem using the Lanczos method. Then, weleave out as many smallest eigenvalues as the number of connected components.Since in most cases a single 3D model is made up of one connected component,we only need to leave out one eigenvalue. The last step of the algorithm dealswith scale normalization and noise reduction, in case it is required, by takingthe logarithm of the spectrum and subtracting the first element from the restof the vector. Detailed steps of the algorithm are summarized in Algorithm 1.

10

Page 11: An Application of Manifold Learning in Global Shape ...variants have been utilized in multiple FEM-based discretization methods. Another approach to constructing the matrix Ais to

Algorithm 1: Laplacian Eigenmap-based scale-invariant global shape de-scriptor

Input: A 3D polygonal model G = (V,E) with n vertices xi and edgelist E

Output: A d-dimensional vector Λ = (λ1, . . . , λd)1 Compute edge weights using equations (6) and (7);2 Construct the sparse, real, and symmetric n× n matrices W , D, and L;3 Find number of connected components (nConComp) from L;4 Solve equation (10) for nConComp+ d eigenvalues;5 Sort them in increasing order and leave out nConComp smallest ones;6 if normalization is required then7 Compute log(λi)− log(λ1) where8 1 ≤ i ≤ d;

9 end

4 Experiments

In Section 4.1, we first present two datasets used in our experiments. Then,in Section 4.2, we qualitatively visualize and measure the competence of theproposed method in discriminating different clusters compared with candidatemethods from the literature. Next, in Section 4.3, we validate the effectivenessof the LESI descriptor to distinct multiple classes by measuring the accuracyof multi-class classification. Finally, in Section 4.4, extensive experiments arecarried out to study the robustness of the proposed shape descriptor with re-spect to noise, scale invariance, and down-sampling. All the algorithms wereimplemented using the MATLAB R2013a environment running on a personalcomputer with Intel(R) Core(TM) i3-4130 CPU @ 3.40GHz and 12GB memory.

4.1 Dataset

To validate the utility of the proposed shape descriptor, we utilized two stan-dard, widely-used, and publicly available datasets of 3D polygon meshes. Thehigh-resolution TOSCA dataset [45] contains 80 three-dimensional non-rigidmodels, including 11 cats, 6 centaurs, 9 dogs, 4 gorillas, 8 horses, 12 womenposes, 3 wolves and two men with 7 and 20 poses respectively. In our exper-iments, we use all models except the gorilla models, as they contain isolatedpoints. The models in each class of the TOSCA dataset are almost identical interms of scale, the number of vertices, quality of triangulation, and structure,which all represent the same object with different poses.

The McGill dataset with articulating parts [46] is used to evaluate the abilityof the descriptors to describe models with poor intra-class quality. The McGilldataset contains 3D models of 30 ants, 30 crabs, 25 glasses, 20 hands, 29 hu-mans, 25 octopuses, 20 pliers, 25 snakes, 31 spiders, and 20 Teddy bears. Theclassification of the McGill dataset models is more challenging due to scale and

11

Page 12: An Application of Manifold Learning in Global Shape ...variants have been utilized in multiple FEM-based discretization methods. Another approach to constructing the matrix Ais to

shape variations.

4.2 Retrieval Results

In this section, we evaluate the general performance of our proposed shape de-scriptor and compare it with several state-of-the-art spectral-based global shapedescriptors including Shape-DNA [17], cShape-DNA [27], and GPS [20] algo-rithms. We chose these methods because they are widely used by researchers(e.g., [47–49]) to develop new descriptors or applications, or to evaluate the per-formance of their proposed descriptors. Moreover, cShape-DNA represents thenormalized version of the original Shape-DNA. Even though there are multipleways to convert a local point descriptor to a global shape fingerprint, in this ar-ticle we focus only on algorithms that have been originally introduced as globalfingerprints. To this end, we take advantage of the source code made available onDr. Kokkinos’s homepage1 [25], as well as the shape descriptor package providedby Li et al. [50] available on a GitHub repository2 to generate the Shape-DNAand GPS descriptors, respectively. We also compare the performance of shaperetrieval using the code provided for evaluation by SHREC’11 [51].

The shape descriptors are compared using the TOSCA dataset to discrimi-nate between different classes of 3D objects. In this experiment, we use the first33 non-zero eigenvalues (d = 33). Then, to visualize the locations of objects inthe shape space, we project them onto a 2D plane using Principle ComponentAnalysis (PCA). Figure 3 displays the effectiveness of our method comparedwith the fingerprints of interest.

(a) Shape-DNA (b) cShape-DNA (c) GPS (d) LESI

Figure 3: 2D PCA projection of shape descriptors computed from (a) originalShape-DNA, (b) cShape-DNA, (c) GPS, and (d) LESI algorithms on TOSCAdataset.

Figure 3 reveals that LESI can differentiate models of various classes signifi-cantly better than the other methods for a refined and normalized dataset. Eventhough all human models (David, Michael, and Victoria) are very similar, it candistinguish the women from the men’s group. However, it fails to discriminatemodels of Michael from David. Despite the large isometric deformations in eachclass, the proposed LESI method clusters all models of the same class togethervery tightly.

1http://vision.mas.ecp.fr/Personnel/iasonas/descriptors.html2https://github.com/ChunyuanLI/spectral_descriptors

12

Page 13: An Application of Manifold Learning in Global Shape ...variants have been utilized in multiple FEM-based discretization methods. Another approach to constructing the matrix Ais to

Table 1: Shape retrieval performance using TOSCA and McGill datasets

Dataset Method NN FT ST E DCG

TOSCA

ShapeDNA 1.0000 0.8091 0.9391 0.4486 0.9584cShapeDNA 0.9474 0.7748 0.8984 0.4748 0.9241GPS 0.4868 0.4244 0.6320 0.3614 0.6787LESI 0.8684 0.8456 0.9430 0.4860 0.9244

McGill

ShapeDNA 0.7922 0.3452 0.4977 0.3411 0.7192cShapeDNA 0.7882 0.3943 0.5483 0.3852 0.7470GPS 0.3843 0.2508 0.4066 0.2588 0.6020LESI 0.9647 0.7046 0.8739 0.6644 0.9251

To demonstrate the power of our method in classifying objects with low intra-class similarity compared with other shape descriptors, the same experiment iscarried out on the McGill dataset. Models of the same class with articulatingparts are in different scales, shape, and structure. The 2D PCA projections of33-dimension descriptors from all four algorithms are shown in Figure 4.

(a) Shape-DNA (b) cShape-DNA (c) GPS (d) LESI

Figure 4: 2D PCA projection of shape descriptors computed from (a) originalShape-DNA, (b) cShape-DNA, (c) GPS, and (d) LESI algorithms on McGilldataset.

As illustrated in Figure 4, the original Shape-DNA is highly sensitive toscales. Multiple methods are presented in [17] to make the descriptor normalizedto scale. cShape-DNA represents a normalized version of it by multiplying thedescriptor with the surface area. Although cShape-DNA can separate modelsfrom each other, classes are not separated efficiently. LESI outperforms theother algorithms by providing distinct descriptors, which can separate classes.Shape descriptors offered by LESI prove superior to the other algorithms inthe shape retrieval and classification tasks, as described below and in the nextsection respectively.

To examine the superiority of LESI quantitatively, we computed multiplestandard retrieval measures including Nearest Neighbor (NN), First Tier (FT),Second Tier (ST), e-Measure (E), and Discounted Cumulative Gain (DCG).These measures represent state-of-the-art quality metrics used when evaluatingmatching results for shape-based search engines [52]. Table (1) reports the

13

Page 14: An Application of Manifold Learning in Global Shape ...variants have been utilized in multiple FEM-based discretization methods. Another approach to constructing the matrix Ais to

results of shape retrieval. Boldface numbers indicate the highest value for eachmeasure per each dataset. From Table (1), it is clear that the LESI descriptoroutperforms all other methods concerning all measures in retrieving modelsfrom the McGill dataset. When retrieving models of the TOSCA dataset, LESIoutperforms all methods concerning FT, ST, and E measures. Shape-DNAoutperforms LESI by a higher value for NN and DCG measures, due to the poordiscrimination between David and Michael performed by the LESI descriptor.However, it does not diminish the validity of our claim that LESI performs wellfor meshes with non-uniform sampling or peculiarities.

4.3 Multi-class Classification Results

In this section, we corroborate the findings of Section 4.2 by training a linearmulti-class SVM classifier to assess the accuracy of LESI compared to othershape descriptors. For this experiment, we utilized the McGill dataset. Inaddition to the shape descriptors evaluated in Section 4.2, we computed anothernormalized version of Shape-DNA by dividing the feature vector by its firstelement (similar to what LESI offers) as suggested in [17]. This way we cancompare the effect of the exponential weighting scheme without the influenceof the normalization method or compactness (offered by cShape-DNA). Using10-fold cross-validation and repeating the experiment 3 times, we report theaverage accuracy for each method in Table (2).

The new LESI approach significantly outperforms all other methods whenusing a two-tailed paired t-test (p < 0.05). The t-test was performed on oneset of 10 folds in order to avoid violating the independence assumption of thet-test. There is a significant improvement in accuracy when comparing theShape-DNA (Normalized) to other variants of the Shape-DNA, which is due inpart to the normalization method. However, the average accuracy of the LESIdescriptor is noticeably higher (95%) when compared to 90% of the Shape-DNA(Normalized).

Table 2: Classification accuracy using McGill dataset

Method Average accuracy

Shape-DNA 21.02%Shape-DNA (Normalized) 90.60%cShape-DNA 71.37%GPS 50.11%LESI 95.69%

Finally, Figure 5 shows the confusion matrix obtained from the linear multi-class SVM using LESI descriptor. The number of correct classifications made foreach class (indicated by the green diagonal), confirms that our method capturesthe discriminative features of the shapes.

14

Page 15: An Application of Manifold Learning in Global Shape ...variants have been utilized in multiple FEM-based discretization methods. Another approach to constructing the matrix Ais to

Figure 5: Confusion matrix obtained from linear multi-class SVM for McGilldataset using LESI descriptors.

4.4 Robustness

In this section, we address the robustness of the LESI shape descriptor to shapevariations, including noise, scale, and down-sampling by performing another setof experiments. First, we generate the disturbed version of every model in theTOSCA dataset. Then, we test the capability of every method mentioned abovein discriminating between different classes. For this purpose, besides plotting the2D PCA projection of shape descriptors, we also compute and plot the pairwiseEuclidean distance matrix, in every case. The distance matrix represents thedissimilarity between each pair of models in the set. It is often used to computeother evaluating metrics such as nearest-neighbor, and first and second tier, toname a few. The dissimilarity of descriptors increases from blue to red, andthe more separate classes differ in color, the better they are discriminating fromeach other.

Resistance to noise. Multiple noisy versions of the TOSCA dataset aregenerated following the idea articulated in [53]. To this end, the surface meshesof all models are disturbed by changing the position of each point along itsnormal vector that is chosen randomly from an interval (−L,L) with the 0mean, where L determines the noise level and is a fraction of the diagonal lengthof the model bounding box. In this experiment, three noise levels L = 0.5%,L = 1%, and L = 2% are tested, where the latter one represents a greater levelof noise. Two-dimensional PCA projections of all descriptors with the presenceof different levels of noise are plotted in Figure 6. Combining these with theresults shown in Figure 3, where no noise is present, demonstrates that theLESI algorithm is highly noise-resistant while the performance of the Shape-DNA and cShape-DNA decreases as the level of noise increases. Moreover, GPSfails in separating different classes of models with the presence of noise. Figure 9

15

Page 16: An Application of Manifold Learning in Global Shape ...variants have been utilized in multiple FEM-based discretization methods. Another approach to constructing the matrix Ais to

Figure 6: 2D PCA projection of shape descriptors computed by (from left toright) Shape-DNA, cShape-DNA, GPS, and LESI algorithms from perturbedTOSCA dataset with (from top to bottom) 0.5%, 1%, 2% noise level, respec-tively.

reflects the effect of noise on the discriminative power of the descriptors. TheLESI algorithm shows consistent results as the level of noise increases from 0%(top row) to 2% (bottom row).

Scale invariance. In order to validate the insensitivity of the LESI de-scriptor to scale variations and compare the robustness of the proposed methodwith other descriptors, each model of the TOSCA dataset is scaled by a factorof 0.5, 0.875, 1.25, 1.625, or 2 randomly. Figure 7 shows that the LESI algo-rithm surpasses other methods in discerning different classes. Comparing theresult of this experiment with the results shown in Figure 3 demonstrates theconsistency of the LESI algorithm with the presence of scale variation. Thedistance matrices in Figure 10 show that the original Shape-DNA algorithm isvery susceptible to scale variations. Even though the cShape-DNA has signifi-cantly improved scale sensitivity of the original Shape-DNA, it does not provideas accurate results as the LESI algorithm does.

Resistance to the sampling rate. To investigate the effect of samplingrates on the discriminative power of the shape descriptors, Bronstein et al. [8]propose to reduce the number of vertices to 20% of its original size. Accord-ingly, the down-sampled version of the TOSCA dataset is generated, and shapedescriptors associated with them are computed. The 2D PCA projections anddistance matrices of descriptors are illustrated in Figures 8 and 11, respectively.Although the original Shape-DNA shows a more accurate result than cShape-DNA, the separation of cat, dog, and wolf models is challenging. Although

16

Page 17: An Application of Manifold Learning in Global Shape ...variants have been utilized in multiple FEM-based discretization methods. Another approach to constructing the matrix Ais to

(a) Shape-DNA (b) cShape-DNA (c) GPS (d) LESI

Figure 7: 2D PCA projection of shape descriptors computed by (a) originalShape-DNA, (b) cShape-DNA, (c) GPS, and (d) LESI algorithms over scaledversion of the TOSCA dataset by a randomly chosen factor of 0.5, 0.875, 1.25,1.625, or 2.

(a) Shape-DNA (b) cShape-DNA (c) GPS (d) LESI

Figure 8: 2D PCA projection of shape descriptors computed by (a) originalShape-DNA, (b) cShape-DNA, (c) GPS, and (d) LESI algorithms from downsampled TOSCA dataset by rate of 20%.

the performance of the LESI method is slightly affected, it still outperformscShape-DNA and GPS methods.

5 Discussion

In this article, motivated by the unique properties of Laplacian Eigenmap (i.e.,locality preservation, structural equivalence, and dimensionality reduction) andinspired by the existing spectral-based shape descriptors, we investigated the ap-plication of manifold learning in deriving a shape fingerprint in order to addressthe limitations tied to popular cotangent-based shape descriptors. We proposeda global descriptor (LESI) with an easy-to-compute and efficient normalizationtechnique that facilitates applications such as shape classification and retrieval.Our method applies fewer restrictions on the class of meshes as well as improvingthe quality of tessellations. Analogous to other spectral descriptors, LESI usesthe spectrum of the LB operator, which is independent of the shape location,is informative (contains a considerable amount of geometrical and topologicalinformation), and above all isometric invariant. We compared the discriminat-ing power of LESI with three prominent descriptors from the literature, namelyShape-DNA, cShape-DNA, and GPS, and found it to be superior.

In the first set of experiments illustrated in Figures 3 and 4, our method sub-

17

Page 18: An Application of Manifold Learning in Global Shape ...variants have been utilized in multiple FEM-based discretization methods. Another approach to constructing the matrix Ais to

NoiseLevel

Shape-DNA cShape-DNA GPS LESI

0%

0.5%

1%

2%

Figure 9: The Euclidean pairwise distance matrix of shape descriptors computedby (from left to right) Shape-DNA, cShape-DNA, GPS, and LESI algorithmsfrom perturbed TOSCA dataset by (from top to bottom) 0%, 0.5%, 1%, 2%noise levels.

Shape-DNA cShape-DNA GPS LESI

RandomScale

Figure 10: The Euclidean pairwise distance matrix of shape descriptors com-puted by (from left to right) Shape-DNA, cShape-DNA, GPS, and LESI algo-rithms over scaled version of the TOSCA dataset by a randomly chosen factorof 0.5, 0.875, 1.25, 1.625, or 2.

stantially outperforms the others. The superiority of LESI is more significantwhen the McGill dataset is used (Table 2 and Figure5). This dataset includeswide variations in mesh structure and scales, causing the failure of the othermethods to generate acceptable results. However, LESI, due to utilizing a dif-ferent method of discretization to form the LB operator, focuses on the vicinityrather than the quality of the triangulation. Therefore, our technique, unlike

18

Page 19: An Application of Manifold Learning in Global Shape ...variants have been utilized in multiple FEM-based discretization methods. Another approach to constructing the matrix Ais to

Shape-DNA cShape-DNA GPS LESI

DownSampled

Figure 11: The Euclidean pairwise distance matrix of shape descriptors com-puted by (from left to right) Shape-DNA, cShape-DNA, GPS, and LESI algo-rithms from from down sampled TOSCA dataset by rate of 20%.

other methods, is not affected by the low quality of polygon meshes.The second set of experiments evaluates the reliability of our method in the

presence of noise, scale variations, as well as different sampling rates. LESIshows impressive robustness against the first two sets of perturbation. Despitethe negative impact of down sampling in LESI descriptor, it continues to showbetter performance when compared to cShape-DNA and GPS. It should benoted that the result could also be improved by increasing the size of the outputvector.

In addition to the discriminating power of the descriptor, degenerate andnon-uniform meshes may also cause failure of an algorithm to converge. Thecotangent weight-based algorithms were not able to compute the descriptors for2 shapes from the McGill dataset. GPS also failed to compute descriptors for 6models of the down sampled TOSCA dataset. However, our technique convergesat all times despite the quality of the polygon mesh structure.

Moreover, LESI, unlike cotangent weight-based techniques, is not confined tothe triangulated meshes as it disregards the mesh geometry [54]. LESI inheritsthis property from the capability of manifold learning techniques in coping withhigh dimensional data. The discretization of the LB operator using cotangentweights on the quadrilateral meshes is not as straightforward as on triangularmeshes. To compute the LB operator on a quadrilateral mesh, all rectanglesneed to be divided into triangles. It could be done easily, however, as for eachquad there are two possible triangulations, the result is not unique.

In the original Laplacian Eigenmaps, the high dimension data requires aconsiderable amount of processing as the list of all connections need to be com-puted for the dataset. In fact, for each point in the high dimension space, agiven number of nearest neighbors need to be extracted which could be chal-lenging and unmanageable. While applying this technique to the 3D meshes,we skip this step as the neighbors are already defined and given in the meshstructure.

This work benefits from the Laplacian Eigenmap technique in a space inwhich the vicinities are given. LESI takes advantage of simple Laplacian com-putation, to form the LB operator, which provides concise and informative shapedescriptors. Experimental results prove that LESI is more effective comparedwith the other powerful descriptors.

19

Page 20: An Application of Manifold Learning in Global Shape ...variants have been utilized in multiple FEM-based discretization methods. Another approach to constructing the matrix Ais to

One limitation of LESI is its inability to separate models of different men(David and Michael). However, it was able to differentiate between the wolf anddog, as well as between women and men.

Although we investigated only the application of Laplacian Eigenmap inintroducing a shape descriptor, there are some other spectral-based manifoldlearning methods, such as Isomap, LLE, and Diffusion map, which have notbeen examined. This can be considered future work.

6 Conclusion

This work presents LESI, a novel scale-invariant global shape descriptor basedon Laplacian Eigenmap that is significantly better when compared to othershape descriptors. We conclude that manifold learning methods can be used todevelop new spectral-based shape descriptors to learn the structure of manifoldsdespite the quality of sampled meshes.

7 Acknowledgment

We acknowledge financial support for Drs. D’Souza, Yu and Ms. Bashiri fromGE Healthcare through the UWM Catalyst Grant program. Our sincere thanksgoes to Dr. Ahmad P. Tafti and Dr. C. David Page for their expertise and con-structive comments. We also acknowledge financial support from the Center forPredictive Computational Phenotyping, supported by the National Institutes ofHealth Big Data to Knowledge (BD2K) Initiative under Award Number U54AI117924 and the grant UL1TR002373 from the Clinical and Translational Sci-ence Award (CTSA) program of the National Center for Advancing Transla-tional Sciences, NIH.

References

[1] Emad Omrani, Ahmad P Tafti, Mojtaba F Fathi, Afsaneh DorriMoghadam, Pradeep Rohatgi, Roshan M D’Souza, and Zeyun Yu. Tribo-logical study in microscale using 3d sem surface reconstruction. TribologyInternational, 103:309–315, 2016.

[2] Lydia Ng, Sayan Pathak, Chihchau Kuan, Chris Lau, Hong-wei Dong,Andrew Sodt, Chinh Dang, Brian Avants, Paul Yushkevich, James Gee,et al. Neuroinformatics for genome-wide 3-d gene expression mapping inthe mouse brain. IEEE/ACM Transactions on Computational Biology andBioinformatics (TCBB), 4(3):382–393, 2007.

[3] Zhanheng Gao, Reihaneh Rostami, Xiaoli Pang, Zhicheng Fu, and ZeyunYu. Mesh generation and flexible shape comparisons for bio-molecules.Molecular Based Mathematical Biology, 4(1):1–13, 2016.

20

Page 21: An Application of Manifold Learning in Global Shape ...variants have been utilized in multiple FEM-based discretization methods. Another approach to constructing the matrix Ais to

[4] G. D. Sosa, S. Rodrıguez, J. Guaje, J. Victorino, M. Mejıa, L. S. Fuentes,A. Ramırez, and H. Franco. 3d surface reconstruction of entomologicalspecimens from uniform multi-view image datasets. In 2016 XXI Sympo-sium on Signal Processing, Images and Artificial Vision (STSIVA), pages1–8, Aug 2016. doi: 10.1109/STSIVA.2016.7743319.

[5] Stefan Riehemann, Martin Palme, Peter Kuehmstedt, Constanze Gross-mann, Gunther Notni, and Josef Hintersehr. Microdisplay-based intraoral3d scanner for dentistry. Journal of Display Technology, 7(3):151–155, 2011.

[6] Chenglei Wu, Derek Bradley, Pablo Garrido, Michael Zollhofer, ChristianTheobalt, Markus Gross, and Thabo Beeler. Model-based teeth reconstruc-tion. ACM Transactions on Graphics (TOG), 35(6):220, 2016.

[7] Yonathan Aflalo, Alexander M Bronstein, Michael M Bronstein, and RonKimmel. Deformable shape retrieval by learning diffusion kernels. In Inter-national Conference on Scale Space and Variational Methods in ComputerVision, pages 689–700. Springer, 2011.

[8] Alexander M Bronstein, Michael M Bronstein, Leonidas J Guibas, andMaks Ovsjanikov. Shape google: Geometric words and expressions forinvariant shape retrieval. ACM Transactions on Graphics (TOG), 30(1):1,2011.

[9] Jin Xie, Yi Fang, Fan Zhu, and Edward Wong. Deepshape: Deep learnedshape descriptor for 3d shape matching and retrieval. In Proceedings ofthe IEEE Conference on Computer Vision and Pattern Recognition, pages1275–1283, 2015.

[10] R Toldo, U Castellani, and A Fusiello. Visual vocabulary signature for3d object retrieval and partial matching. In Proceedings of the 2nd Eu-rographics conference on 3D Object Retrieval, pages 21–28. EurographicsAssociation, 2009.

[11] Shuhui Bu, Zhenbao Liu, Junwei Han, Jun Wu, and Rongrong Ji. Learn-ing high-level feature by deep belief networks for 3-d model retrieval andrecognition. IEEE Transactions on Multimedia, 16(8):2154–2167, 2014.

[12] Davide Boscaini, Jonathan Masci, Emanuele Rodola, Michael M Bronstein,and Daniel Cremers. Anisotropic diffusion descriptors. Computer GraphicsForum, 35(2):431–441, 2016.

[13] Alexander M Bronstein. Spectral descriptors for deformable shapes. arXivpreprint arXiv:1110.5015, 2011.

[14] Dan Raviv, Ron Kimmel, and Alfred M Bruckstein. Graph isomorphismsand automorphisms via spectral signatures. IEEE transactions on patternanalysis and machine intelligence, 35(8):1985–1993, 2013.

21

Page 22: An Application of Manifold Learning in Global Shape ...variants have been utilized in multiple FEM-based discretization methods. Another approach to constructing the matrix Ais to

[15] Dela De Youngster, Eric Paquet, Herna Viktor, and Emil Petriu. Anisometry-invariant spectral approach for protein-protein docking. In Bioin-formatics and Bioengineering (BIBE), 2013 IEEE 13th International Con-ference on, pages 1–6. IEEE, 2013.

[16] Maks Ovsjanikov, Mirela Ben-Chen, Justin Solomon, Adrian Butscher, andLeonidas Guibas. Functional maps: a flexible representation of maps be-tween shapes. ACM Transactions on Graphics (TOG), 31(4):30, 2012.

[17] Martin Reuter, Franz-Erich Wolter, and Niklas Peinecke. Laplace–beltramispectra as ’shape-dna’ of surfaces and solids. Computer-Aided Design, 38(4):342–366, 2006.

[18] Jian Sun, Maks Ovsjanikov, and Leonidas Guibas. A concise and prov-ably informative multi-scale signature based on heat diffusion. Computergraphics forum, 28(5):1383–1392, 2009.

[19] Mathieu Aubry, Ulrich Schlickewei, and Daniel Cremers. The wave kernelsignature: A quantum mechanical approach to shape analysis. In ComputerVision Workshops (ICCV Workshops), 2011 IEEE International Confer-ence on, pages 1626–1633. IEEE, 2011.

[20] Raif M Rustamov. Laplace-beltrami eigenfunctions for deformation invari-ant shape representation. In Proceedings of the fifth Eurographics sympo-sium on Geometry processing, pages 225–233. Eurographics Association,2007.

[21] Martin Reuter, Silvia Biasotti, Daniela Giorgi, Giuseppe Patane, andMichela Spagnuolo. Discrete laplace–beltrami operators for shape analysisand segmentation. Computers & Graphics, 33(3):381–390, 2009.

[22] Mikhail Belkin, Jian Sun, and Yusu Wang. Discrete laplace operator onmeshed surfaces. In Proceedings of the twenty-fourth annual symposium onComputational geometry, pages 278–287. ACM, 2008.

[23] Yair Goldberg, Alon Zakai, Dan Kushnir, and Ya’acov Ritov. Manifoldlearning: The price of normalization. Journal of Machine Learning Re-search, 9(Aug):1909–1939, 2008.

[24] Mikhail Belkin and Partha Niyogi. Laplacian eigenmaps for dimensionalityreduction and data representation. Neural computation, 15(6):1373–1396,2003.

[25] Michael M Bronstein and Iasonas Kokkinos. Scale-invariant heat kernelsignatures for non-rigid shape recognition. In Computer Vision and PatternRecognition (CVPR), 2010 IEEE Conference on, pages 1704–1711. IEEE,2010.

22

Page 23: An Application of Manifold Learning in Global Shape ...variants have been utilized in multiple FEM-based discretization methods. Another approach to constructing the matrix Ais to

[26] Zhenzhong Kuang, Zongmin Li, Qian Lv, Tian Weiwei, and Yujie Liu.Modal function transformation for isometric 3d shape representation. Com-puters & Graphics, 46:209–220, 2015.

[27] Zhanheng Gao, Zeyun Yu, and Xiaoli Pang. A compact shape descriptorfor triangular surface meshes. Computer-Aided Design, 53:62–69, 2014.

[28] Bruno Levy. Laplace-beltrami eigenfunctions towards an algorithm that”understands” geometry. In Shape Modeling and Applications, 2006. SMI2006. IEEE International Conference on, pages 13–13. IEEE, 2006.

[29] Gabriel Taubin. A signal processing approach to fair surface design. InProceedings of the 22nd annual conference on Computer graphics and in-teractive techniques, pages 351–358. ACM, 1995.

[30] Uwe F Mayer. Numerical solutions for the surface diffusion flow in threespace dimensions. Computational and Applied Mathematics, 20(3):361–379,2001.

[31] Guoliang Xu. Discrete laplace–beltrami operators and their convergence.Computer aided geometric design, 21(8):767–784, 2004.

[32] Mathieu Desbrun, Mark Meyer, Peter Schroder, and Alan H Barr. Im-plicit fairing of irregular meshes using diffusion and curvature flow. InProceedings of the 26th annual conference on Computer graphics and inter-active techniques, pages 317–324. ACM Press/Addison-Wesley PublishingCo., 1999.

[33] Mark Meyer, Mathieu Desbrun, Peter Schroder, and Alan H Barr. Discretedifferential-geometry operators for triangulated 2-manifolds. In Visualiza-tion and mathematics III, pages 35–57. Springer, 2003.

[34] Guoliang Xu. Convergence analysis of a discretization scheme for gaussiancurvature over triangular surfaces. Computer Aided Geometric Design, 23(2):193–207, 2006.

[35] Joshua B Tenenbaum, Vin De Silva, and John C Langford. A global ge-ometric framework for nonlinear dimensionality reduction. science, 290(5500):2319–2323, 2000.

[36] Vin D Silva and Joshua B Tenenbaum. Global versus local methods innonlinear dimensionality reduction. In Advances in neural information pro-cessing systems, pages 721–728, 2003.

[37] Sam T Roweis and Lawrence K Saul. Nonlinear dimensionality reductionby locally linear embedding. science, 290(5500):2323–2326, 2000.

[38] Lawrence K Saul and Sam T Roweis. Think globally, fit locally: unsuper-vised learning of low dimensional manifolds. Journal of machine learningresearch, 4(Jun):119–155, 2003.

23

Page 24: An Application of Manifold Learning in Global Shape ...variants have been utilized in multiple FEM-based discretization methods. Another approach to constructing the matrix Ais to

[39] Ronald R Coifman and Stephane Lafon. Diffusion maps. Applied andcomputational harmonic analysis, 21(1):5–30, 2006.

[40] Mikhail Belkin and Partha Niyogi. Laplacian eigenmaps and spectral tech-niques for embedding and clustering. In Advances in neural informationprocessing systems (NIPS), pages 585–591, 2002.

[41] Christian Wachinger and Nassir Navab. Manifold learning for multi-modalimage registration. In BMVC, pages 1–12. Citeseer, 2010.

[42] Christian Wachinger and Nassir Navab. Structural image representationfor image registration. In Computer Vision and Pattern Recognition Work-shops (CVPRW), 2010 IEEE Computer Society Conference on, pages 23–30. IEEE, 2010.

[43] Andrew L Dulmage and Nathan S Mendelsohn. Coverings of bipartitegraphs. Canadian Journal of Mathematics, 10(4):516–534, 1958.

[44] Hermann Weyl. Uber die asymptotische verteilung der eigenwerte.Nachrichten von der Gesellschaft der Wissenschaften zu Gottingen,Mathematisch-Physikalische Klasse, 1911:110–117, 1911.

[45] Alexander M Bronstein, Michael M Bronstein, and Ron Kimmel. Numericalgeometry of non-rigid shapes. Springer Science & Business Media, 2008.

[46] Kaleem Siddiqi, Juan Zhang, Diego Macrini, Ali Shokoufandeh, SylvainBouix, and Sven Dickinson. Retrieving articulated 3-d models using medialsurfaces. Machine vision and applications, 19(4):261–275, 2008.

[47] Mahsa Mirloo and Hossein Ebrahimnezhad. Non-rigid 3d object re-trieval using directional graph representation of wave kernel signa-ture. Multimedia Tools and Applications, Apr 2017. ISSN 1573-7721. doi: 10.1007/s11042-017-4617-x. URL https://doi.org/10.1007/

s11042-017-4617-x.

[48] Majid Masoumi and A Ben Hamza. Global spectral graph wavelet signaturefor surface analysis of carpal bones. arXiv preprint arXiv:1709.02782, 2017.

[49] Davide Boscaini, Jonathan Masci, Simone Melzi, Michael M Bronstein,Umberto Castellani, and Pierre Vandergheynst. Learning class-specific de-scriptors for deformable shapes using localized spectral convolutional net-works. Computer Graphics Forum, 34(5):13–23, 2015.

[50] Chunyuan Li and A Ben Hamza. Spatially aggregating spectral descriptorsfor nonrigid 3d shape retrieval: a comparative survey. Multimedia Systems,20(3):253–281, 2014.

[51] Z Lian, A Godil, B Bustos, M Daoudi, J Hermans, S Kawamura, Y Kurita,G Lavoua, and P Dp Suetens. Shrec’11 track: Shape retrieval on non-rigid3d watertight meshes. In Eurographics Workshop on 3D Object Retrieval(3DOR), 2011.

24

Page 25: An Application of Manifold Learning in Global Shape ...variants have been utilized in multiple FEM-based discretization methods. Another approach to constructing the matrix Ais to

[52] Philip Shilane, Patrick Min, Michael Kazhdan, and Thomas Funkhouser.The princeton shape benchmark. In Shape modeling applications, 2004.Proceedings, pages 167–178. IEEE, 2004.

[53] Yong-Jin Liu, Zhanqing Chen, and Kai Tang. Construction of iso-contours,bisectors, and voronoi diagrams on triangulated surfaces. IEEE Transac-tions on Pattern Analysis and Machine Intelligence, 33(8):1502–1517, 2011.

[54] Yongjie Zhang, Chandrajit Bajaj, and Guoliang Xu. Surface smoothingand quality improvement of quadrilateral/hexahedral meshes with geomet-ric flow. International Journal for Numerical Methods in Biomedical Engi-neering, 25(1):1–18, 2009.

25