15
IEEE TRANSACTIONS ON SIGNAL PROCESSING 1 Robust L 2 E Estimation of Transformation for Non-Rigid Registration Jiayi Ma, Weichao Qiu, Ji Zhao, Yong Ma, Alan L. Yuille, and Zhuowen Tu Abstract—We introduce a new transformation estimation al- gorithm using the L2E estimator, and apply it to non-rigid reg- istration for building robust sparse and dense correspondences. In the sparse point case, our method iteratively recovers the point correspondence and estimates the transformation between two point sets. Feature descriptors such as shape context are used to establish rough correspondence. We then estimate the transformation using our robust algorithm. This enables us to deal with the noise and outliers which arise in the correspondence step. The transformation is specified in a functional space, more specifically a reproducing kernel Hilbert space. In the dense point case for non-rigid image registration, our approach consists of matching both sparsely and densely sampled SIFT features, and it has particular advantages in handling significant scale changes and rotations. The experimental results show that our approach greatly outperforms state-of-the-art methods, particularly when the data contains severe outliers. Index TermsL2E estimator, registration, dense correspon- dence, regularization, outlier, non-rigid. I. I NTRODUCTION Building “correct” registration, alignment or correspon- dence is a fundamental problem in computer vision, medical image analysis, and pattern recognition [1], [2], [3], [4]. Many tasks in these fields – such as stereo matching, shape matching, image registration and content-based image retrieval – can be formulated as matching two sets of feature points, more specifically, as a point matching problem [1]. The feature points in these tasks are typically the locations of sparse or dense interest points extracted from an image, or the edge points sampled from a shape contour. The registration problem then reduces to determining the correct correspondence, and to find the underlying spatial transformation between two point sets extracted from the input data. The registration problem can be categorized into sparse correspondence or dense correspondence, depending on the specific application. The task of sparse correspondence aims to match sparsely distributed points, such as matching contour points in shape matching. It also frequently arises in image matching, for example, for images taken from different view- points while the scene consists of mostly rigid objects, sparse feature matching methods have proven highly effective. The J. Ma and Y. Ma are with the Electronic Information School, Wuhan University, Wuhan, 430072, China. E-mail: [email protected]; may- [email protected]. W. Qiu and A. Yuille are with the Department of Statistics, UCLA, Los Angeles, CA, 90095. E-mail: [email protected]; [email protected]. J. Zhao is with the Samsung Advanced Institute of Technology, Beijing, 100027, China. E-mail: [email protected]. Z. Tu is with the Department of Cognitive Science, UCSD, La Jolla, CA, 92697. E-mail: [email protected]. problem of dense correspondence is typically associated with image alignment/registration, which aims to overlaying two or more images with shared content, either at the pixel level (e.g., stereo matching [5] and optical flow [6], [7]) or the object/scene level (e.g., pictorial structure model [8] and SIFT flow [4]). It is a crucial step in all image analysis tasks in which the final information is gained from the combination of various data sources, e.g., in image fusion, change detec- tion, multichannel image restoration, as well as object/scene recognition. The registration problem can also be categorized into rigid or non-rigid registration depending on the form of the data. Rigid registration, which only involves a small number of parameters, is relatively easy and has been widely studied [1], [2], [9], [10], [11]. By contrast, non-rigid registration is more difficult because the underlying non-rigid transformations are often unknown, complex, and hard to model [3]. But non-rigid registration is very important because it is required for many real world tasks including hand-written character recognition, shape recognition, deformable motion tracking and medical image registration. In this paper, we focus on the non-rigid case and introduce a new algorithm for both sparse point set registration and dense image correspondence. To illustrate our main idea, we take the point set registration problem as an example. There are two unknown variables we have to solve for in this problem: the correspondence and the transformation. Although solving for either variable without information regarding the other is difficult, an iterated estimation framework can be used [2], [12], [3]. In this iterative process, the estimate of the correspondence is used to refine the estimate of the trans- formation, and vice versa. But a problem arises if there are errors in the correspondence as happens in many applications particularly if the transformation is large and/or there are outliers in the data (e.g., data points that are not undergoing the non-rigid transformation). In this situation, the estimate of the transformation will degrade badly unless it is performed robustly. The key issue of our approach is to robustly esti- mate the transformations from the correspondences using a robust estimator named the L 2 -Minimizing Estimate (L 2 E) [13], [14]. More precisely, our approach iteratively recovers the point correspondences and estimates the transformation between two point sets. In the first step of the iteration, feature descriptors such as shape context are used to establish correspondence. The correspondence set typically includes a large set of correspondence, and many of which are wrong. In the second step, we estimate the transformation using the robust estimator L 2 E. This estimator enable us to deal with

IEEE TRANSACTIONS ON SIGNAL PROCESSING 1 Robust L E ...ayuille/Pubs15/JiayiMaRobustL2E.pdf · features to obtain a higher level of image alignment (e.g., object/scene level). Here

  • Upload
    others

  • View
    1

  • Download
    0

Embed Size (px)

Citation preview

Page 1: IEEE TRANSACTIONS ON SIGNAL PROCESSING 1 Robust L E ...ayuille/Pubs15/JiayiMaRobustL2E.pdf · features to obtain a higher level of image alignment (e.g., object/scene level). Here

IEEE TRANSACTIONS ON SIGNAL PROCESSING 1

Robust L2E Estimation of Transformation forNon-Rigid Registration

Jiayi Ma, Weichao Qiu, Ji Zhao, Yong Ma, Alan L. Yuille, and Zhuowen Tu

Abstract—We introduce a new transformation estimation al-gorithm using the L2E estimator, and apply it to non-rigid reg-istration for building robust sparse and dense correspondences.In the sparse point case, our method iteratively recovers thepoint correspondence and estimates the transformation betweentwo point sets. Feature descriptors such as shape context areused to establish rough correspondence. We then estimate thetransformation using our robust algorithm. This enables us todeal with the noise and outliers which arise in the correspondencestep. The transformation is specified in a functional space, morespecifically a reproducing kernel Hilbert space. In the dense pointcase for non-rigid image registration, our approach consists ofmatching both sparsely and densely sampled SIFT features, andit has particular advantages in handling significant scale changesand rotations. The experimental results show that our approachgreatly outperforms state-of-the-art methods, particularly whenthe data contains severe outliers.

Index Terms—L2E estimator, registration, dense correspon-dence, regularization, outlier, non-rigid.

I. INTRODUCTION

Building “correct” registration, alignment or correspon-dence is a fundamental problem in computer vision, medicalimage analysis, and pattern recognition [1], [2], [3], [4]. Manytasks in these fields – such as stereo matching, shape matching,image registration and content-based image retrieval – canbe formulated as matching two sets of feature points, morespecifically, as a point matching problem [1]. The featurepoints in these tasks are typically the locations of sparse ordense interest points extracted from an image, or the edgepoints sampled from a shape contour. The registration problemthen reduces to determining the correct correspondence, and tofind the underlying spatial transformation between two pointsets extracted from the input data.

The registration problem can be categorized into sparsecorrespondence or dense correspondence, depending on thespecific application. The task of sparse correspondence aimsto match sparsely distributed points, such as matching contourpoints in shape matching. It also frequently arises in imagematching, for example, for images taken from different view-points while the scene consists of mostly rigid objects, sparsefeature matching methods have proven highly effective. The

J. Ma and Y. Ma are with the Electronic Information School, WuhanUniversity, Wuhan, 430072, China. E-mail: [email protected]; [email protected].

W. Qiu and A. Yuille are with the Department of Statistics, UCLA, LosAngeles, CA, 90095. E-mail: [email protected]; [email protected].

J. Zhao is with the Samsung Advanced Institute of Technology, Beijing,100027, China. E-mail: [email protected].

Z. Tu is with the Department of Cognitive Science, UCSD, La Jolla, CA,92697. E-mail: [email protected].

problem of dense correspondence is typically associated withimage alignment/registration, which aims to overlaying twoor more images with shared content, either at the pixel level(e.g., stereo matching [5] and optical flow [6], [7]) or theobject/scene level (e.g., pictorial structure model [8] and SIFTflow [4]). It is a crucial step in all image analysis tasks inwhich the final information is gained from the combinationof various data sources, e.g., in image fusion, change detec-tion, multichannel image restoration, as well as object/scenerecognition.

The registration problem can also be categorized into rigidor non-rigid registration depending on the form of the data.Rigid registration, which only involves a small number ofparameters, is relatively easy and has been widely studied [1],[2], [9], [10], [11]. By contrast, non-rigid registration is moredifficult because the underlying non-rigid transformations areoften unknown, complex, and hard to model [3]. But non-rigidregistration is very important because it is required for manyreal world tasks including hand-written character recognition,shape recognition, deformable motion tracking and medicalimage registration.

In this paper, we focus on the non-rigid case and introduce anew algorithm for both sparse point set registration and denseimage correspondence. To illustrate our main idea, we takethe point set registration problem as an example. There aretwo unknown variables we have to solve for in this problem:the correspondence and the transformation. Although solvingfor either variable without information regarding the otheris difficult, an iterated estimation framework can be used[2], [12], [3]. In this iterative process, the estimate of thecorrespondence is used to refine the estimate of the trans-formation, and vice versa. But a problem arises if there areerrors in the correspondence as happens in many applicationsparticularly if the transformation is large and/or there areoutliers in the data (e.g., data points that are not undergoingthe non-rigid transformation). In this situation, the estimate ofthe transformation will degrade badly unless it is performedrobustly. The key issue of our approach is to robustly esti-mate the transformations from the correspondences using arobust estimator named the L2-Minimizing Estimate (L2E)[13], [14]. More precisely, our approach iteratively recoversthe point correspondences and estimates the transformationbetween two point sets. In the first step of the iteration,feature descriptors such as shape context are used to establishcorrespondence. The correspondence set typically includes alarge set of correspondence, and many of which are wrong.In the second step, we estimate the transformation using therobust estimator L2E. This estimator enable us to deal with

Page 2: IEEE TRANSACTIONS ON SIGNAL PROCESSING 1 Robust L E ...ayuille/Pubs15/JiayiMaRobustL2E.pdf · features to obtain a higher level of image alignment (e.g., object/scene level). Here

IEEE TRANSACTIONS ON SIGNAL PROCESSING 2

the noise and outliers in the correspondences. The non-rigidtransformation is modeled in a functional space, called thereproducing kernel Hilbert space (RKHS) [15], in which thetransformation function has an explicit kernel representation.

We also apply our robust estimation of transformation tothe dense image correspondence/registration problem, whichaligns the images of the same scene taken under differentimaging conditions. The majority of the methods are basedon sparse feature correspondences and consists of four stepssuch as feature detection, feature matching, transformationestimation (rigid/non-rigid), and image resampling and trans-formation [16]. Recently, some new registration methods havealso been developed, where the goal is to align differentinstances of the same object/scene category, such as the pic-torial structure model for object recognition [8] and the SIFTflow for scene alignment [4]. Rather than operating on sparsefeature correspondences, these methods match dense sampledfeatures to obtain a higher level of image alignment (e.g.,object/scene level). Here we generalize our robust algorithm,and introduce a novel image registration method which con-sists of matching both sparsely and densely sampled features.Our formulation contains two coupling variables: the non-rigidgeometric transformation and the discrete dense flow field,where the former corresponds to the sparse feature matchingand the latter corresponds to the dense feature matching.Ideally, the two variables are equivalent. We alternatively solvefor one variable under the assumption that the other is known.

The main contributions of our work include: (i) the proposalof a new robust algorithm to estimate a spatial transforma-tion/mapping from correspondences with noise and outliers;(ii) application of this algorithm to non-rigid sparse corre-spondence such as point set registration and sparse imagefeature correspondence; (iii) further application of this robustalgorithm to non-rigid dense correspondence such as imageregistration, and it can handle significantly changes in scaleand rotation; (iv) both the sparse and dense matching algo-rithms are demonstrably effective and outperform state-of-the-art methods. This article is an extension of our earlierwork [17]. The primary new contributions are an expandedderivation and discussion of the robust L2E estimator (SectionIII) and a new algorithm for non-rigid image registration basedon the proposed robust algorithm (Section V).

The rest of the paper is organized as follows. Section 2describes relevant previous work, followed by our robust L2Ealgorithm for estimating transformations from point correspon-dences with unknown outliers in Section 3. In Section 4, weapply our robust algorithm to non-rigid sparse correspondencesuch as point set registration and sparse image feature corre-spondence. We further apply our robust algorithm to non-rigiddense correspondence, i.e., image registration, in Section 5.Section 6 demonstrates the experimental results of our sparseand dense matching algorithms on both 2D and 3D, syntheticand real data. Finally, we conclude this paper in Section 7.

II. RELATED WORK

We apply our robust algorithm to sparse and dense cor-respondence, e.g., non-rigid point set registration and image

registration. It is beyond the scope of this paper to give athorough review on these topics. In this section we brieflyoverview some works that are most relevant to our approach.

A. Point Set Registration Methods

The iterated closest point (ICP) algorithm [2] is one ofthe best known point registration approaches. It uses nearest-neighbor relationships to assign a binary correspondence, andthen uses estimated correspondence to refine the spatial trans-formation. Moreover, it requires a good initial transformationthat places the two data sets in approximate registration.Belongie et al. [12] introduced a method for shape registrationbased on the shape context descriptor, which incorporates theneighborhood structure of the point set and thus helps establishcorrespondence between the point sets. These methods how-ever ignore robustness when they recover the transformationfrom the correspondence.

For robust point feature matching, the random sampleconsensus (RANSAC) [18] is a widely used algorithm incomputer vision. It uses a hypothesize-and-verify and tries toget as small an outlier-free subset as feasible to estimate agiven parametric model by resampling. RANSAC has sever-al variants such as MLESAC [19], LO-RANSAC [20] andPROSAC [21]. Although these methods are very successful inmany situations they have had limited success if the under-lying spatial transformation between point features are non-parametric, for example if the real correspondence is non-rigid.In related work, Chui and Rangarajan [3] established a generalframework for estimating correspondence and transformationsfor non-rigid point matching. They modeled the transformationas a thin-plate spline and did robust point matching by analgorithm (TPS-RPM) which involved deterministic annealing[22] and soft-assignment [23]. Alternatively, the coherencepoint drift (CPD) algorithm [24] uses Gaussian radial basisfunctions instead of thin-plate splines. These algorithms arerobust compared to ICP in the non-rigid case, but the jointestimation of correspondence and transformation increases thealgorithm complexity.

Another interesting point matching approach is the kernelcorrelation (KC) based method [25]. The cost function ofKC is proportional to the correlation of two kernel densityestimates. Jian and Vemuri [26] presented a unified frameworkbased on a divergence family that allows for interpretingseveral of the existing point set registration methods, e.g.[27], [3], [24], as special cases of the divergence family.In particular, the L2 distance between Gaussians is stronglyrelated to the L2E estimator used in our method. Zhengand Doermann [28] introduced the notion of a neighborhoodstructure for the general point matching problem, and proposeda matching method, the robust point matching-preserving localneighborhood structures (RPM-LNS) algorithm, which waslater generalized in [29] by introducing an optimal compatibili-ty coefficient for the relaxation labeling process to solve a non-rigid point matching problem. Huang et al. [30] proposed toimplicitly embed the shapes of interest in a higher-dimensionalspace of distance transforms, and align them similar to non-rigid image registration algorithms. The method performs well

Page 3: IEEE TRANSACTIONS ON SIGNAL PROCESSING 1 Robust L E ...ayuille/Pubs15/JiayiMaRobustL2E.pdf · features to obtain a higher level of image alignment (e.g., object/scene level). Here

IEEE TRANSACTIONS ON SIGNAL PROCESSING 3

on relatively simple data sets. Other related work includes theEM-like algorithm for matching rigid and articulated shapes[31], as well as the graph matching approach for establishingfeature correspondences [32].

B. Image Registration Methods

Image registration, as was mentioned above, is widely usedin computer vision, medical imaging, remote sensing, etc. Anearly example of a widely-used image registration algorithmis the optical flow technique [6], [7]. It computes a densecorrespondence field by directly minimizing pixel-to-pixeldissimilarities, and hence tends to operate on very similarimages, e.g., two adjacent frames in a video sequence. Typicalassumptions in optical flow algorithms include brightnessconstancy and piecewise smoothness of the pixel displacementfield.

Due to changes of lighting, perspective and noise, the pixelvalues are often not reliable for registration [33]. Nevertheless,the development of various local invariant features has broughtabout significant progress in this area [34], [35]. The feature-based approaches are to some extent similar to the point setregistration methods. They work by extracting a set of sparsefeatures and then matching them to each other. The spatialform of the correspondence then can be described by paramet-ric models such as affine and homography, or non-parametricmodels such as radial basis functions, thin-plate splines andelastic transforms, relating pixel coordinates in one image topixel coordinates in another. This type of methods has theadvantage of being more robust to typical appearance changesand scene movements, and is potentially faster, if implementedthe right way. However, they are often designed for matchingrigid scenes (e.g., in stereo matching and image stitching) orsmall non-rigid motions (e.g., in medical imaging), but lesseffective for handling objects and scenes with significant non-rigidity. This can be explained since the dense correspondencehere is interpolated from sparse matching rather then throughpixel-wise correspondence, which may be problematic whenthe real correspondence is non-rigid and the transform modelis not known apriori. For comprehensive reviews of this classof methods please refer to [16], [36], [37].

Recently, Liu et al. [4] proposed a SIFT flow algorithmfor non-rigid matching of highly different scenes. Instead ofmatching brightness of pixels in optical flow algorithms, theSIFT flow algorithm matches densely sampled SIFT featuresbetween two images. It has demonstrated impressive dense,pixel-wise correspondence results, however, it is not robust tosignificantly changes in scale and rotation. HaCohen et al. [38]proposed a non-rigid dense correspondence method (NRDC)which combines dense local matching with robustness tooutliers based on Generalized PatchMatch [39]. Moreover, itcan address large scale change and rotation. Ma et al. [40]proposed a regularized Gaussian field criterion for multimodalimage registration, such as visible and infrared face images. Inthis paper, we introduce a novel dense correspondence methodmatching both sparsely and densely sampled SIFT featuresbased on the robust L2E estimator and SIFT flow. We showthat our method significantly outperforms both SIFT flow and

NRDC in a set of quantitative evaluation. Besides, Vemuriet al. [41] and Liu et al. [42] proposed a robust multimodalimage registration method based on matching dominant localfrequency image representations, and they are the first whoapplied the L2E estimator in the image registration problem.

III. ESTIMATING TRANSFORMATION FROMCORRESPONDENCES BY L2E

Given a set of point correspondences S = {(xi,yi)}ni=1,which are typically perturbed by noise and by outlier pointswhich undergo different transformations, the goal is to esti-mate a transformation f : yi = f(xi) and fit the inliers.

In this paper we make the assumption that the noise on theinliers is Gaussian on each component with zero mean anduniform standard deviation σ (our approach can be directlyapplied to other noise models). More precisely, an inlier pointcorrespondence (xi,yi) satisfies yi − f(xi) ∼ N(0, σ2I),where I is an identity matrix of size d× d, with d being thedimension of the point. The data {yi− f(xi)}ni=1 can then bethought of as a sample set from a multivariate normal densityN(0, σ2I) which is contaminated by outliers. The main idea ofour approach is then, to find the largest portion of the sampleset (e.g., the underlying inlier set) that “matches” the normaldensity model, and hence estimate the transformation f forthe inlier set. Next, we introduce a robust estimator namedL2-minimizing estimate (L2E) which we use to estimate thetransformation f .

A. Problem Formulation Using L2E: Robust Estimation

Parametric estimation is typically done using maximumlikelihood estimation (MLE). It can be shown that MLE isthe optimal estimator if it is applied to the correct probabilitymodel for the data (or to a good approximation). But MLEcan be badly biased if the model is not a sufficiently goodapproximation or, in particular, there are a significant fractionof outliers. In the point matching problem, it is generallyaccepted that incorrect matches, or outliers, cannot be avoidedin the matching process where only local feature descriptorsare compared. In this case, a robust estimator of the transfor-mation f is desirable because the point correspondence set Susually contains outliers. There are two choices: (i) to builda more complex model that includes the outliers – which iscomplex since it involves modeling the outlier process usingextra (hidden) variables which enable us to identify and rejectoutliers, or (ii) to use an estimator which is different fromMLE but less sensitive to outliers, as described in Huber’srobust statistics [43]. In this paper, we use the second methodand adopt the L2E estimator [13], [14], a robust estimatorwhich minimizes the L2 distance between densities, and isparticularly appropriate for analyzing massive data sets wheredata cleaning (to remove outliers) is impractical.

For the parametric setting with a density model p(z|θ),consider minimizing an estimate of L2 distance with respect

Page 4: IEEE TRANSACTIONS ON SIGNAL PROCESSING 1 Robust L E ...ayuille/Pubs15/JiayiMaRobustL2E.pdf · features to obtain a higher level of image alignment (e.g., object/scene level). Here

IEEE TRANSACTIONS ON SIGNAL PROCESSING 4

MLE

z

log

(z|

,1) z=

L2E

z=

z

(z|

,1)

Fig. 1: The penalty curves of MLE (left) and L2E (right) for an one-dimensional Gaussian model.

to θ:

θ = arg minθ

∫[p(z|θ)− p(z|θ0)]2dz

= arg minθ

[∫p(z|θ)2dz − 2

∫p(z|θ)p(z|θ0)dz

+

∫p(z|θ0)2dz

]= arg min

θ

[∫p(z|θ)2dz − 2E[p(z|θ)]

], (1)

where the true parameter θ0 is unknown, and we omit the (un-known) constant

∫p(z|θ0)2dz. Based on the L2 distance, the

L2E estimator for model p(z|θ) then recommends estimatingthe parameter θ by minimizing the criterion:

θL2E = arg minθ

[∫p(z|θ)2dz − 2

n

n∑i=1

p(zi|θ)

]. (2)

To get some intuition for why L2E is robust, we consider asimple one-dimensional Gaussian model p(z|θ) = φ(z|µ, 1),where µMLE = arg maxµ

∑ni=1 log φ(zi|µ, 1) and µL2E =

arg minµ

[1

2√π− 2

n

∑ni=1 φ(zi|µ, 1)

]. The penalty curves of

MLE (i.e. − log φ(z|µ, 1)) and L2E (i.e., −φ(z|µ, 1)) areshown in Fig. 1. From the figure, we see that the penalty curveof MLE is quadratic, and then it is reluctant to assign lowprobability to any points (which becomes infinite as φ(zi|µ, 1)tends to 0), including outliers, and hence tends to be biased byoutliers. By contrast, the penalty of L2E can be approximatelyseen as a truncated form of the MLE penalty, and then itcan assign low probabilities to many points, hopefully to theoutliers, without paying too high a penalty. Moreover, unlikea truncated penalty, the L2E functional is derivable and it hasthe advantage of computational convenience.

To further demonstrate the robustness of L2E, we presenta line-fitting example which contrasts the behavior of MLEand L2E, see Fig. 2. The goal is to fit a linear regres-sion model, y = αx + ε, with residual ε ∼ N(0, 1),by estimating α using MLE and L2E. This gives, re-spectively, αMLE = arg maxα

∑ni=1 log φ(yi − αxi|0, 1)

and αL2E = arg minα

[1

2√π− 2

n

∑ni=1 φ(yi − αxi|0, 1)

]. As

shown in Fig. 2, L2E is very resistant when we contaminatethe data by outliers, but MLE does not show this desirableproperty. L2E always has a global minimum at approximately0.5 (the correct value for α) but MLE’s estimates becomesteadily worse as the amount of outliers increases. Observe,in the bottom right figure, that L2E also has a local minimumnear α = 2, which becomes deeper as the number n of outliers

0 2 4 6 8 10

0

10

20

30

40

50

x

y

0 2 4 6 8 10−5

0

5

10

15

20

25

x

y

0 1 2 3 4 5−6

−5

−4

−3

−2

−1

0

x 104

α

log−

likel

ihoo

d

100

160

120140

180200

0 0.5 1 1.5 2 2.5−9

−7

−5

−3

−1

1

α

log−

likel

ihoo

d

x 103

120140

160200

180

100

0 1 2 3 4 5−1.2

−0.8

−0.4

0

0.4

α

L2E

100

120

200180160140

α = 0.5

0 0.5 1 1.5 2 2.5−1.2

−0.8

−0.4

0

0.4

α

L2E

200180160

140120

100

Fig. 2: Comparison between L2E and MLE for linear regression as thenumber of outliers varies. Top row: data samples, where the inliers areshown by cyan pluses, and the outliers by magenta circles. The goal isto estimate the slope α of the line model y = αx. We vary the number ofdata samples n = 100, 120, · · · , 200, by always adding 20 new outliers(retaining the previous samples). In the second column the outliers aregenerated from another line model y = 2x. Middle and bottom rows: thecurves of MLE and L2E respectively. The MLE estimates are correctfor n = 100 but rapidly degrade as we add outliers, see how the peakof the log-likelihood changes in the second row. By contrast, (see thirdrow) L2E estimates α correctly even when half the data is outliers andalso develops a local minimum to fit the outliers when appropriate (thirdrow, right column). Best viewed in color.

increases so that the two minima become approximately equalwhen n = 200. This is appropriate because, in this case, thecontaminated data also comes from the same linear parametricmodel with slope α = 2, e.g., y = 2x.

We now apply the L2E formulation in (2) to the pointmatching problem, assuming that the noise of the inliersis given by a normal distribution, and obtain the followingfunctional criterion:

L2E(f , σ2) =1

2d(πσ)d/2− 2

n

n∑i=1

Φ(yi − f(xi)|0, σ2I

). (3)

where Φ is a multi-dimensional Gaussian function. We modelthe non-rigid transformation f by requiring it to lie within aspecific functional space, namely a reproducing kernel Hilbertspace (RKHS) [15], [44], [45]. Note that other parameterizedtransformation models, for example, thin-plate splines (TPS)[46], [47], [48], can also be easily incorporated into ourformulation.

We define an RKHS H by a positive definite matrix-valued kernel Γ(x,x′) : IRd × IRd → IRd×d. The optimaltransformation f which minimizes the L2E functional (3) then

Page 5: IEEE TRANSACTIONS ON SIGNAL PROCESSING 1 Robust L E ...ayuille/Pubs15/JiayiMaRobustL2E.pdf · features to obtain a higher level of image alignment (e.g., object/scene level). Here

IEEE TRANSACTIONS ON SIGNAL PROCESSING 5

takes the form f(x) =∑ni=1 Γ(x,xi)ci [45], [49], where the

coefficient ci is a d×1 dimensional vector (to be determined).Hence, the minimization over the infinite dimensional Hilbertspace reduces to finding a finite set of n coefficients ci. But inpoint correspondence problem the point set typically containshundreds or thousands of points, which causes significant com-plexity problems (in time and space). Consequently, we adopta sparse approximation, and randomly pick only a subset ofsize m input points {xi}mi=1 to have nonzero coefficients in theexpansion of the solution. This follows [50], [51] who foundthat this approximation works well and that simply selecting arandom subset of the input points in this manner, performs noworse than more sophisticated and time-consuming methods.Therefore, we seek a solution of form

f(x) =

m∑i=1

Γ(x, xi)ci. (4)

The chosen point set {xi}mi=1 are somewhat analogous to“control points” [1]. By including a regularization term forimposing smooth constraint on the transformation, the L2Efunctional (3) becomes:

L2E(f , σ2) =1

2d(πσ)d/2− 2

n

n∑i=1

1

(2πσ2)d/2

e−‖yi−∑m

j=1 Γ(xi,xj)cj‖22σ2 + λ‖f‖2Γ, (5)

where λ > 0 controls the strength of regularization, andthe stabilizer ‖f‖2Γ is defined by an inner product, e.g.,‖f‖2Γ = 〈f , f〉Γ. By choosing a diagonal decomposable kernel[49]: Γ(xi,xj) = κ(xi,xj) · I = e−β‖xi−xj‖

2

I with βdetermining the width of the range of interaction betweensamples (i.e. neighborhood size), the L2E functional (5) maybe conveniently expressed in the following matrix form:

L2E(C, σ2) =1

2d(πσ)d/2− 2

n

n∑i=1

1

(2πσ2)d/2e−‖yTi −Ui,·C‖2

2σ2

+ λ tr(CTΓC), (6)

where kernel matrix Γ ∈ IRm×m is called the Gram matrixwith Γij = κ(xi, xj) = e−β‖xi−xj‖

2

, U ∈ IRn×m with Uij =

κ(xi, xj) = e−β‖xi−xj‖2

, Ui,· denotes the i-th row of thematrix U, C = (c1, · · · , cm)T is the coefficient matrix ofsize m× d, and tr(·) denotes the trace.

B. Estimation of the Transformation

Estimating the transformation requires taking the derivativeof the L2E cost function, see equation (6), with respect to thecoefficient matrix C, which is given by:

∂L2E

∂C=

2UT[P ◦ (Q⊗ 11×d)]

nσ2(2πσ2)d/2+ 2λΓC, (7)

where P = UC−Y and Y = (y1, · · · ,yn)T are matrices ofsize n×d, Q = exp{diag(PPT)/2σ2} is an n×1 dimensionalvector, diag(·) is the diagonal of a matrix, 11×d is an 1 × ddimensional row vector of all ones, ◦ denotes the Hadamardproduct, for example (A ◦ B)ij = Aij · Bij , and ⊗ denotes

Algorithm 1: Estimate of Transformation from Correspon-dences

Input: Correspondence set S = {(xi,yi)}ni=1,parameters γ, β, λ

Output: Optimal transformation f1 Construct Gram matrix Γ and matrix U;2 Initialize parameter σ2 and C;3 Deterministic annealing:4 Using the gradient (7), optimize the objective function

(6) by a numerical technique (e.g., the quasi-Newtonalgorithm with C as the old value);

5 Update the parameter C← arg minC L2E(C, σ2);6 Anneal σ2 = γσ2;7 The transformation f is determined by equation (4).

the Kronecker product, for example, for A ∈ IRm×n and B ∈

IRp×q , A⊗B =

a11B · · · a1nB...

. . ....

am1B · · · amnB

.

By using the derivative in equation (7), we can employ effi-cient gradient-based numerical optimization techniques suchas the quasi-Newton method and the nonlinear conjugategradient method to solve the optimization problem. But thecost function (6) is convex only in the neighborhood of theoptimal solution. Hence to improve convergence we use acoarse-to-fine strategy by applying deterministic annealing onthe inlier noise parameter σ2. This starts with a large initialvalue for σ2 which is gradually reduced by σ2 7→ γσ2,where γ is the annealing rate. Our algorithm is outlined inAlgorithm 1. Note that our proposed algorithm can be easilyextended to the rigid or affine model, by using a rotation matrixR or a nonsingular matrix A together with a translation vectort, i.e., f(x) = Rx + t or f(x) = Ax + t. Here we omit thedetails of the derivation and focus on the non-rigid case only.Computational complexity. By examining equations (6) and(7), we see that the costs of updating the objective functionand gradient are both O(dm2 + dmn). For the numericaloptimization method, we choose the Matlab Optimizationtoolbox (e.g., the Matlab function fminunc), which implicitlyuses the BFGS Quasi-Newton method with a mixed quadraticand cubic line search procedure. Thus the total complexity isapproximately O(dm2 +dm3 +dmn). In our implementation,the number m of the control points required to construct thetransformation f in equation (4) is in general not large, andso use m = 15 for all the results in this paper (increasingm only gave small changes to the results). The dimension dof the data in feature point matching for vision applicationsis typically 2 or 3. Therefore, the complexity of our methodcan be simply expressed as O(n), which is about linear in thenumber of correspondences. This is important since it enablesour method to be applied to large scale data.

C. Implementation Details

The performance of point matching algorithms depends,typically, on the coordinate system in which points are ex-pressed. We use data normalization to control for this. More

Page 6: IEEE TRANSACTIONS ON SIGNAL PROCESSING 1 Robust L E ...ayuille/Pubs15/JiayiMaRobustL2E.pdf · features to obtain a higher level of image alignment (e.g., object/scene level). Here

IEEE TRANSACTIONS ON SIGNAL PROCESSING 6

specifically, we perform a linear re-scaling of the correspon-dences so that the points in the two sets both have zero meanand unit variance.

We define the transformation f as the initial position plusa displacement function v: f(x) = x + v(x) [24], and solvefor v instead of f . This can be achieved simply by setting theoutput yi to be yi − xi.Parameter settings. Our method uses deterministic annealingto deal with the non-convexity, and so it will fail if theannealing fails. The deterministic annealing is more likely towork if we anneal slowly, and the annealing rate is controlledby parameter γ. Without considering the efficiency issue, wecan always set it as large as possible (e.g., close to 1), andthen use more iterations to produce satisfied results. Twoother parameters are β and λ, which control the influence ofthe smoothness constraint on the transformation f . Parameterβ determines how wide the range of interaction betweensamples. Since we have used data normalization so that thepoints in the two sets both have zero mean and unit variance,the optimal value of parameter β in general will be similar ondifferent examples. Parameter λ controls the trade-off betweenthe closeness to the data and the smoothness of the solution.It is mainly influenced by the degrees of data degradation.We set γ = 0.5, β = 0.8 and λ = 0.1 throughout this paper.Finally, the parameter σ2 and C in line 2 of Algorithm 1 wereinitialized to 0.05 and 0 respectively.

IV. APPLICATION TO NON-RIGID SPARSECORRESPONDENCE

In this section, we apply our robust algorithm to estab-lishment of non-rigid sparse correspondence. Two tasks areconsidered such as non-rigid point set registration and non-rigid image feature correspondence.

A. Non-Rigid Point Set Registration

Point set registration aims to align two point sets {xi}ni=1

(the model point set) and {yj}lj=1 (the target point set).Typically, in the non-rigid case, it requires estimating a non-rigid transformation f which warps the model point set tothe target point set. We have shown above that once we haveestablished the correspondence between the two point setseven with noise and outliers, we are able to estimate the under-lying transformation between them. Since our method does notjointly solve the transformation and point correspondence, inorder to use Algorithm 1 to solve the transformation betweentwo point sets, we need initial correspondences.

In general, if the two point sets have similar shapes, thecorresponding points have similar neighborhood structureswhich could be incorporated into a feature descriptor. Thusfinding correspondences between two point sets is equivalentto finding for each point in one point set (e.g., the model)the point in the other point set (e.g., the target) that hasthe most similar feature descriptor. Fortunately, the initialcorrespondences need not be very accurate, since our methodis robust to noise and outliers. Inspired by these facts, weuse shape context [12] as the feature descriptor in the 2Dcase, using the Hungarian method for matching with the

Algorithm 2: Non-Rigid Point Set Registration

Input: Two point sets {xi}ni=1, {yj}lj=1

Output: Aligned model point set {xi}ni=1

1 Compute feature descriptors for the target point set{yj}lj=1;

2 repeat3 Compute feature descriptors for the model point set

{xi}ni=1;4 Estimate the initial correspondences based on the

feature descriptors of two point sets;5 Solve the transformation f warping the model point

set to the target point set using Algorithm 1;6 Update model point set {xi}ni=1 ← {f(xi)}ni=1;7 until reach the maximum iteration number;8 The aligned model point set {xi}ni=1 is given by{f(xi)}ni=1 in the last iteration.

χ2 test statistic as the cost measure. In the 3D case, thespin image [52] can be used as a feature descriptor, wherethe local similarity is measured by an improved correlationcoefficient. Then the matching is performed by a methodwhich encourages geometrically consistent groups.

The two steps of estimating correspondences and transfor-mations are iterated to obtain a reliable result. In this paper,we use a fixed number of iterations, typically 10 but morewhen the noise is big or when there are a large percentageof outliers contained in the original point sets. We summarizeour point set registration method in Algorithm 2.

B. Non-Rigid Image Feature Correspondence

The image feature correspondence task aims to find visualcorrespondences between two sets of sparse feature points{xi}ni=1 and {yj}lj=1 with corresponding feature descriptorsextracted from two input images. In this paper, we assumethat the underlying relationship between the input images isnon-rigid. Our method for this task is to estimate correspon-dences by matching feature descriptors using a smooth spatialmapping f . More specifically, we first estimate the initialcorrespondences based on the feature descriptors, and thenuse the correspondences to learn a spatial mapping f fittingthe inliers by algorithm 1.

Once we have obtained the spatial mapping f , we then haveto establish accurate correspondences. We predefine a thresh-old τ and judge a correspondence (xi,yj) to be an inlier pro-vided it satisfies the following condition: e−‖yj−f(xi)‖

2/2σ2

>τ . We set τ = 0.5 in this paper.

Note that the feature descriptors in the point set registrationproblem are calculated based on the point sets themselves, andare recalculated in each iteration. However, the descriptors ofthe feature points here are fixed and calculated from imagesin advance. Hence the iterative technique for recovering cor-respondences, estimating spatial mapping, and re-estimatingcorrespondences can not be used here. In practice, we find thatour method works well without iteration due to the followingtwo reasons: i) the initial correspondences between the given

Page 7: IEEE TRANSACTIONS ON SIGNAL PROCESSING 1 Robust L E ...ayuille/Pubs15/JiayiMaRobustL2E.pdf · features to obtain a higher level of image alignment (e.g., object/scene level). Here

IEEE TRANSACTIONS ON SIGNAL PROCESSING 7

two feature point sets in general contain most of the groundtruth correspondences; ii) we focus on determining the rightcorrespondences which does not need precise recovery of theunderlying transformation. Our approach then plays a role ofrejecting outliers.

V. APPLICATION TO NON-RIGID DENSECORRESPONDENCE

We now focus on the image registration/dense matchingproblem. Given two input images, e.g. a model image anda target image, the goal is to establish dense correspondencesbetween the two images. In this paper we assume that theunderlying relationship between the input images is non-rigid.Next, we introduce a novel dense matching method whichconsists of matching both sparsely and densely sampled SIFTfeatures.

A. Problem Formulation

To align two images, we extract a set of sparse feature corre-spondences {(xi,yi)}ni=1, e.g. SIFT feature correspondences[34], with xi,yi ∈ IR2 being the spatial locations of the featurepoints in the model and target images respectively. Generally,the correspondences contain some unknown outliers, and therobust L2E estimator can be used to recover a transformationf according to Algorithm 1. This sparse feature detection andmatching has been a standard approach to image registrationof the same scene [16]. However, for image pairs containingunknown non-rigid motions, this procedure in general cannotobtain accurate dense correspondences. To address this prob-lem, here we also match densely sampled SIFT features basedon the SIFT flow algorithm [4].

We define the transformation between the sparse correspon-dences as f(x) = x + v(x), where v is the displacementfunction with the form in equation (4). To match dense SIFTfeatures, we introduce a dense discrete flow w: let p bethe grid coordinate of the model image, w(p) then is thedisplacement vector at p, i.e., the point p in the model imagecorresponds to the point p + w(p) in the target image. Inaddition, w is only allowed to have integral entries. Let s1

and s2 be the per-pixel SIFT features for two images. Theset ε contains all the spatial neighborhoods (a four-neighborsystem is used). Denote w(p) = (w1(p),w2(p))T, and l thetotal number of evaluation pixels. The cost function in ourdense correspondence problem is then defined as:

E(v,w, σ2) =1

4πσ− 1

nπσ2

n∑i=1

e−‖yi−xi−v(xi)‖

2

2σ2 + λ‖v‖2Γ

l

∑p

min(‖s1(p)− s2(p + w(p))‖1, t)

l

∑(p,q)∈ε

2∑i=1

min(α|wi(p)−wi(q)|, u)

l

∑p

η‖v(p)−w(p)‖2. (8)

We briefly go through all the components of the cost function.The first three terms are from the L2E estimator as in equation

(5), which performs a robust estimation of the displacementfunction v from sparse correspondences. The fourth, fifth andsixth terms correspond to the data term, smoothness termand small displacement term of the SIFT flow algorithm,which performs dense matching between two images. The onlydifference is that the last term in the SIFT flow algorithm is∑

p ‖w(p)‖1, rather than∑

p ‖v(p) − w(p)‖2 in our algo-rithm. Here this term acts as a bridge between the sparse andthe dense feature matchings; it constraints the displacementfunction v and flow w to be consistent.

B. Optimization

There are two unknown variables in the cost function–thedisplacement function v and the flow w. While solving foreither variable without information regarding the other is quitedifficult, an interesting fact is that solving for one variable oncethe other is known is much simpler than solving the original,coupled problem.

Now we first consider the terms of cost function (8) thatrelated to v, which can be expressed in the following matrixform:

E(C, σ2) =1

4πσ− 1

nπσ2

n∑i=1

e−‖yTi −xT

i −Ui,·C‖2

2σ2

+ λtr(CTΓC) +δη

l‖VC−W‖2F , (9)

where the matrices Γ, U and C are defined as the same asin equation (6), matrix V ∈ IRl×m with Vij = κ(pi, xj) =

e−β‖pi−xj‖2

, W = (w1, · · · ,wl)T is the flow field of size

l × 2, and ‖ · ‖F denotes the Frobenius norm.As described in Algorithm 1, the cost function (9) can be

minimized by using efficient gradient-based numerical opti-mization techniques combining with deterministic annealing,where the derivative of the objective function with respect tothe coefficient matrix C is given by:∂E(C, σ2)

∂C=

UT[P ◦ (Q⊗ 11×2)]

nπσ4+2λΓC+

2δη

lVT(VC−W),

(10)where P = X + UC − Y and Q = exp{diag(PPT)/2σ2}are defined similar to those in equation (7).

Next we consider the terms of cost function (8) related tow, which involves the last three terms

E(w) =∑p

min(‖s1(p)− s2(p + w(p))‖1, t)

+∑

(p,q)∈ε

2∑i=1

min(α|wi(p)−wi(q)|, u)

+∑p

η‖v(p)−w(p)‖2. (11)

To solve the flow w, we utilize the SIFT flow algorithm [4]by modifying the small displacement term from

∑p ‖w(p)‖1

to∑

p ‖v(p)−w(p)‖2.The two steps of estimating displacement function v and

flow w are iterated to obtain a reliable result. Note that thevalue of objective function (8) will decrease in each stepduring the iteration, which guarantees the convergence of thealgorithm. The number of iteration is fixed to 3.

Page 8: IEEE TRANSACTIONS ON SIGNAL PROCESSING 1 Robust L E ...ayuille/Pubs15/JiayiMaRobustL2E.pdf · features to obtain a higher level of image alignment (e.g., object/scene level). Here

IEEE TRANSACTIONS ON SIGNAL PROCESSING 8

Algorithm 3: Non-Rigid Dense Image CorrespondenceInput: A pair of images, parameters γ, β, λ, α, η, u, δOutput: A dense flow w

1 Extract a set of sparse SIFT correspondence from theimage pair: S = {(xi,yi)}ni=1;

2 Establish dense SIFT features {s1} and {s2} for twoimages respectively according to the orientations andscales of the sparse correspondences;

3 Construct Gram matrix Γ, matrix U and V;4 repeat5 if the first iteration6 Initialize parameter σ2;7 Compute C by using Algorithm 1;8 else9 Initialize parameter σ2, and initialize C as its old

value;10 Optimize the objective function (9) by a

quasi-Newton method together with deterministicannealing to compute C;

11 end12 Update the displacement function v← VC;13 Optimize the objective function (11) by using the

SIFT flow algorithm to compute the flow w;14 until reach the maximum iteration number;15 The dense image correspondence can be obtain from w

after the iteration.

C. Implementation Details

Observe that the last term of objective function (9) is theaverage difference between v and w on the image lattice.Therefore, to estimate the coefficient matrix C, it is possible todownsample the image lattice to achieve a significant speedupwithout much performance degradation. In our evaluation, weuse the uniform sampling strategy with a sampling interval 10pixels. Besides, we initialize C independent of W by usingAlgorithm 1. We also use data normalization so that the twosets of sparse feature points {xi}ni=1 and {yi}ni=1 both havezero mean and unit variance.

The SIFT flow algorithm suffers from the scale and rotationproblem. Here we address this issue in our dense matchingscheme. Note that the sparse SIFT matching provides theorientations and scales for each correspondence, and for eachinlier correspondence, the orientation difference and scale ratiobetween the two feature points are in general constants. Basedon this observation, we choose a small part of the sparseSIFT correspondences (e.g., 20%) with the highest matchingscores which are in general inliers, and then set the orientationand scale of the dense SIFT features to the mean orientationdifference and mean scale ratio of the chosen correspondences.

The parameters in our dense correspondence algorithm areset according to Algorithm 1 and the SIFT flow algorithm.There is an additional parameter δ need to be set in equation(9) after the first iteration, which controls the trade-off betweenthe sparse matching and dense matching. We set it accordingto the equation δη = 103. Our dense image correspondencemethod is summarized in Algorithm 3.

VI. EXPERIMENTAL RESULTS

In order to evaluate the performance of our algorithm, weconducted three types of experiments: i) point set registrationfor 2D shapes; ii) sparse feature correspondence on 2D imagesand 3D surfaces; iii) 2D dense image correspondence.

A. Results on Non-Rigid Point Set Registration

We tested our method on the same synthesized data as in[3] and [28]. The data consists of two different shape models:a fish and a Chinese character. For each model, there are fivesets of data designed to measure the robustness of registrationalgorithms under deformation, occlusion, rotation, noise andoutliers. In each test, one of the above distortions is appliedto a model set to create a target set, and 100 samples aregenerated for each degradation level. We use the shape contextas the feature descriptor to establish initial correspondences. Itis easy to make shape context translation and scale invariant,and in some applications, rotation invariance is also required.We use the rotation invariant shape context as in [28].

Fig. 3a shows the registration results of our method onsolving different degrees of deformations and occlusions. Wealso give a quantitative evaluation in Fig. 3b, where the recallis computed as the metric used in [26]. Here the recall isdefined as the proportion of true positive correspondencesto the ground truth correspondences and a true positive cor-respondence is counted when the pair falls within a givenaccuracy threshold in terms of pairwise distance, e.g., theEuclidean distance between a point in the warped modeland the corresponding point in the target. As shown in thefigure, we see that for both datasets with moderate degradation,our method is able to produce an almost perfect alignment.Moreover, the matching performance degrades gradually andgracefully as the degree of degradation in the data increases.Consider the results on the occlusion test in the fifth column, itis interesting that even when the occlusion ratio is 50 percentour method can still achieve a satisfactory registration result.Therefore our method can be used to provide a good initialalignment for more complicated problem-specific registrationalgorithms. The average run time of our method on this datasetwith about 100 points is about 0.5 seconds on an Intel Core2.5 GHz PC with Matlab code.

To provide a quantitative comparison, we report the resultsof five state-of-the-art algorithms such as shape context [12],TPS-RPM [3], RPM-LNS [28], GMMREG [26] and CPD[24] which are implemented using publicly available codes.For the GMMREG algorithm, we choose the L2 distance andTPS kernel for evaluation. The registration error on a pair ofshapes is quantified as the average Euclidean distance betweena point in the warped model and the corresponding point inthe target. Then the registration performance of each algorithmis compared by the mean and standard deviation of theregistration error of all the 100 samples in each distortion level.The statistical results, error means, and standard deviations foreach setting are summarized in Fig. 3c. In the deformation testresults (e.g., 1st and 3rd rows), six algorithms achieve similarregistration performance in both fish and Chinese character atlow deformation levels, and our method generally gives better

Page 9: IEEE TRANSACTIONS ON SIGNAL PROCESSING 1 Robust L E ...ayuille/Pubs15/JiayiMaRobustL2E.pdf · features to obtain a higher level of image alignment (e.g., object/scene level). Here

IEEE TRANSACTIONS ON SIGNAL PROCESSING 9

(a)

0 0.005 0.01 0.015 0.020

20

40

60

80

100

Accuracy threshold

Rec

all (

%)

0.020.0350.050.0650.08

0 0.005 0.01 0.015 0.020

20

40

60

80

100

Accuracy threshold

Rec

all (

%)

0.10.20.30.40.5

0 0.01 0.02 0.03 0.040

20

40

60

80

100

Accuracy thresholdR

ecal

l (%

)

0.020.0350.050.0650.08

0 0.01 0.02 0.03 0.040

20

40

60

80

100

Accuracy threshold

Rec

all (

%)

0.10.20.30.40.5

(b)

0.02 0.035 0.05 0.065 0.080.04

0.02

0

0.02

0.04

0.06

0.08

Degree of deformation

Ave

rage

err

or

SCTPS RPMRPM LNSGMMREGCPDOurs

0 0.1 0.2 0.3 0.4 0.50.04

0

0.04

0.08

0.12

Occlusion ratio

Ave

rage

err

or

SCTPS RPMRPM LNSGMMREGCPDOurs

0.02 0.035 0.05 0.065 0.080.04

0.02

0

0.02

0.04

0.06

0.08

Degree of deformation

Ave

rage

err

or

SCTPS RPMRPM LNSGMMREGCPDOurs

0 0.1 0.2 0.3 0.4 0.5

0

0.04

0.08

0.12

0.16

Occlusion ratio

Ave

rage

err

or

SCTPS RPMRPM LNSGMMREGCPDOurs

(c)

Fig. 3: Point set registration on 2D non-rigid shapes. (a) Results of our method on the fish (top) and Chinese character (bottom) shapes [3], [28],with deformation and occlusion presented in every two rows. The goal is to align the model point set (blue pluses) onto the target point set (redcircles). For each group of experiments, the upper figure is the model and target point sets, and the lower figure is the registration result. Fromleft to right, increasing degree of degradation. (b) Quantitative evaluations of our method on the corresponding datasets with different degrees ofdegradation. Each curve is generated based on 100 trials. (c) Comparisons of the registration performance of our method with shape context (SC)[12], TPS-RPM [3], RPM-LNS [28], GMMREG [26] and CPD [24] on the corresponding datasets. The error bars indicate the registration error meansand standard deviations over 100 trials.

performance as the degree of deformation increases. In theocclusion test results (e.g., 2nd and 4th rows), we observethat our method shows much more robustness compared withthe other five algorithms. It is not surprised that our algorithmcan achieve the best performance, since shape context andRPM-LNS do not consider the robust issue during estimatingthe transformation, while TPS-RPM, GMMREG and CPD donot consider using local shape features to help establish pointcorrespondences. Note that the local shape features come fromthe point sets themselves.

More experiments on rotation, noise and outliers are alsoperformed on the two shape models, as shown in Fig. 4.From the results, we again see that our method is able togenerate good alignment when the degradation is moderate,and the registration performance degrades gradually and is stillacceptable as the amount of degradation increases. Note thatour method is not affected by rotation which is not surprising

because we use the rotation invariant shape context as thefeature descriptor.

In conclusion, our method is efficient for most non-rigidpoint set registration problems with moderate, and in somecases severe, distortions. It can also be used to provide agood initial alignment for more complicated problem-specificregistration algorithms.

B. Results on Non-Rigid Sparse Image Feature Correspon-dence

In this section, we perform experiments on real images, andtest the performance of our method for sparse image featurecorrespondence. These images contain deformable objects andconsequently the underlying relationships between the imagesare non-rigid.

Fig. 5 contains a newspaper with different amounts ofspatial warps. We aim to establish correspondences between

Page 10: IEEE TRANSACTIONS ON SIGNAL PROCESSING 1 Robust L E ...ayuille/Pubs15/JiayiMaRobustL2E.pdf · features to obtain a higher level of image alignment (e.g., object/scene level). Here

IEEE TRANSACTIONS ON SIGNAL PROCESSING 10

Fig. 4: From top to bottom, results on rotation, noise and outliers presented in every two rows. For each group of experiments, the upper figure isthe data, and the lower figure is the registration result. From left to right, increasing degree of degradation.

sparse image features in each image pair. In our evaluation,we first extract SIFT [34] feature points in each input im-age, and estimate the initial correspondences based on thecorresponding SIFT descriptors. Our goal is then to reject theoutliers contained in the initial correspondences and, at thesame time, to keep as many inliers as possible. Performanceis characterized by precision and recall.

The results of our method are presented in Fig. 5. The inlierand true positive, false positive, true negative, false negativeare defined within some tolerance radius on human annotated,per-pixel ground truth. For the leftmost pair, the deformationof the newspaper is relatively slight. There are 466 initialcorrespondences with 95 outliers, and the inlier percentage isabout 79.61%. After using our method to reject outliers, 370out of the 371 inliers are preserved, and simultaneously allthe 95 outliers are rejected. The precision-recall pair is about(100.00%, 99.73%). On the rightmost pair, the deformationis relatively large and the inlier percentage in the initialcorrespondences is only about 45.71%. In this case, ourmethod still obtains a good precision-recall pair (100.00%,98.96%). Note that there are still a few false positives andfalse negatives in the results since we could not preciselyestimate the true warp functions between the image pairs inthis framework. The average run time of our method on theseimage pairs is about 0.4 seconds on an Intel Core 2.5 GHzPC with Matlab code.

In addition, we also compared our method to three state-of-

TABLE I: Performance comparison on the image pairs in Fig. 5. Thevalues in the first row are the inlier percentages (%), and the pairs arethe precision-recall pairs (%).

Inlier pct. 79.61 56.57 51.84 45.71

RANSAC [18] (100.00, 97.12) (97.63, 91.56) (92.26, 93.46) (99.35, 92.22)ICF [53] (96.05, 98.38) (83.95, 86.73) (80.43, 95.48) (75.42, 92.71)VFC [49] (100.00, 97.04) (98.59, 99.53) (98.09, 99.35) (98.94, 97.91)Ours (sparse) (100.00, 99.73) (99.06, 99.53) (98.09, 99.35) (100.00, 98.96)Ours (dense) (100.00, 99.73) (100.00, 99.53) (99.36, 100.00) (100.00, 98.96)

the-art methods, such as random sample consensus (RANSAC)[18], identifying point correspondences by correspondencefunction (ICF) [53] and vector field consensus (VFC) [49],[54], [55]. We choose the affine model in RANSAC since re-cent work [56] justifies a simple RANSAC-driven deformableregistration technique with an affine model that is at least asaccurate as other methods based on the optimization of fullydeformable models. The ICF uses support vector regressionto learn a correspondence function pair which maps pointsin one image to their corresponding points in another, andthen reject outliers by the estimated correspondence functions.While the VFC converts the outlier rejection problem into arobust vector field interpolation problem, and interpolates asmooth field to fit the potential inliers as well as estimates aconsensus inlier set. The results are shown in Table I. We seethat all the four algorithms work well when the deformationcontained in the image pair is relatively slight. As the amount

Page 11: IEEE TRANSACTIONS ON SIGNAL PROCESSING 1 Robust L E ...ayuille/Pubs15/JiayiMaRobustL2E.pdf · features to obtain a higher level of image alignment (e.g., object/scene level). Here

IEEE TRANSACTIONS ON SIGNAL PROCESSING 11

Fig. 5: Results of image feature correspondence on 2D image pairs of resolution 384 × 256 with deformable objects. From left to right, increasingdegree of deformation. The inlier percentages in the initial correspondences are 79.61%, 56.57%, 51.84% and 45.71% respectively, and thecorresponding precision-recall pairs are (100.00%, 99.73%), (99.06%, 99.53%), (99.09%, 99.35%) and (100.00%, 98.96%) respectively. The linesindicate matching results (blue = true positive, green = false negative, red = false positive). Best viewed in color.

of deformation increases, the performance of RANSAC andICF begin to degenerate, especially the ICF method. But VFCand our method seem to be relatively unaffected even whenthe number of outliers exceeds the number of inliers. Still,our method gains slightly better results compared to VFC.Besides, we also test our dense correspondence method (i.e.,Algorithm 3) on this task, as shown in the last row. Morespecifically, we first estimate a dense flow for each pixel in theimage by using Algorithm 3, and then check the correctnessof each sparse correspondence according to the flow. Wesee that the results are further improved by using our densecorrespondence method.

Our next experiment involves feature point matching on3D surfaces. We adopt MeshDOG and MeshHOG [57] asthe feature point detector and descriptor to determine theinitial correspondences. For the dataset, we use a surfacecorrespondence benchmark1 [58], which contains meshes rep-resenting people and animals in a variety of poses. Here weselect four classes and each class four meshes with largelevel of deformations, as shown in Fig. 6. The ground truthcorrespondences are supplied by the dataset.

Fig. 6 presents the matching results of our method. We seethat our method in general can obtain good performance, bothin precision and recall. However, if there exists a body partwith large degree of deformation and simultaneously few inliercorrespondences, then all the correspondences on that part maybe treated as outliers by our algorithm and hence removed,such as the legs of cat and dog, the hands of centaur andperson. A quantitative comparison of our method with VFC[49] is reported in Table II (we do not compare to RANSACand ICF since they are not applicable in this case). Clearly,our method has a better precision-recall trade-off. The averageprecision-recall pairs on the eight meshes are about (94.45%,86.07%) and (95.98%, 94.33%) for VFC and our method.

C. Results on Non-Rigid Image Registration

In this section, we test our non-rigid dense correspondencemethod on a set of challenging image pairs with sharedcontent, which includes image transformations such as view-point change, scale change, rotation, as well as non-rigiddeformation. We also compare our method with existing state-of-the-art dense correspondence methods.

1Available at http://www.cs.princeton.edu/∼vk/projects/CorrsBlended/docdata.php.

Fig. 6: Results on 3D surfaces of deformable objects. From left toright, top to bottom: Cat1, Cat2, Dog1, Dog2, Centaur1, Centaur2,Person1, Person2. The inlier percentages in the initial correspon-dences are 93.38%, 93.03%, 80.57%, 82.46%, 77.48%, 76.92%, 51.98%and 27.40% respectively, and the corresponding precision-recall pairsare (99.43%, 95.08%), (99.20%, 94.03%), (96.37%, 99.52%), (96.84%,93.40%), (95.51%, 94.77%), (95.44%, 86.79%), (95.02%, 99.60%) and(86.99%, 91.45%) respectively. For visibility, only 100 randomly selectedcorrespondences are shown. The lines indicate matching results (blue= true positive, green = false negative, red = false positive). Best viewedin color.

Fig. 7 shows a visual comparison between our method andSIFT flow on six pairs of real-world scenes. We aim to alignthe model images (second column) onto the target images (firstcolumn). In the first two rows and middle two rows, we testthe robustness of our method to large scale change and rotationrespectively. The warped model images of SIFT flow and ourmethod are shown in the third and fourth columns. Clearly,SIFT flow will fail in these cases since in the algorithm

Page 12: IEEE TRANSACTIONS ON SIGNAL PROCESSING 1 Robust L E ...ayuille/Pubs15/JiayiMaRobustL2E.pdf · features to obtain a higher level of image alignment (e.g., object/scene level). Here

IEEE TRANSACTIONS ON SIGNAL PROCESSING 12

TABLE II: Performance comparison on the surface pairs in Fig. 6. The values in the first row are the inlier percentages (%), and the pairs are theprecision-recall pairs (%). From left to right: results of Cat1, Cat2, Dog1, Dog2, Centaur1, Centaur2, Person1, Person2.

Inlier pct. 93.38 93.03 80.57 82.46 77.48 76.92 51.98 27.40

VFC [49] (99.76, 76.10) (99.39, 93.46) (97.17, 92.41) (99.78, 74.53) (100.00, 82.17) (99.76, 78.87) (92.88, 99.60) (66.87, 91.45)Ours (99.43, 95.08) (99.20, 94.03) (96.37, 99.52) (96.84, 93.40) (95.51, 94.77) (95.44, 86.79) (95.02, 99.60) (86.99, 91.45)

(a) (b) (c) (d)

Fig. 7: Qualitative comparison of matches on real-world scenes. Thegoal is to align the model images (b) onto the target images (a); (c) and(d) show the matching results (i.e., warped model images) of SIFT flow[4] and our dense correspondence method respectively. The upper tworows test the scale change, the middle two rows test the rotation, andthe bottom two rows test the non-rigid deformation.

the dense SIFT features in the model and target images arecomputed with the same scale and orientation. However, dueto the use of the scale and orientation information of sparsefeature points for computing the dense feature, our method isnot influenced by the rotation and scale change. It can producealmost perfect alignments under these image transformations.We also test the performance of our method in case ofnon-rigid deformation, as shown in the bottom two rows inFig. 7. Again, our method displays its robustness and is ableto generate good alignment even under extreme deformation(e.g., the last row). In contrast, the SIFT flow algorithm worksquite well when the deformation is moderate (e.g., the next-to-last row), but fails in the case of extreme deformation (e.g.,the last row).

Next, we give a quantitative comparison of our method withthree existing state-of-the-art dense correspondence methods.The first method is implemented based on sparse SIFT featurecorrespondence, which has been a standard approach to imageregistration of the same scene. More specifically, we firstexact a set of sparse SIFT feature correspondence from the

image pair, and then utilize our robust L2E method (i.e.,Algorithm 1) to compute a mapping function/transformationfrom the sparse correspondence, finally we transform themodel image to the target image by means of the mappingfunction. The second method is the SIFT flow method whichmatches dense sampled SIFT features. The third method isthe NRDC method [38] which is based on matching imagepatches [39]. For the SIFT flow and NRDC algorithms, weimplement them based on the publicly available codes.

Since ground truth data for real scenes of this kind isscarce, we turn to a standard dataset by Mikolajczyk etal. [59] that has been used for evaluating sparse featurebased correspondence algorithms. The dataset contains im-age pairs either of planar scenes or taken by camera ina fixed position during acquisition. The images, therefore,always obey homography. The ground truth homographiesare supplied by the dataset. The dataset contains six imagetransformations including rotation, scale change, viewpointchange, image blur, JPEG compression, and illumination. Wefocus our comparison on large geometric deformations, i.e.,the first three transformations, and picked the related twosubsets named “zoom + rotation” and “viewpoint”, as shownin Fig. 8. In addition, we also generate an image pair of non-rigid deformation for each subset. More specifically, in Fig. 8,we deform the model image in the second row in each subsetby using the Moving Least Squares algorithm [60], [61], andthen align the deformed model image to the target image. Theground truth correspondence can be obtained based on thehomography and non-rigid transformation.

The results are shown in Fig. 8, in which we highlightonly regions of correspondences that fall within a radius of5 pixels from the ground-truth. The detail matching perfor-mance comparisons between the four methods are given inthe last column. From the results, we see that the sparsecorrespondence based method works quite well in the case ofhomography, but degrades seriously after we add the non-rigiddeformations. The SIFT flow method completely fails due tothe scale change or rotation. The image patch matching basedmethod method NRDC can obtain satisfied results in mostcases, although the captured correspondences typically are notso many. Besides, it also degrades badly in extreme cases,e.g., in the last row of Fig. 8 which contains extreme scalechange, rotation, as well as non-rigid deformation. In contrast,our dense correspondence method is much more robust to allthese degenerations, and in general able to capture much largerand more reliable matches than the other methods.

VII. CONCLUSION

In this paper, we have presented a new approach for non-rigid registration. A key characteristic of our approach is the

Page 13: IEEE TRANSACTIONS ON SIGNAL PROCESSING 1 Robust L E ...ayuille/Pubs15/JiayiMaRobustL2E.pdf · features to obtain a higher level of image alignment (e.g., object/scene level). Here

IEEE TRANSACTIONS ON SIGNAL PROCESSING 13

0 3 6 9 12 150.2

0.4

0.6

0.8

1

Distance from ground truth match (pixel)

Per

cent

age

of c

orre

ct p

ixel

s

SparseSIFTflowNRDCOurs

0 3 6 9 12 150

0.2

0.4

0.6

0.8

1

Distance from ground truth match (pixel)

Per

cent

age

of c

orre

ct p

ixel

s

SparseSIFTflowNRDCOurs

0 3 6 9 12 150

0.2

0.4

0.6

0.8

1

Distance from ground truth match (pixel)

Per

cent

age

of c

orre

ct p

ixel

s

SparseSIFTflowNRDCOurs

0 3 6 9 12 150

0.2

0.4

0.6

0.8

1

Distance from ground truth match (pixel)

Per

cent

age

of c

orre

ct p

ixel

s

SparseSIFTflowNRDCOurs

0 3 6 9 12 150

0.2

0.4

0.6

0.8

1

Distance from ground truth match (pixel)

Per

cent

age

of c

orre

ct p

ixel

s

SparseSIFTflowNRDCOurs

0 3 6 9 12 150

0.2

0.4

0.6

0.8

1

Distance from ground truth match (pixel)

Per

cent

age

of c

orre

ct p

ixel

s

SparseSIFTflowNRDCOurs

(a) (b) (c) (d) (e) (f) (g)

Fig. 8: Correspondence evaluation on two examples from the dataset of Mikolajczyk et al. [59]. The goal is to align the model images (b) onto thetarget images (a); (c), (d), (e) and (f) show the matching results (i.e., warped model images) of sparse correspondence based method, SIFT flow[4], NRDC [38], and our dense correspondence method, we highlight only regions of matches that fall within a radius of 5 pixels from the ground-truth, and the black regions denote that there are no ground truth correspondences; (g) shows the quantitative comparison of the four methods.The upper group of images contains view point change and non-rigid deformation; the bottom group contains scale change, rotation, as well asnon-rigid deformation.

estimation of transformation from correspondences based on arobust estimator named L2E. The computational complexityof estimation of transformation is linear in the scale ofcorrespondences. We applied our robust method to sparsecorrespondence such as non-rigid point set registration andsparse image feature correspondence, and dense correspon-dence such as non-rigid image registration. Experiments onpublic datasets for non-rigid point registration, real imagesfor sparse image feature correspondence and non-rigid imageregistration demonstrate that our approach yields results con-sistently outperform those of state-of-the-art methods such asCPD and SIFT flow when there are significant deformations,occlusions, rotations, and/or scale changes in the data.

ACKNOWLEDGMENT

The authors gratefully acknowledge the financial supportfrom the National Science Foundation (No. 0917141) andChina Scholarship Council (No. 201206160008). Z. Tu issupported by NSF CAREER award IIS-0844566 and NSFaward IIS-1216528.

REFERENCES

[1] L. G. Brown, “A Survey of Image Registration Techniques,” ACMComputing Surveys, vol. 24, no. 4, pp. 325–376, 1992.

[2] P. J. Besl and N. D. McKay, “A Method for Registration of 3-DShapes,” IEEE Transactions on Pattern Analysis and Machine Intelli-gence, vol. 14, no. 2, pp. 239–256, 1992.

[3] H. Chui and A. Rangarajan, “A new point matching algorithm for non-rigid registration,” Computer Vision and Image Understanding, vol. 89,pp. 114–141, 2003.

[4] C. Liu, J. Yuen, and A. Torralba, “SIFT Flow: Dense CorrespondenceAcross Scenes and Its Applications,” IEEE Transactions on PatternAnalysis and Machine Intelligence, vol. 33, no. 5, pp. 978–994, 2011.

[5] D. Scharstein and R. Szeliski, “A taxonomy and evaluation of densetwo-frame stereo correspondence algorithms,” International Journal ofComputer Vision, vol. 47, no. 1, pp. 7–42, 2002.

[6] B. D. Lucas and T. Kanade, “An Iterative Image Registration Techniquewith An Application to Stereo Vision,” in Proceedings of the Interna-tional Joint Conference on Artificial Intelligence, 1981, pp. 674–679.

[7] B. K. P. Horn and B. G. Schunck, “Determining optical flow,” Artificialintelligence, vol. 17, no. 1, pp. 185–203, 1981.

[8] P. F. Felzenszwalb and D. P. Huttenlocher, “Pictorial Structures forObject Recognition,” International Journal of Computer Vision, vol. 61,no. 1, pp. 55–79, 2005.

[9] A. W. Fitzgibbon, “Robust Registration of 2D and 3D Point Sets,” Imageand Vision Computing, vol. 21, pp. 1145–1153, 2003.

Page 14: IEEE TRANSACTIONS ON SIGNAL PROCESSING 1 Robust L E ...ayuille/Pubs15/JiayiMaRobustL2E.pdf · features to obtain a higher level of image alignment (e.g., object/scene level). Here

IEEE TRANSACTIONS ON SIGNAL PROCESSING 14

[10] S. Rusinkiewicz and M. Levoy, “Efficient Variants of the ICP Algorith-m,” in 3-D Digital Imaging and Modeling, 2001, pp. 145–152.

[11] B. Luo and E. R. Hancock, “Structural Graph Matching Using the EMAlgorithm and Singular Value Decomposition,” IEEE Transactions onPattern Analysis and Machine Intelligence, vol. 23, no. 10, pp. 1120–1136, 2001.

[12] S. Belongie, J. Malik, and J. Puzicha, “Shape Matching and ObjectRecognition Using Shape Contexts,” IEEE Transactions on PatternAnalysis and Machine Intelligence, vol. 24, no. 24, pp. 509–522, 2002.

[13] D. W. Scott, “Parametric Statistical Modeling by Minimum IntegratedSquare Error,” Technometrics, vol. 43, no. 3, pp. 274–285, 2001.

[14] A. Basu, I. R. Harris, N. L. Hjort, and M. C. Jones, “Robust and EfficientEstimation by Minimising A Density Power Divergence,” Biometrika,vol. 85, no. 3, pp. 549–559, 1998.

[15] N. Aronszajn, “Theory of Reproducing Kernels,” Transactions of theAmerican Mathematical Society, vol. 68, no. 3, pp. 337–404, 1950.

[16] B. Zitova and J. Flusser, “Image Registration Methods: A Survey,” Imageand Vision Computing, vol. 21, pp. 977–1000, 2003.

[17] J. Ma, J. Zhao, J. Tian, Z. Tu, and A. Yuille, “Robust estimation ofnonrigid transformation for point set registration,” in IEEE conferenceon Computer Vision and Pattern Recognition, 2013, pp. 2147–2154.

[18] M. A. Fischler and R. C. Bolles, “Random Sample Consensus: AParadigm for Model Fitting with Application to Image Analysis andAutomated Cartography,” Communications of the ACM, vol. 24, no. 6,pp. 381–395, 1981.

[19] P. H. S. Torr and A. Zisserman, “MLESAC: A New Robust Estimatorwith Application to Estimating Image Geometry,” Computer Vision andImage Understanding, vol. 78, no. 1, pp. 138–156, 2000.

[20] O. Chum, J. Matas, and J. Kittler, “Locally Optimized RANSAC,” inProceedings of Pattern Recognition Symposium of the German Associ-ation for Pattern Recognition (DAGM), 2003, pp. 236–243.

[21] O. Chum and J. Matas, “Matching with PROSAC - Progressive SampleConsensus,” in Proceedings of IEEE conference on Computer Vision andPattern Recognition, 2005, pp. 220–226.

[22] A. L. Yuille, “Generalized Deformable Models, Statistical Physics, andMatching Problems,” Neural Computation, vol. 2, no. 1, pp. 1–24, 1990.

[23] A. Rangarajan, H. Chui, and F. L. Bookstein, “The Softassign ProcrustesMatching Algorithm,” in Information Processing in Medical Imaging.Springer, 1997, pp. 29–42.

[24] A. Myronenko and X. Song, “Point Set Registration: Coherent PointDrift,” IEEE Transactions on Pattern Analysis and Machine Intelligence,vol. 32, no. 12, pp. 2262–2275, 2010.

[25] Y. Tsin and T. Kanade, “A Correlation-Based Approach to Robust PointSet Registration,” in Proceedings of European Conference on ComputerVision, 2004, pp. 558–569.

[26] B. Jian and B. C. Vemuri, “Robust Point Set Registration Using GaussianMixture Models,” IEEE Transactions on Pattern Analysis and MachineIntelligence, vol. 33, no. 8, pp. 1633–1645, 2011.

[27] ——, “A robust algorithm for point set registration using mixture ofgaussians,” in IEEE International Conference on Computer Vision, 2005,pp. 1246–1251.

[28] Y. Zheng and D. Doermann, “Robust Point Matching for NonrigidShapes by Preserving Local Neighborhood Structures,” IEEE Transac-tions on Pattern Analysis and Machine Intelligence, vol. 28, no. 4, pp.643–649, 2006.

[29] J.-H. Lee and C.-H. Won, “Topology Preserving Relaxation Labelingfor Nonrigid Point Matching,” IEEE Transactions on Pattern Analysisand Machine Intelligence, vol. 33, no. 2, pp. 427–432, 2011.

[30] X. Huang, N. Paragios, and D. N. Metaxas, “Shape Registration inImplicit Spaces Using Information Theory and Free Form Deformation-s,” IEEE Transactions on Pattern Analysis and Machine Intelligence,vol. 28, no. 8, pp. 1303–1318, 2006.

[31] R. Horaud, F. Forbes, M. Yguel, G. Dewaele, and J. Zhang, “Rigid andArticulated Point Registration with Expectation Conditional Maximiza-tion,” IEEE Transactions on Pattern Analysis and Machine Intelligence,vol. 33, no. 3, pp. 587–602, 2011.

[32] L. Torresani, V. Kolmogorov, and C. Rother, “Feature Correspondencevia Graph Matching: Models and Global Optimization,” in Proceedingsof European Conference on Computer Vision, 2008, pp. 596–609.

[33] W. E. L. Grimson, “Computational Experiments with A Feature BasedStereo Algorithm,” IEEE Transactions on Pattern Analysis and MachineIntelligence, no. 1, pp. 17–34, 1985.

[34] D. Lowe, “Distinctive Image Features from Scale-Invariant Keypoints,”International Journal of Computer Vision, vol. 60, no. 2, pp. 91–110,2004.

[35] J. Matas, O. Chum, M. Urban, and T. Pajdla, “Robust Wide BaselineStereo from Maximally Stable Extremal Regions,” in British MachineVision Conference, vol. 1, 2002, pp. 384–393.

[36] R. Szeliski, “Image Alignment and Stitching: A Tutorial,” Foundationsand Trends in Computer Graphics and Vision, vol. 2, no. 1, pp. 1–104,2006.

[37] A. Sotiras and N. Paragios, “Deformable Image Registration: A Survey,”2012.

[38] Y. HaCohen, E. Shechtman, D. B. Goldman, and D. Lischinski, “Non-Rigid Dense Correspondence with Applications for Image Enhancemen-t,” ACM Transactions on Graphics, vol. 30, no. 4, p. 70, 2011.

[39] C. Barnes, E. Shechtman, D. B. Goldman, and A. Finkelstein, “TheGeneralized Patchmatch Correspondence Algorithm,” in Proceedings ofEuropean Conference on Computer Vision. Springer, 2010, pp. 29–43.

[40] J. Ma, J. Zhao, Y. Ma, and J. Tian, “Non-rigid visible and infrared faceregistration via regularized gaussian fields criterion,” Pattern Recogni-tion, vol. 48, no. 3, pp. 772–784, 2015.

[41] B. C. Vemuri, J. Liu, and J. L. Marroquın, “Robust multimodal im-age registration using local frequency representations,” in InformationProcessing in Medical Imaging, 2001, pp. 176–182.

[42] J. Liu, B. C. Vemuri, and J. L. Marroquın, “Local Frequency Repre-sentations for Robust Multimodal Image Registration,” IEEE Trans. onMedical Imaging, vol. 21, no. 5, pp. 462–469, 2002.

[43] P. J. Huber, Robust Statistics. New York: John Wiley & Sons, 1981.[44] A. L. Yuille and N. M. Grzywacz, “A Mathematical Analysis of the

Motion Coherence Theory,” International Journal of Computer Vision,vol. 3, no. 2, pp. 155–175, 1989.

[45] C. A. Micchelli and M. Pontil, “On Learning Vector-Valued Functions,”Neural Computation, vol. 17, no. 1, pp. 177–204, 2005.

[46] G. Wahba, Spline Models for Observational Data. SIAM, Philadelphia,PA, 1990.

[47] J. Ma, J. Zhao, Y. Zhou, and J. Tian, “Mismatch Removal via CoherentSpatial Mapping,” in Proceedings of International Conference on ImageProcessing, 2012, pp. 1–4.

[48] J. Chen, J. Ma, C. Yang, and J. Tian, “Mismatch removal via coherentspatial relations,” Journal of Electronic Imaging, vol. 23, no. 4, pp.043 012–043 012, 2014.

[49] J. Zhao, J. Ma, J. Tian, J. Ma, and D. Zhang, “A Robust Methodfor Vector Field Learning with Application to Mismatch Removing,”in Proceedings of IEEE conference on Computer Vision and PatternRecognition, 2011, pp. 2977–2984.

[50] R. Rifkin, G. Yeo, and T. Poggio, “Regularized Least-Squares Clas-sification,” in Advances in Learning Theory: Methods, Model andApplications, 2003.

[51] J. Ma, J. Zhao, J. Tian, X. Bai, and Z. Tu, “Regularized vector fieldlearning with sparse approximation for mismatch removal,” PatternRecognition, vol. 46, no. 12, pp. 3519–3532, 2013.

[52] A. E. Johnson and M. Hebert, “Using Spin-Images for Efficient ObjectRecognition in Cluttered 3-D Scenes,” IEEE Transactions on PatternAnalysis and Machine Intelligence, vol. 21, no. 5, pp. 433–449, 1999.

[53] X. Li and Z. Hu, “Rejecting Mismatches by Correspondence Function,”International Journal of Computer Vision, vol. 89, no. 1, pp. 1–17, 2010.

[54] J. Ma, J. Zhao, J. Tian, A. L. Yuille, and Z. Tu, “Robust point matchingvia vector field consensus,” IEEE Transactions on Image Processing,vol. 23, no. 4, pp. 1706–1721, 2014.

[55] J. Ma, Y. Ma, J. Zhao, and J. Tian, “Image feature matching viaprogressive vector field consensus,” IEEE Signal Processing Letters,vol. 22, no. 6, pp. 767–771, 2015.

[56] Q.-H. Tran, T.-J. Chin, G. Carneiro, M. S. Brown, and D. Suter, “Indefence of RANSAC for outlier rejection in deformable registration,” inECCV. Springer, 2012, pp. 274–287.

[57] A. Zaharescu, E. Boyer, K. Varanasi, and R. Horaud, “Surface FeatureDetection and Description with Applications to Mesh Matching,” inProceedings of IEEE conference on Computer Vision and PatternRecognition, 2009, pp. 373–380.

[58] V. G. Kim, Y. Lipman, and T. Funkhouser, “Blended intrinsic maps,”ACM Transactions on Graphics, vol. 30, no. 4, p. 79, 2011.

[59] K. Mikolajczyk, T. Tuytelaars, C. Schmid, A. Zisserman, J. Matas,F. Schaffalitzky, T. Kadir, and L. van Gool, “A Comparison of AffineRegion Detectors,” International Journal of Computer Vision, vol. 65,no. 1, pp. 43–72, 2005.

[60] S. Schaefer, T. McPhail, and J. Warren, “Image Deformation UsingMoving Least Squares,” ACM Transactions on Graphics, vol. 25, no. 3,pp. 533–540, 2006.

[61] D. Levin, “The Approximation Power of Moving Least-Squares,” Math-ematics of Computation, vol. 67, no. 224, pp. 1517–1531, 1998.

Page 15: IEEE TRANSACTIONS ON SIGNAL PROCESSING 1 Robust L E ...ayuille/Pubs15/JiayiMaRobustL2E.pdf · features to obtain a higher level of image alignment (e.g., object/scene level). Here

IEEE TRANSACTIONS ON SIGNAL PROCESSING 15

Jiayi Ma received the B.S. degree from the Depart-ment of Mathematics, and the Ph.D. Degree fromthe School of Automation, Huazhong University ofScience and Technology, Wuhan, China, in 2008 and2014, respectively. From 2012 to 2013, he was withthe Department of Statistics, University of Californiaat Los Angeles. He is now a Post-Doctoral with theElectronic Information School, Wuhan University.His current research interests include in the areasof computer vision, machine learning, and patternrecognition.

Weichao Qiu received the B.E. and M.S. degreefrom Huazhong University of Science and Technol-ogy, Wuhan, China in 2011 and 2014 respectively.He is currently a Ph.D. student of the Statistics de-partment, UCLA. His research interest is computervision.

Ji Zhao received the B.S. degree in automation fromNanjing University of Posts and Telecommunicationin 2005. He received the Ph.D. degree in controlscience and engineering from HUST in 2012. Since2012, he is a postdoctoral research associate at theRobotics Institute, Carnegie Mellon University. Hisresearch interests include image classification, imagesegmentation and kernel-based learning.

Yong Ma graduated from the Department of Au-tomatic Control, Beijing Institute of Technology,Beijing, China, in 1997. He received the Ph.D.degree from the Huazhong University of Science andTechnology (HUST), Wuhan, China, in 2003. Hisgeneral field of research is in signal and systems.His current research projects include remote sensingof the Lidar and infrared, as well as Infrared imageprocessing, pattern recognition, interface circuits tosensors and actuators. Between 2004 and 2006, hewas a Lecturer at the University of the West of

England, Bristol, U.K. Between 2006 and 2014, he was with the WuhanNational Laboratory for Optoelectronics, HUST, Wuhan, where he was aProfessor of electronics. He is now a Professor with the Electronic InformationSchool, Wuhan University.

Alan L. Yuille received his B.A. in mathematicsfrom the University of Cambridge in 1976, and com-pleted his Ph.D. in theoretical physics at Cambridgein 1980 studying under Stephen Hawking. Followingthis, he held a postdoc position with the PhysicsDepartment, University of Texas at Austin, and theInstitute for Theoretical Physics, Santa Barbara. Hethen joined the Artificial Intelligence Laboratory atMIT (1982-1986), and followed this with a facultyposition in the Division of Applied Sciences at Har-vard (1986-1995), rising to the position of associate

professor. From 1995-2002 Alan worked as a senior scientist at the Smith-Kettlewell Eye Research Institute in San Francisco. In 2002 he accepted aposition as full professor in the Department of Statistics at the University ofCalifornia, Los Angeles. He has over two hundred peer-reviewed publicationsin vision, neural networks, and physics, and has co-authored two books: DataFusion for Sensory Information Processing Systems (with J. J. Clark) andTwo- and Three- Dimensional Patterns of the Face (with P. W. Hallinan, G.G. Gordon, P. J. Giblin and D. B. Mumford); he also co-edited the bookActive Vision (with A. Blake). He has won several academic prizes and is aFellow of IEEE.

Zhuowen Tu is an assistant professor in the De-partment of Cognitive Science, and the Departmentof Computer Science and Engineering, University ofCalifornia, San Diego. Before joining UCSD, he wasan assistant professor at UCLA. Between 2011 and2013, he took a leave to work at Microsoft ResearchAsia. He received his PhD from the Ohio StateUniversity and his ME from Tsinghua University.Zhuowen Tu received NSF CAREER award in 2009and was awarded with David Marr prize in 2003.