14
1426 IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 29, 2020 Low-Rank Quaternion Approximation for Color Image Processing Yongyong Chen , Xiaolin Xiao , and Yicong Zhou , Senior Member, IEEE Abstract— Low-rank matrix approximation (LRMA)-based methods have made a great success for grayscale image process- ing. When handling color images, LRMA either restores each color channel independently using the monochromatic model or processes the concatenation of three color channels using the concatenation model. However, these two schemes may not make full use of the high correlation among RGB channels. To address this issue, we propose a novel low-rank quaternion approximation (LRQA) model. It contains two major compo- nents: first, instead of modeling a color image pixel as a scalar in conventional sparse representation and LRMA-based methods, the color image is encoded as a pure quaternion matrix, such that the cross-channel correlation of color channels can be well exploited; second, LRQA imposes the low-rank constraint on the constructed quaternion matrix. To better estimate the singular values of the underlying low-rank quaternion matrix from its noisy observation, a general model for LRQA is proposed based on several nonconvex functions. Extensive evaluations for color image denoising and inpainting tasks verify that LRQA achieves better performance over several state-of-the-art sparse representation and LRMA-based methods in terms of both quantitative metrics and visual quality. Index Terms— Low-rank matrix approximation, image denois- ing, image inpainting, quaternion singular value decomposition, nonconvex approximation. I. I NTRODUCTION A S AN emerging mathematical tool, low-rank matrix approximation (LRMA) has been applied in a broad family of real applications, such as image denoising [1]–[3], image inpainting [4], [5], image deblurring [6], background- foreground separation [1], [7], subspace clustering [8], and so on. Due to the fact that the high-dimensional data inherently possess a low-rank structure [9], the goal of LRMA is to exactly and efficiently recover the underlying low-rank matrix from its corrupted observation [1]. Manuscript received March 28, 2019; revised August 8, 2019; accepted September 5, 2019. Date of publication September 19, 2019; date of current version November 7, 2019. This work was funded in part by the Science and Technology Development Fund, Macau SAR (File no. 189/2017/A3), and by the Research Committee at University of Macau under Grants MYRG2016-00123-FST and MYRG2018-00136-FST. The associate editor coordinating the review of this manuscript and approving it for publication was Dr. Nikolaos Mitianoudis. (Corresponding author: Yicong Zhou.) Y. Chen and Y. Zhou are with the Department of Computer and Information Science, University of Macau, Macau 999078, China (e-mail: [email protected]; [email protected]). X. Xiao is with the School of Computer Science and Engineering, South China University of Technology, Guangzhou 510006, China, and also with the Department of Computer and Information Science, University of Macau, Macau 999078, China (e-mail: [email protected]). Digital Object Identifier 10.1109/TIP.2019.2941319 TABLE I SUMMARY OF EXISTING NONCONVEX SURROGATE FUNCTIONS φ(x ,γ) Most of LRMA-based variants follow two lines [2], [10], [11]: matrix factorization and matrix rank minimization. In the first line, it usually factorizes the matrix to be recovered into two or three thin matrices. An intuitive example is the popular nonnegative matrix factorization [12]. It factorizes the original data matrix into two factor matrices with non- negative constraints. And their product can well approximate the original data matrix under the Frobenius-norm fidelity. Recently, the work in [13] proposed a bilinear factor matrix norm minimization for LRMA. The main issue of the matrix factorization-based LRMA algorithms is lack of the rank values in real applications [3], [10], [14]. The second line is usually achieved by different rank approximation regularizers, such as the nuclear norm [15]. Candès and Recht [16] has proven that the nuclear norm is the tightest convex relax- ation of the NP-hard rank minimization function. However, many works [1]–[3], [17] have pointed out that the nuclear norm-based LRMA may yield sub-optimal results of the origi- nal rank minimization. The reason is that each singular value is treated equally, which is contrary to the fact that the large sin- gular values may contain more information of the original data. To better approximate the rank function, many nonconvex sur- rogates have been proposed, such as the Schatten γ -norm [18], weighted nuclear norm [1], log-determinant penalty [19] (sum- marized in Table I) and achieved the promising performance in grayscale image processing. Among them, the most popular one is the weighted nuclear norm (WNNM), which assigns different weights to different singular values by consider- ing the physical significance of them. Following this line, Xie et al. [2] proposed the weighted Schatten γ -norm min- imization (WSNM) by combining WNNM and the Schatten γ -norm for image denoising and background-foreground separation. In [20], the authors proposed a general- ized nonconvex nonsmooth LRMA solver, named itera- tively reweighted nuclear norm algorithm. To promote the applicability of these nonconvex surrogates, Nie et al. [21] 1057-7149 © 2019 IEEE. Personal use is permitted, but republication/redistribution requires IEEE permission. See http://www.ieee.org/publications_standards/publications/rights/index.html for more information.

Low-Rank Quaternion Approximation for Color Image Processing · 2020. 4. 17. · fits color image processing tasks. Using QR, a color image is represented by a pure quaternion matrix

  • Upload
    others

  • View
    5

  • Download
    0

Embed Size (px)

Citation preview

Page 1: Low-Rank Quaternion Approximation for Color Image Processing · 2020. 4. 17. · fits color image processing tasks. Using QR, a color image is represented by a pure quaternion matrix

1426 IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 29, 2020

Low-Rank Quaternion Approximation forColor Image Processing

Yongyong Chen , Xiaolin Xiao , and Yicong Zhou , Senior Member, IEEE

Abstract— Low-rank matrix approximation (LRMA)-basedmethods have made a great success for grayscale image process-ing. When handling color images, LRMA either restores eachcolor channel independently using the monochromatic modelor processes the concatenation of three color channels usingthe concatenation model. However, these two schemes may notmake full use of the high correlation among RGB channels.To address this issue, we propose a novel low-rank quaternionapproximation (LRQA) model. It contains two major compo-nents: first, instead of modeling a color image pixel as a scalarin conventional sparse representation and LRMA-based methods,the color image is encoded as a pure quaternion matrix, suchthat the cross-channel correlation of color channels can bewell exploited; second, LRQA imposes the low-rank constrainton the constructed quaternion matrix. To better estimate thesingular values of the underlying low-rank quaternion matrixfrom its noisy observation, a general model for LRQA is proposedbased on several nonconvex functions. Extensive evaluations forcolor image denoising and inpainting tasks verify that LRQAachieves better performance over several state-of-the-art sparserepresentation and LRMA-based methods in terms of bothquantitative metrics and visual quality.

Index Terms— Low-rank matrix approximation, image denois-ing, image inpainting, quaternion singular value decomposition,nonconvex approximation.

I. INTRODUCTION

AS AN emerging mathematical tool, low-rank matrixapproximation (LRMA) has been applied in a broad

family of real applications, such as image denoising [1]–[3],image inpainting [4], [5], image deblurring [6], background-foreground separation [1], [7], subspace clustering [8], and soon. Due to the fact that the high-dimensional data inherentlypossess a low-rank structure [9], the goal of LRMA is toexactly and efficiently recover the underlying low-rank matrixfrom its corrupted observation [1].

Manuscript received March 28, 2019; revised August 8, 2019; acceptedSeptember 5, 2019. Date of publication September 19, 2019; date of currentversion November 7, 2019. This work was funded in part by the Scienceand Technology Development Fund, Macau SAR (File no. 189/2017/A3),and by the Research Committee at University of Macau under GrantsMYRG2016-00123-FST and MYRG2018-00136-FST. The associate editorcoordinating the review of this manuscript and approving it for publicationwas Dr. Nikolaos Mitianoudis. (Corresponding author: Yicong Zhou.)

Y. Chen and Y. Zhou are with the Department of Computer andInformation Science, University of Macau, Macau 999078, China (e-mail:[email protected]; [email protected]).

X. Xiao is with the School of Computer Science and Engineering, SouthChina University of Technology, Guangzhou 510006, China, and also withthe Department of Computer and Information Science, University of Macau,Macau 999078, China (e-mail: [email protected]).

Digital Object Identifier 10.1109/TIP.2019.2941319

TABLE I

SUMMARY OF EXISTING NONCONVEX SURROGATE FUNCTIONS φ(x, γ )

Most of LRMA-based variants follow two lines [2],[10], [11]: matrix factorization and matrix rank minimization.In the first line, it usually factorizes the matrix to be recoveredinto two or three thin matrices. An intuitive example is thepopular nonnegative matrix factorization [12]. It factorizesthe original data matrix into two factor matrices with non-negative constraints. And their product can well approximatethe original data matrix under the Frobenius-norm fidelity.Recently, the work in [13] proposed a bilinear factor matrixnorm minimization for LRMA. The main issue of the matrixfactorization-based LRMA algorithms is lack of the rankvalues in real applications [3], [10], [14]. The second line isusually achieved by different rank approximation regularizers,such as the nuclear norm [15]. Candès and Recht [16] hasproven that the nuclear norm is the tightest convex relax-ation of the NP-hard rank minimization function. However,many works [1]–[3], [17] have pointed out that the nuclearnorm-based LRMA may yield sub-optimal results of the origi-nal rank minimization. The reason is that each singular value istreated equally, which is contrary to the fact that the large sin-gular values may contain more information of the original data.To better approximate the rank function, many nonconvex sur-rogates have been proposed, such as the Schatten γ -norm [18],weighted nuclear norm [1], log-determinant penalty [19] (sum-marized in Table I) and achieved the promising performancein grayscale image processing. Among them, the most popularone is the weighted nuclear norm (WNNM), which assignsdifferent weights to different singular values by consider-ing the physical significance of them. Following this line,Xie et al. [2] proposed the weighted Schatten γ -norm min-imization (WSNM) by combining WNNM and the Schattenγ -norm for image denoising and background-foregroundseparation. In [20], the authors proposed a general-ized nonconvex nonsmooth LRMA solver, named itera-tively reweighted nuclear norm algorithm. To promote theapplicability of these nonconvex surrogates, Nie et al. [21]

1057-7149 © 2019 IEEE. Personal use is permitted, but republication/redistribution requires IEEE permission.See http://www.ieee.org/publications_standards/publications/rights/index.html for more information.

Page 2: Low-Rank Quaternion Approximation for Color Image Processing · 2020. 4. 17. · fits color image processing tasks. Using QR, a color image is represented by a pure quaternion matrix

CHEN et al.: LOW-RANK QUATERNION APPROXIMATION FOR COLOR IMAGE PROCESSING 1427

proposed a parameter-free nonconvex LRMA model for imageinpainting.

When handling color images using LRMA, we mayencounter two challenges. The first challenge is how tofully exploit the high correlation among RGB channels.An intuitive way is to perform LRMA channel-by-channel andthen combine all results together. That is, LRMA is appliedto each channel independently. This scheme is termed asthe monochromatic model. An obvious shortcoming of themonochromatic model is that the correlation among threechannels are completely ignored, resulting in performancedegradation in color image processing. Another popular wayis to concatenate three RGB channels to make use of thechannel correlation. This scheme is named as the concate-nation model [22]. For example, Xu et al. [23] proposeda multi-channel weighted nuclear norm minimization model.It first vectorizes each similar patch and then concatenates allvectors to form a noisy patch matrix. From the perspectiveof tensor-unfolding [9], the concatenation model-based colorimage processing methods just use one unfolding matrix andignore other two unfolding matrices. In summary, these abovetwo schemes may fail to make full use of the correlationinformation among three color channels.

The second challenge is how to preserve the local detailsof color images under the LRMA framework. In addi-tion to the channel correlation, the nonlocal self-similarity(NSS) [24] prior is another intrinsic characteristic for imageprocessing tasks, such as image denoising, image deblurring.The underlying assumption of NSS is that for an exemplarpatch, there are many similar local patches across a naturalimage [1], [24]. For example, Zhang et al. [25] used principalcomponent analysis on the similar patches group to remove theGaussian noise. Combining the sparse representation and theNSS prior, the work in [26] proposed a nonlocally centralizedsparse representation model for image restoration. Borrowingthe above idea, the advanced version of sparse representation,LRMA as well as NSS was fully considered in the worksof [1], [2].

The above-mentioned methods may cause unsatisfactoryresults for color image processing since they were originallydesigned for grayscale image processing. Recently, the quater-nion representation (QR)-based color image processing meth-ods have attracted much attention due to its rapid developmentin both theory [27]–[29] and applications [22], [30]–[36].Due to the essential characteristic of QR which includesone real part and three imaginary components, it perfectlyfits color image processing tasks. Using QR, a color imageis represented by a pure quaternion matrix instead of threeindependent matrices or a concatenation matrix. This meansthat all color channels are handled simultaneously in thequaternion domain, resulting in the superior power of preserv-ing cross-channel correlations of color images. For example,inspired by the sparse representation [37], Xu et al. [22]developed a vector sparse representation model for colorimage denoising, inpainting, and super-resolution. The workin [33] extended the sparse representation and collabora-tive representation classifications into the quaternion domainand achieved good performance for color face recognition.

In [35], Zhu et al. proposed a quaternion convolution neuralnetwork (QCNN) model to extend CNN from the real domaininto the quaternion domain. It contains a series of basic modelslike quaternion convolution layer. To explore the fundamentalreason of the success of QCNN, the work in [38] used theHamilton product to capture internal latent relations withinfeatures. Following [35], Zeng et al. [34] proposed the quater-nion principal component analysis network for color imageclassification. To preserve the spatial structures of images,Xiao and Zhou [30] proposed a two-dimensional quater-nion sparse principal component analysis method for colorimage recognition. Generally speaking, the main advantageof QR over the real valued methods is the generalizationcapability [22], [30], [38].

Considering the huge success of LRMA and the powerfuldata representation advance of QR, one natural question isthat is it possible and reasonable to extend LRMA into thequaternion domain to handle the three color channels ina holistic manner? Therefore, in this paper, we propose anovel model, i.e., low-rank quaternion approximation (LRQA)with applications to color image denoising and color imageinpainting. Our idea is motivated by the recently proposedLRMA [2], [3], [17] and the quaternion representation-basedmethods in [22], [30], and [33]. However, different fromthese methods in [2], [3], [17], LRQA incorporates thequaternion representation into LRMA to handle three colorchannels as a whole for the better capture of the highcorrelation. A general model for LRQA based on threenonconvex functions is proposed. To derive the closed-formsolution of LRQA, we first compute the complex singularvalue decomposition (SVD) instead of quaternion SVD toreduce the high computation cost and then adopt the differenceof convex method to linearize three nonconvex functions.The main contributions of this paper are summarized asfollows:

• We propose a novel model, namely, low-rank quaternionapproximation (LRQA) for color image processing. Dif-ferent from LRMA-based methods treating color imagepixel as a scalar, LRQA encodes each color image pixelas a pure quaternion number, such that more corre-lation information among RGB channels can be wellpreserved.

• To achieve a better quaternion rank estimation, we pro-pose a general model for LRQA based on three noncon-vex functions and derive a general solver.

• LRQA is further extended to handle the color imagedenoising and inpainting tasks. The experimental resultson both simulated and real color images demonstrate theirsuperiority over several state-of-the-art methods.

The rest of this paper is structured as follows. Section IIbriefly reviews related works on LRMA and quaternion-basedmethods. In Section III, we discuss the basics of quaternionrepresentation. Section IV proposes LRQA in detail and ana-lyzes the general solver for LRQA. Two real applications ofLRQA are described in Section V. We evaluate the perfor-mance of the proposed LRQA in Section VI and concludethis paper in Section VII.

Page 3: Low-Rank Quaternion Approximation for Color Image Processing · 2020. 4. 17. · fits color image processing tasks. Using QR, a color image is represented by a pure quaternion matrix

1428 IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 29, 2020

II. RELATED WORK

In this section, we briefly review the highly related topics:low-rank matrix approximation and quaternion-based methods.

A. Low-Rank Matrix Approximation

Given a real matrix Y ∈ Rm×n , LRMA can be formulatedas the following minimization problem:

minX

1

2‖X − Y‖2

F + λ ∗ rank(X), (1)

where λ is a nonnegative parameter and rank(·) is the rankfunction. In Eq. (1), the first term is the data fidelity termwhich aims to measure the difference between X and Y , whilethe second one is called regularizer, such as the nuclear norm‖X‖∗ = ∑

i σi . A large number of extensions of LRMAhave been presented, including robust principal componentanalysis [39], matrix completion [40], low-rank representa-tion [8], [41], robust matrix sensing [42] and so on. To addressthe essential issue of the nuclear norm, truncated nuclearnorm [4], capped nuclear norm [43], partial sum nuclearnorm [44], weighted nuclear norm [1], and weighted Schattenγ -norm [2] have successively been proposed as summarizedin Table I. In a word, the main idea of these nonconvexapproximations is that no matter how big a singular value is,its contribution for the rank function should be 1.

The existing LRMA-based approaches work well forgrayscale images, but degrade sharply when handling colorimages [23]. The main reasons are that, one the one hand, theseapproaches handle each color channel in a channel-by-channelmanner, which means that each color channel is treated as anindependent grayscale image. On the other hand, they processthe concatenation of three color channels. As discussed in theintroduction part, however, it is still intractable to preserve thecorrelation information among RGB channels.

B. Quaternion-Based Methods

Quaternion-based methods have received increasing atten-tion in various applications, such as face recognition [33], [46],dimension reduction [30], image denoising [22], since it cantreat a color image as a whole, such that the cross-channelcorrelation of color images can be well encoded. Consideringthe fact that an image signal can be sparsely represented bya linear combination of the dictionary atoms, Xu et al. [22]developed a quaternion-based sparse representation model forseveral color image processing tasks. Following this line,Zou et al. [33] proposed two quaternion methods namedquaternion collaborative and sparse representation-based clas-sification models for color face recognition, which aims tolearn the collaborative or sparse coefficients of a testing sam-ple, and defined specific operators to convert the quaternionvariables into real domain.

Essentially, the proposed LRQA is different from [22], [33]since the methods in [22] and [33] belong to theone-dimensional vector sparse representation methods, whileour proposed LRQA belongs to the two-dimensional low-rankmatrix approximation method, which imposes the sparsity onthe singular values of the quaternion matrix.

III. PRELIMINARIES

In this section, we briefly review some mathematical nota-tions and definitions for quaternion number and quaternionrepresentation. In this paper, scalars, vectors, matrices, andtensors are denoted as lowercase letters, boldface lowercaseletters, boldface capital letters, and boldface Euler scriptletters, e.g., x , x, X, and X, respectively. Following [22], [31],a dot above the variable (e.g., x ∈ H) is to denote the variablein quaternion space H.

As a generalization of the real space R and complexspace C, the quaternion space was firstly introduced byHamilton [47]. A quaternion number x includes one realnumber and three imaginary parts, i.e.,

x = x0 + x1i + x2j + x3k, (2)

where xl ∈ R (l = 0, 1, 2, 3) is real number, and i, j andk are three imaginary units. If x0 = 0, x is called the purequaternion.

Several important definitions of norm, properties and oper-ator of quaternion numbers and matrices are introduced asfollowing: Given a quaternion number x and a quaternionmatrix X ∈ Hm×n ,

• The conjugate operator x , the modulus |x |, and theinverse x−1 of x : x = x0 − (x1i + x2j + x3k), |x | =√

x20 + x2

1 + x22 + x2

3 , x−1 = x∗|x|2 ; The conjugate opera-

tor X, the transpose operator XT , the conjugate transposeX�: X = (xi, j ), XT = (x j,i), X� = (X)T = XT = (x j,i);

• The F-norm of X: ‖X‖F = ( m∑i=1

n∑j=1

|xi, j |2) 1

2 .

• X ∈ Hm×m is called unitary quaternion matrix if and onlyif XX� = X�X = Im where Im ∈ Rm×m is a real identitymatrix.

Definition 1 (Cayley-Dickson construction [48] and Com-plex adjoint form [28]): For any quaternion matrix X =X0 + X1i + X2j + X3k ∈ Hm×n , it can be represented byan ordered pair of complex matrices Xa, Xb ∈ Cm×n , i.e.,

X = Xa + Xbj, (3)

where Xa = X0 + X1i, Xb = X2 + X3i. Then the complexadjoint form of X is formulated as

χX =[

Xa Xb

−Xb Xa

]∈ C

2m×2n . (4)

Please refer to [28], [49] for detailed introduction of quater-nion algebra.

In Section IV-B, a third-order tensor X of size m × n × 3is transformed into a quaternion matrix X ∈ H

m×n by settingthree frontal slices of X as the three imaginary parts of X,i.e., X = X(:, :, 1)i + X(:, :, 2)j + X(:, :, 3)k. In Section V,the estimated quaternion matrix X is transformed back to acolor image by extracting the three imaginary parts of X asthree color channels.

IV. LOW-RANK QUATERNION APPROXIMATION

Considering that the existing LRMA-based methods maysuffer from performance degradation of color image process-

Page 4: Low-Rank Quaternion Approximation for Color Image Processing · 2020. 4. 17. · fits color image processing tasks. Using QR, a color image is represented by a pure quaternion matrix

CHEN et al.: LOW-RANK QUATERNION APPROXIMATION FOR COLOR IMAGE PROCESSING 1429

ing since they are initially designed in the real settings forgrayscale image processing, we propose the low-rank quater-nion approximation (LRQA) for the better capture of thecross-channel relationship. We first solve two challenges ofLRQA, and then propose the general framework of LRQA forbetter quaternion rank approximation.

A. LRQA

Given a quaternion matrix Y , the proposed LRQA isdefined as

minX

1

2‖X − Y‖2

F + λ ∗ rank(X), (5)

where rank(X) is the rank function of quaternion matrix X.As for how to construct the quaternion matrix Y, please referto Section V.

Different from LRMA in real domain, LRQA in Eq. (5)poses at least two challenges for color image analysis: thefirst one is how to define the rank of a quaternion matrix;the other one is how to obtain the closed-form solution ofquaternion rank regularized minimization problem. In thefollowing section, we will give the affirmative answer to theseabove two questions. Our solution is inspired by the quaternionsingular value decomposition [28], [29], [32].

Theorem 1 ( [28] Quaternion singular value decomposition(QSVD)): For any quaternion matrix Y ∈ H

m×n of rank r ,there exist a pair of unitary quaternion matrices U ∈ Hm×m

and V ∈ Hn×n such that

Y = U(

�r 00 0

)V�, (6)

where � is the conjugate transpose operator and �r ∈ Rr×r isa real diagonal matrix.

The above QSVD first introduced by Zhang [28]in 1996 also enjoys similar forms and properties of the SVDfor a real matrix. For example, all singular values of anyquaternion matrix are nonnegative and have the decreasingorder property. The bigger singular values also preserve themajor color image information, such as the color informationand the structural details. This observation can be foundin [32]. Based on QSVD, we can give the following definition.

Definition 2 (Quaternion rank): The rank of a quaternionmatrix can be defined as the number of nonzero singularvalues.Until now, we have given the positive answer to the firstchallenge of Eq. (5). Actually, the quaternion rank is the

l0-norm of singular value vector {σi }min(m,n)i=1 . Unfortunately,

the l0-norm is nonconvex. Borrowing the idea of the nuclearnorm in real domain, we consider the following quaternionnuclear norm:

Definition 3 (Quaternion nuclear norm): The quaternionnuclear norm of a quaternion matrix X is defined as the sumof all nonzero singular values, i.e., ‖X‖� = �iσi (X).

Intuitively, the nuclear norm is extended into the quaterniondomain by Definition 3 and enjoys the same form since allsingular values are still real numbers. By using the quaternion

nuclear norm, the nonconvex optimization problem (5) can berelaxed as:

minX

1

2‖Y − X‖2

F + λ‖X‖�. (7)

Here, we call (7) as quaternion nuclear norm minimization(QNNM). The remaining problem is how to obtain the optimalsolution of QNNM (7).

Theorem 2: (Quaternion singular value thresholding(QSVT)): For any λ > 0 and quaternion matrix Y, the closed-form solution of QNNM (7) is:

X∗ = USλ(�)V�, (8)

where Y = U�V� is the QSVD of Y. The soft thresholdingoperator is defined as Sλ(�) = diag(max{σi (Y) − λ, 0}).

The proof of Theorem 2 can be found in the supplemen-tary material Appendix A. Essentially, our quaternion nuclearnorm is the l1-norm of the singular value vector, while thequaternion rank is the l0-norm of the singular value vector.From the above theorem, we can see that QSVT shrinks eachsingular value equally, that is, each singular value is minus thesame threshold. Therefore, the quaternion nuclear norm stillover-penalizes larger singular values as well as the nuclearnorm in real domain since l1-norm is a biased estimator ofthe l0-norm [1], [3]. This observation encourages us to seek amore accurate approximation of quaternion rank.

B. General Model of LRQA

Motivated by the promising performance of nonconvexsurrogates in LRMA, we propose a general model of LRQAas follows:

minX

1

2‖Y − X‖2

F + λ∑

i

φ(σi (X), γ

), (9)

where φ(·) is defined in Table I. γ is a nonnegative parameterrelated to the specific φ(·). In particular, we consider threespecial cases:

‖X‖l,γ =∑

i

(1 − e−σi (X)

γ ); (10)

‖X‖g,γ =∑

i

(1 + γ ) ∗ σi (X)

γ + σi (X); (11)

‖X‖γω,γ =

∑i

ωi ∗ σγi (X); (12)

where ωi is a nonnegative weight scalar which is used tobalance the contribution for quaternion rank of i -th singularvalue. Note that the above three nonconvex rank surrogatesare based on the laplace, geman, and weighted Schatten-γfunctions as defined in Table I, respectively.

Here, we first omit to give the closed-form solution of (9)and design an experiment on synthetic data to illustratethe superiority of the general model in Eq. (9) comparedwith the QNNM model in Eq. (7). We first use Matlabscript randn(m, r, 3) and randn(r, n, 3) to generate two tensorsA ∈ Rm×r×3 and B ∈ Rr×n×3. Then, the low-ranktensor X ∈ Rm×n×3 is as the product of A and B,

Page 5: Low-Rank Quaternion Approximation for Color Image Processing · 2020. 4. 17. · fits color image processing tasks. Using QR, a color image is represented by a pure quaternion matrix

1430 IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 29, 2020

Fig. 1. Comparison of QNNM in Eq. (7) and the general model in Eq. (9). It iseasy to observe that the singular values obtained by QNNM are overshrunksince each singular value is regularized equally. While the singular valuesestimated by the general model is closer to the original singular values.

i.e., X = A ⊗ B1. And the additive Gaussian noise with zeromean and standard variance τ = 5 is added into X to generatethe noisy observation Y, that is, Y = X + τ ∗ randn(m, n, 3).Finally, we generate quaternion matrices X ∈ Hm×n andY ∈ Hm×n . Due to the fact that the distribution of singularvalues can represent the low-rank property [1], [3], [23], [40],we plot the singular values of clean quaternion matrix X,noisy observation Y, and obtained quaternion matrices by hardthresholding operator, models (7) and (9) in Fig. 1. In Fig. 1,x-axis denotes the number of singular values while y-axis rep-resents the corresponding singular values. We can observe thatthe distribution of singular values of noisy observation Y (thered curve) deviates significantly that of clean quaternion matrixX (the green curve) due to the existence of Gaussian noise.Compared with the general model (the black curve), QNNMovershrinks each singular value since each singular value isthresholded with a same amount. While, the singular valuesobtained by the general model is closer to the original singularvalues. Meanwhile, there exits a rapid decline trend of originalsingular values of X, indicating that the quaternion matrix Xstill has the low-rank characteristic. These observations canalso verify the reasonableness of (9).

The remaining question is how to efficiently solvemodel (9). Actually, it is intractable to solve the general modelin Eq. (9) due to the complicated operations of the quaternionalgebra and the nonconvexity of φ(·). Here, we focus on solv-ing model (9) in complex domain due to the striking fact thatcomputing SVD of the complex adjoint matrix is equivalent tocomputing QSVD of the corresponding quaternion matrix andthe lower computation [22]. Based on Definition 1 in Eqs (3)and (4), the quaternion matrix Y can be conveniently andefficiently converted into the equivalent complex matrix χY,then the QSVD of Y can be obtained by

U = colo(U1) + colo(−U2)j, (13)

� = rowo(colo(�′)), (14)

V = colo(V1) + colo(−V2)j, (15)

where χY = U�′V� is the SVD of complex matrix χY; U =[U1; U2] ∈ C2m×2m , V = [V1; V2] ∈ C2n×2n ; rowo(U) andcolo(U) denote to extract the odd rows and odd columns of

1Here, ⊗ is the tensor-tensor product, more details can be found in [50].

Algorithm 1 Solving LRQA in Eq. (9) via DC

complex matrix U, respectively. We first give the followingtheorem and then derive the solution of Eq. (9).

Theorem 3: Suppose Y = U�YV� is the QSVD of Y ∈H

m×n and �Y = diag(σY), then the optional solution ofEq. (9) is X∗ = U�XV�, where �X = diag(σ ∗) and σ ∗is obtained by

σ ∗ = argminσ≥0

1

2‖σ − σY‖2

2 + λ ∗ φ(σ, γ ). (16)

The proof of Theorem 3 can be found in the supplementarymaterial Appendix B. Based on Theorem 3, solving Eq. (9)is transformed into solving Eq. (16) in real domain. How-ever, φ(·) is continuous, concave, smooth, differentiable, andmonotonically increasing on [0,+∞). Following [3], [17],the difference of convex (DC) method [51] is exploited tosolve Eq. (16). The main idea of DC is to iteratively optimizeit by linearizing the concave function φ(·) and preserving thel2-norm term, resulting in the following formula

σ t+1 = argminσ≥0

1

2‖σ − σY‖2

2 + λ ∗ ∂φ(σ t )T σ, (17)

where the superscript t denotes the number of iterations;∂φ(σ t ) is the gradient of φ(·) at σ t . The closed-form solutionof Eq. (17) is

σ t+1 = max{σY − λ ∗ ∂φ(σ t ), 0}. (18)

After several iterations, DC methods can converge to a localoptimal σ ∗. Finally, the optimal solution of Eq. (9) is X∗ =Udiag(σ ∗)V�. The whole procedure of solving Eq. (9) isoutlined in Algorithm 1.

V. LRQA FOR COLOR IMAGE DENOISING

AND INPAINTING

In this section, we apply the proposed LRQA to twocolor image processing tasks: color image denoising andinpainting.

A. LRQA for Color Image Denoising

Image denoising is an ancient yet hot image processing topicand its aim is to remove unwanted noise n and restore theclean image x from the noisy observation y, which can beformulated as

y = x + n, (19)

where n is generally assumed to be additive white Gaussiannoise with zero mean and variance τ 2

n . However, this inverse

Page 6: Low-Rank Quaternion Approximation for Color Image Processing · 2020. 4. 17. · fits color image processing tasks. Using QR, a color image is represented by a pure quaternion matrix

CHEN et al.: LOW-RANK QUATERNION APPROXIMATION FOR COLOR IMAGE PROCESSING 1431

Fig. 2. The flowchart of color image denoising by LRQA.

problem (19) is severely ill-posed without any priors ofimage x . Therefore, regularization for image prior knowledgeis needed to stabilize the inversion process. Here, we considerthe nonlocal self-similarity (NSS) and low-rank property forimage denoising task.

Following the procedure of NSS, we first divide the noisycolor image y into many overlapping color patches of sizem ×m ×3 (for example, 6 ×6 ×3). Instead of modeling colorimage pixel as a scalar in traditional sparse representationand LRMA-based methods [1], [2], we use a pure quaternionnumber to represent a color image pixel. This is also the maindifference with these existing denoising methods [1], [2], [17].For an exemplar patch of size m ×m ×3 at position i denotedby yi ∈ Hm2

(yi = yr i + ygj + ybk, where yr , yg, yb ∈ Rm2

are the corresponding patch vectors of R, G, B channels), itssimilar patch group consists of n nearest neighbor patcheswithin its local searching window (where n includes yi ) andthen is stacked as quaternion column vectors of quaternionmatrix Yi = [yi1, yi2, · · · , yin] ∈ Hm2×n . The abovewhole procedure is depicted in Fig 2. The key assumptionof NSS is that for any exemplar patch yi , there exist manysimilar local patches across a natural image, which impliesthat the constructed matrix Yi is low-rank [1]. Similarly,we assume that our constructed quaternion matrix Yi alsohas the low-rank characteristic in the quaternion domain.Based on this assumption, the clean patch group Xi can beobtained by

minX

1

τ 2n

‖Y − X‖2F +

∑i

φ(σi (X), γ

), (20)

where ‖Y − X‖2F of Eq. (20) is the data fidelity term regu-

larized by the Frobenius-norm, which is consistent with theassumption in (19) because l2-norm is optimal to suppressingGaussian noise. The optimal solution of Eq. (20) can beobtained by Algorithm 1.

Once the clean patch group X is learned by Eq. (20),we transform quaternion matrix X into real matrix X andaggregate all patches together to form the clean color image x .Following [1], [2], [23], we adopt the iterative regularization

Algorithm 2 Color image denoising by LRQA in Eq. (20)

scheme [54] as follows

y(k) = x (k−1) + δ(y − x (k−1)), (21)

where k and δ are the iteration number and the relaxationparameter, respectively. The flowchart of LRQA for colorimage denoising is shown in Fig. 2. The whole procedure ofcolor image denoising is summarized in Algorithm 2.

B. LRQA for Color Image Inpainting

In this section, we further apply the proposed LRQA tocolor image inpainting. Image inpainting targets at filling themissing pixels from the incomplete observation D. As statedin [4], [13], and [16], the real images can be approximatelyrecovered by a low-rank matrix. Therefore, we use the pro-posed LRQA to recover the color image by the followingmodel:

minX

∑i

φ(σi (X), γ

)

s.t . P�(X) = P�(D), (22)

where � is the set of observed entries of D. P�(X) denotesthe same size quaternion matrix as X with P�(X)i, j = Di, j

if (i, j) ∈ � and zero otherwise.To make X separable, we use the variable-splitting tech-

nique [55] and introduce one auxiliary quaternion variable Einto Eq. (22) yielding the following model:

minX,E

∑i

φ(σi (X), γ

)

s.t . X + E = D, P�(E) = 0. (23)

The partial augmented Lagrangian function of Eq. (23) isdefined as

Lμ(X, E; )=∑

i

φ(σi (X), γ

)+ μ

2‖D−X−E+

μ‖2

F , (24)

where μ > 0 is the penalty parameter. is the Lagrangianmultiplier.

Page 7: Low-Rank Quaternion Approximation for Color Image Processing · 2020. 4. 17. · fits color image processing tasks. Using QR, a color image is represented by a pure quaternion matrix

1432 IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 29, 2020

Algorithm 3 Color image inpainting by LRQA in Eq. (22)

Under the complex altering direction method of multipliersframework [56], we can iteratively update all variables asfollows

Xk+1 = argminX

∑i

φ(σi (X), γ

)+ μk

2‖D−X−Ek + k

μk‖2

F ,

(25)

Ek+1 = argminP�(E)=0

‖D − Xk+1 − E + k

μk‖2

F , (26)

k+1 = k + μk(D − Xk+1 − Ek+1), (27)

μk+1 = min{β ∗ μk, μmax}, (28)

where 0 < β < 1+√5

2 . μmax is the maximum value of thepenalty parameter μ. The closed-form solution of Eq. (25) canbe obtained by Algorithm 1. The optimal Ek+1 of Eq. (26) isD− Xk+1 + k

μk . To guarantee the observed entries unchanged,

P�(E) = 0 is held throughout the iteration. The wholeprocedure of LRQA for color image inpainting is summarizedin Algorithm 3, in which the stopping criterion is defined asfollows

‖D − Xk+1 − Ek+1‖2F

‖D‖2F

≤ tol, (29)

where tol > 0 is a pre-defined tolerance.

VI. EXPERIMENTS RESULTS

In this section, we conduct several experiments to verifythe effectiveness of the proposed LRQA including two realcolor image processing tasks: denoising and inpainting. Allexperiments are run in Matlab R2016a on a 64-bit workstationwith a 1.90GHz CPU and 16GB memory.

A. Simulated Noisy Color Image Denoising

Data: We select 12 widely used color images as shownin Fig. 4. To generate simulated noisy color images, the addi-tive white Gaussian noise with zero mean and variance τ 2

nis added to these clean color images. Three noise levelsτ = 10, 30, 50 are considered in our experiments.

Compared methods: We compare the proposed LRQAwith several state-of-the-art SR and LRMA-based denoisingmethods, including LSCD [52], KQSVD [22], NNM [16], andWNNM [53]. For a comprehensive comparison, we imple-ment WNNM with two versions by the concatenation model

and the monochromatic model, denoted as WNNM-1 andWNNM-2, respectively. Note that KQSVD is recently pro-posed color image denoising method based on the vectorsparse representation model and quaternion algebra. LRQAusing the nuclear norm, Laplace function, Geman function, andweighted Schatten norm are denoted as LRQA-1, LRQA-2,LRQA-3, and LRQA-4, respectively.

Parameter setting: In Algorithm 2, when τ = 10, 30, 50,we set patch size to 6×6, 7×7, 8×8, respectively. The numberof similar patches group is set to 70, 90, and 120, respectively.The iteration K is 8, 12, and 14, respectively. In all cases,the iterative relaxation parameter δ is fixed to 0.1. Exceptfor parameter γ in LRQA-4, there exists another parameterω which is set as

ωki = c

√n/(σi (Xk−1) + �),

where c = 3 ∗ √2; n is the number of similar patches group;

σi (Xk−1i ) is the i -th singular value of Xk−1; � is a small value

to avoid being divided by zero.Evaluation measures: The quantitative quality indexes,

i.e., Peak Signal-to-Noise Ration (PSNR) and Structure Sim-ilarity (SSIM) [57], are selected to measure the quality of alldenoising methods. In general, the higher values the PSNR andSSIM are, the better the denoising quality is. All competingdenoising methods are from the source codes and follow theparameter settings in the orginal papers.

The PSNR and SSIM values of all denoising methods arereported in Table II. The highest PSNR and SSIM valuesare highlighted in bold, and the second-highest ones areunderlined. The average PSNR values of all denoising methodsare also computed. From this table, we have the followingobservations:

• In most cases, the proposed LRQA yields the bet-ter denoising results than the best competing methodWNNM. Among them, LRQA-4

(LRQA using the

weighted Schatten γ -norm as defined in Eq. (12))

achieved the highest PSNR and SSIM values. The averageimprovement of the proposed LRQA-4 over WNNMis around 0.113dB, 0.323dB, and 0.366dB when τ =10, 30, 50, respectively. Besides, when the variance ofGaussian noise increases, the advantage of LRQA overWNNM is more obvious. These observations verify theeffectiveness of the low-rank quaternion approximationover the low-rank matrix approximation in real domainfor color image denoising. The main reason is that theproposed LRQA models tree color channels into threeimaginary parts such that the whole color image can beprocessed in a holistic manner and more details of colorimage can be preserved.

• LRQA-1 (LRQA using the quaternion nuclear norm)performs worse than our proposed LRQA variants inall cases. This is because that the quaternion nuclearnorm shrinks each singular value equally, resulting inover-penalizing larger singular values. This observationis consistent with that in Fig. 1.

• WNNM-1 outperforms WNNM-2 in all cases, in whichWNNM-1 concatenates three color channels, whilethe second one restores each channel of a color image

Page 8: Low-Rank Quaternion Approximation for Color Image Processing · 2020. 4. 17. · fits color image processing tasks. Using QR, a color image is represented by a pure quaternion matrix

CHEN et al.: LOW-RANK QUATERNION APPROXIMATION FOR COLOR IMAGE PROCESSING 1433

TABLE II

QUANTITATIVE EVALUATION (PSND/SSIM) OF THE DIFFERENT IMAGE DENOISING ALGORITHMS

TABLE III

AVERAGE IMPROVEMENT (PSNR(DB)) OF LRQA OVER KQSVD [22]

individually and then combined all results together. Thisindicates that the concatenation model-based denoisingmethod may capture more image information than themonochromatic model-based ones.

We also give the visual comparison between all competingmethods and LRQA as shown in Figs. 3 and 5. Obviously,some Gaussian noise still remains in the denoised images byLSCD. KQSVD and NNM can remove Gaussian noise to someextent, but oversmooth the results as shown in the highlighted

red rectangle in Figs. 3 and 5. The proposed LRQA-4 achievesthe best visual performance among all denoising methods.

To compare the proposed LRQA with the quaternion-baseddenoising method: KQSVD [22], we also give the averageimprovement of LRQA over KQSVD as shown in Table IIIand their average runtime of them. The average runtimeof KQSVD on all color images is 112.35s while these ofthe proposed LRQA-1, LRQA-2, LRQA-3 and LRQA-4 are1988.12s, 2112.32s, 2123.25s, and 2205.22s, respectively.One can see that KQSVD is running faster than the proposedLRQA. The main computation burden of LRQA is performingthe quaternion singular value decomposition and nonlocal self-similarity. It is natural because the proposed LRQA takes thelow-rank property and the nonlocal self-similarity prior intoconsideration. However, all proposed LRQA methods achievehigher PSNR values over KQSVD. Specifically, LRQA-3 andLRQA-4 improve more than 1.0dB.

Page 9: Low-Rank Quaternion Approximation for Color Image Processing · 2020. 4. 17. · fits color image processing tasks. Using QR, a color image is represented by a pure quaternion matrix

1434 IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 29, 2020

Fig. 3. Color image denoising results on “Baboon.” (a) Original image; (b) Noisy image corrupted by Gaussian noise with variance τ = 50, respectively;The denoised image reconstructed by: (c) LSCD [52], (d) KQSVD [22], (e) NNM [16], (f) WNNM-1 [53], (g) QLRMA-1, (h) QLRMA-2, (i) QLRMA-3,(j) QLRMA-4. The figure is viewed better in zoomed PDF.

Fig. 4. The 12 color images of size 256 × 256 × 3.

B. Real Noisy Color Image Denoising

In order to give an overall evaluation of LRQA,in this section, we use two real noisy color images2,i.e., Dog and Bears, to test the proposed LRQA for realnoise removal. These two images are mainly contaminatedby the Gaussian noise [11]. Following [1], we first use themethod in [59] to estimate the noise level, since there is nogroundtruth. We select WNNM [53], NC [60], TLR_HTV [11]as the competing methods. Fig. 6 displays the denoising resultsof all methods. We can observe that some noise also left inthe images by NC and TLR_HTV may oversmooth the image.The images by WNNM and LRQA are visually similar. Whilethe structural information among three channels and imagedetails are well captured by LRQA. This further verify theadvantage and effectiveness of the proposed LRQA.

C. Color Image Inpainting

In this section, we test the performance of the proposedLRQA for color image inpainting. In particular, we justconsider the model in Eq. (22) using the quaternion nuclear

2http://demo.ipol.im/demo/125/

norm, laplace function and geman function, corresponding toLRQA-1, LRQA-2, and LRQA-3, respectively, since LRQAusing other nonconvex functions has similar inpainting per-formance. We compare LRQA with several representativeLRMA-based image inpainting methods, including NNM [16],TNNR [4], WNNM, MC-NC [21], LMaFit [58], D-N [13]and F-N [13]. Among them, the first four methods belong tothe matrix rank minimization-based LRMA ones, while theothers are the matrix factorization-based LRMA ones. It isworth noting that D-N and F-N are recently proposed methodsfor image inpainting task. 8 color images are selected as thetest samples as shown in Fig. 7. Following the experimentalsettings in [4], [13], all competing methods are performed oneach channel of the test images individually. We also use thePSNR and SSIM as the quantitative quality.

The quantitative PSNR and SSIM values of all inpaintingmethods are reported in Table IV, in which ρ representsthe ratio of the missing pixels. In most cases, LRQA-2 andLRQA-3 outperform other competing methods. When the ratioof missing pixels ρ = 0.50, 0.75, the average improve-ment of LRQA-2 over the best competing method TNNR isaround 0.353dB and 0.969dB, respectively. When ρ = 0.85,LRQA-2 achieves average 1.051dB improvement over F-N,further verifying the advantages of LRQA. MC-NC achievesthe comparable results with the proposed LRQA-2 whenρ = 0.50, while it performs extremely worse than mostof other methods when ρ = 0.85. The visual comparisonbetween LRQA and all competing inpainting methods on theseventh images with 85% random missing pixels is shownin Fig. 8. It is obvious that the proposed LRQA achievesthe best inpainting results. In particular, LMaFit needs topredefine the rank prior, which is difficult to obtain inreal applications [11], and can not well restore the details.

Page 10: Low-Rank Quaternion Approximation for Color Image Processing · 2020. 4. 17. · fits color image processing tasks. Using QR, a color image is represented by a pure quaternion matrix

CHEN et al.: LOW-RANK QUATERNION APPROXIMATION FOR COLOR IMAGE PROCESSING 1435

Fig. 5. Color image denoising results on “Monarch.” (a) Original image; (b) Noisy image corrupted by Gaussian noise with variance τ = 50; The denoisedimage reconstructed by: (c) LSCD [52], (d) KQSVD [22], (e) NNM [16], (f) WNNM-1 [53], (g) QLRMA-1, (h) QLRMA-2, (i) QLRMA-3, (j) QLRMA-4.The figure is viewed better in zoomed PDF.

TABLE IV

QUANTITATIVE EVALUATION (PSNR/SSIM) OF THE DIFFERENT IMAGE INPAINTING ALGORITHMS

From the comparison between Fig. 8 (d) and 8 (i) restored byNNM and QNNM, respectively, NNM contains quite blurryparts while LRQA-1 can restore more details, which directly

verifies the advantage of the quaternion representation. D-Nand F-N yield similar inpainting results with LRQA, however,the proposed LRQA-2 and LRQA-3 well preserve the light

Page 11: Low-Rank Quaternion Approximation for Color Image Processing · 2020. 4. 17. · fits color image processing tasks. Using QR, a color image is represented by a pure quaternion matrix

1436 IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 29, 2020

Fig. 6. Real color image denoising results on “Dog” and “Bears.”(a) Real noisy image. The denoised image reconstructed by: (b) NC [60],(c) WNNM [53], (d) TLR-HTV [11], (e) LRQA. The figure is viewed betterin zoomed PDF.

Fig. 7. The 8 color images for color image inpainting.

Fig. 8. Color image inpainting results on Img7 with 85% random missingpixels.

of the early morning sun (see the highlighted green rectanglein Fig. 8 (g) and (h)).

D. Model Analysis

In this section, we aim to analyze the LRQA, including theparameter selection, empirical convergence, and runtime. Takethe color image inpiainting with 85% random missing pixelsas an example.

Parameter selection: There is only one free parameter γ inthe proposed Algorithms 2 and 3 based on the laplace function.For color image denoising, γ is experimentally selected fromthe interval [0.1 1.15]. For color image inpainting, γ isselected from [10 60]. Fig. 9 reports the PSNR and SSIMvalues of LRQA for color image denoising and inpainting withdifferent γ . We can see that when γ ranges from 0.5 to 1.15,LRMA performs better to remove noise, while when γ rangesfrom 20 to 40, LRMA achieves promising performance forfilling the missing pixels. The parameter analysis of otherLRQA variants can be implemented in the similar way.

Empirical convergence of LRQA: We also plot the empir-ical convergence of LRQA for color image inpainting on allimages as shown in Fig. 10, where the Y-axis denotes therelative error defined in Eq. (29). The convergence curves on

Fig. 9. PSNR and SSIM values of LRQA for color image inpainting withdifferent γ on Img6 (top) and PSNR and SSIM values of LRQA for colorimage denoising with different γ on Img7 (bottom).

Fig. 10. Relative error versus iteration on all images.

TABLE V

AVERAGE RUNTIME (IN SECONDS) FOR COLOR IMAGE INPAINTING

all images have a slight fluctuation around the 10-th iteration.But after 40 iterations, LRQA converges.

Runtime: To further investigate how LRQA efficientlyprocesses color images, we report the average runtimes of allmethods for color image inpainting on all images as shownin Table V. It can be seen that the matrix factorization-basedmethods, including LMaFit, D-N and F-N, are fastest. Thereason is that they exploit the matrix factorization to depictthe low-rank property and do not require computing SVD.WNNM is the slowest method since it not only needs toperform SVD, but also process three times on three colorchannels. The proposed LRQA achieves similar runtime withthat of TNNR.

VII. CONCLUSION

In this paper, we proposed a novel color image processingmodel, named low-rank quaternion approximation (LRQA),

Page 12: Low-Rank Quaternion Approximation for Color Image Processing · 2020. 4. 17. · fits color image processing tasks. Using QR, a color image is represented by a pure quaternion matrix

CHEN et al.: LOW-RANK QUATERNION APPROXIMATION FOR COLOR IMAGE PROCESSING 1437

which targets at recovering a low-rank quaternion matrix fromits noisy observation. Different from the existing sparse rep-resentaion and low-rank matrix approximation (LRMA)-basedmethods, LRQA can process the three channels in a holisticway. Inspired by the excellent performance of the nonconvexfunctions for LRMA, we proposed a general model for LRQA,and then solve the optimization problem. Extensive experi-ments for color image denoising and inpainting tasks havedemonstrated the effectiveness of the proposed LRQA. In thefuture, we would like to extend LRQA to other color imageprocessing tasks, such as image deblurring, image super-resolution, image deraining and so on.

APPENDIX APROOF OF Theorem 2

Following [40], we need to prove that Eq. (7) has oneunique optimal solution X# and X# equals to X∗ defined inEq. (8). Proof: It is obvious that Eq. (7) has one uniqueoptimal solution since these two terms of Eq. (7) are convex.The optimal solution is denoted as X# and it must satisfy thefollowing formula:

0 ∈ X# − Y + λ∂‖X#‖�, (30)

where ∂‖X#‖� denots the subgradient of the quaternionnuclear norm at X#. For an arbitrary quaernion matrixX ∈ Hm×n , ∂‖X‖� is defined as

∂‖X‖� = {UV� + W : W ∈ Hm×n,

U�W = 0, WV = 0, ‖W‖2 ≤ 1}, (31)

where X = U�V� and ‖ · ‖2 is the spectral norm.Next, we aim to prove that X∗ satisfies the equality in

Eq. (30). The QSVD of Y is

Y = U�V� = U1�1V�1 + U2�2V�

2, (32)

where U = [U1, U2], V = [V1, V2], U1, V1 and �1 are thesingular values vectors and singular values corresponding tosingular values larger than λ, while U2, V2 and �2 correspondto singular values smaller than or equal to λ. Based on theabove definitions and Eq. (8), X∗ can be reformulated as

X∗ = U1(�1 − λI)V�1. (33)

then,

Y − X∗ = λ(U1V�1 + W), W = 1

λU2�2V�

2 . (34)

Since U and V are unitary matrices, U�1U2 = 0 and V�

2V1 = 0.Therefore, U�

1W = 0, WV1 = 0. Since the singular values of�2 are smller than or equal to λ, we have ‖W‖2 ≤ 1. Fromall above discussions, we can obtain that

0 ∈ X∗ − Y + λ∂‖X∗‖�, (35)

which means that X∗ = X#. This completes the proof.

APPENDIX BPROOF OF Theorem 3

Proof: Y = U�YV� is the QSVD of Y ∈ Hm×n and�Y = diag(σ (Y)) = U�YV.

1

2‖Y − X‖2

F + λ∑

i

φ(σi (X)) (36)

= 1

2‖�Y − U�XV‖2

F + λ∑

i

φ(σi (X)) (37)

Eq. (37) is based on the fact that Frobenius norm is unitarilyvariant. Suppose A = U�XV and A and X have the samesingular values, thus Eq. (37) is equal to

= 1

2‖�Y − A‖2

F + λ∑

i

φ(σi (A)) (38)

≥ 1

2‖�Y − �A‖2

F + λ∑

i

φ(σi (A)) (39)

= 1

2‖�Y − �X‖2

F + λ∑

i

φ(σi (X)) (40)

= 1

2‖σ(Y) − σ(X)‖2

2 + λφ(σ(X)) (41)

≥ 1

2‖σ(Y) − σ ∗‖2

2 + λφ(σ ∗) (42)

where σ(Y) and σ(X) are the singular values vectors ofY and X, respectively. Using the Hoffman-Wielandt inequality,we can derive Eq. (39). Then, solving the original model inEq. (9) is transformed by solving Eq. (42). Finally, the optimalsolution X∗ is X∗ = Udiag(σ ∗)V�.

REFERENCES

[1] S. Gu, Q. Xie, D. Meng, W. Zuo, X. Feng, and L. Zhang, “Weightednuclear norm minimization and its applications to low level vision,” Int.J. Comput. Vis., vol. 121, no. 2, pp. 183–208, Jan. 2017.

[2] Y. Xie, S. Gu, Y. Liu, W. Zuo, W. Zhang, and L. Zhang,“Weighted schatten p-norm minimization for image denoising andbackground subtraction,” IEEE Trans. Image Process., vol. 25, no. 10,pp. 4842–4857, Oct. 2016.

[3] Y. Chen, Y. Guo, Y. Wang, D. Wang, C. Peng, and G. He, “Denoising ofhyperspectral images using nonconvex low rank matrix approximation,”IEEE Trans. Geosci. Remote Sens., vol. 55, no. 9, pp. 5366–5380,Sep. 2017.

[4] Y. Hu, D. Zhang, J. Ye, X. Li, and X. He, “Fast and accurate matrixcompletion via truncated nuclear norm regularization,” IEEE Trans.Pattern Anal. Mach. Intell., vol. 35, no. 9, pp. 2117–2130, Sep. 2013.

[5] H. Xue, S. Zhang, and D. Cai, “Depth image inpainting: Improving lowrank matrix completion with low gradient regularization,” IEEE Trans.Image Process., vol. 26, no. 9, pp. 4311–4320, Sep. 2017.

[6] W. Ren, X. Cao, J. Pan, X. Guo, W. Zuo, and M.-H. Yang, “Imagedeblurring via enhanced low-rank prior,” IEEE Trans. Image Process.,vol. 25, no. 7, pp. 3426–3437, Jul. 2016.

[7] S. Wang, Y. Wang, Y. Chen, P. Pan, Z. Sun, and G. He, “Robust PCAusing matrix factorization for background/foreground separation,” IEEEAccess, vol. 6, pp. 18945–18953, 2018.

[8] G. Liu, Z. Lin, S. Yan, J. Sun, Y. Yu, and Y. Ma, “Robust recoveryof subspace structures by low-rank representation,” IEEE Trans. PatternAnal. Mach. Intell., vol. 35, no. 1, pp. 171–184, Jan. 2013.

[9] D. Goldfarb and Z. Qin, “Robust low-rank tensor recovery: Models andalgorithms,” SIAM J. Matrix Anal. Appl., vol. 35, no. 1, pp. 225–253,2014.

Page 13: Low-Rank Quaternion Approximation for Color Image Processing · 2020. 4. 17. · fits color image processing tasks. Using QR, a color image is represented by a pure quaternion matrix

1438 IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 29, 2020

[10] X. Zhou, C. Yang, H. Zhao, and W. Yu, “Low-rank modeling and itsapplications in image analysis,” ACM Comput. Surv., vol. 47, no. 2,p. 36, 2015.

[11] Y. Chen, S. Wang, and Y. Zhou, “Tensor nuclear norm-based low-rankapproximation with total variation regularization,” IEEE J. Sel. TopicsSignal Process., vol. 12, no. 6, pp. 1364–1377, Dec. 2018.

[12] D. D. Lee and H. S. Seung, “Learning the parts of objects by non-negative matrix factorization,” Nature, vol. 401, no. 6755, p. 788,1999.

[13] F. Shang, J. Cheng, Y. Liu, Z. Luo, and Z. Lin, “Bilinear factor matrixnorm minimization for robust PCA: Algorithms and applications,” IEEETrans. Pattern Anal. Mach. Intell., vol. 40, no. 9, pp. 2066–2080,Sep. 2018.

[14] F. Xu, Y. Chen, C. Peng, Y. Wang, X. Liu, and G. He, “Denoising ofhyperspectral image using low-rank matrix factorization,” IEEE Geosci.Remote Sens. Lett., vol. 14, no. 7, pp. 1141–1145, Jul. 2017.

[15] M. Fazel, “Matrix rank minimization with applications,” Ph.D. disserta-tion, Dept. Elect. Eng., Stanford Univ., Stanford, CA, USA, 2002.

[16] E. J. Candès and B. Recht, “Exact matrix completion via con-vex optimization,” Found. Comput. Math., vol. 9, no. 6, p. 717,Dec. 2009.

[17] Z. Kang, C. Peng, and Q. Cheng, “Robust PCA via nonconvex rankapproximation,” in Proc. IEEE Int. Conf. Data Mining, Nov. 2015,pp. 211–220.

[18] L. E. Frank and J. H. Friedman, “A statistical view of some chemo-metrics regression tools,” Technometrics, vol. 35, no. 2, pp. 109–135,1993.

[19] Z. Kang, C. Peng, J. Cheng, and Q. Cheng, “LogDet rank minimizationwith application to subspace clustering,” Comput. Intell. Neurosci.,vol. 2015, Jan. 2015, Art. no. 68.

[20] C. Lu, J. Tang, S. Yan, and Z. Lin, “Generalized nonconvex nonsmoothlow-rank minimization,” in Proc. IEEE Conf. Comput. Vis. PatternRecognit., Jun. 2014, pp. 4130–4137.

[21] F. Nie, Z. Hu, and X. Li, “Matrix completion based on non-convexlow-rank approximation,” IEEE Trans. Image Process., vol. 28, no. 5,pp. 2378–2388, May 2018.

[22] Y. Xu, L. Yu, H. Xu, H. Zhang, and T. Nguyen, “Vector sparserepresentation of color image using quaternion matrix analysis,”IEEE Trans. Image Process., vol. 24, no. 4, pp. 1315–1329,Apr. 2015.

[23] J. Xu, L. Zhang, D. Zhang, and X. Feng, “Multi-channel weightednuclear norm minimization for real color image denoising,” in Proc.IEEE Int. Conf. Comput. Vis., Oct. 2017, pp. 1096–1104.

[24] A. Buades, B. Coll, and J.-M. Morel, “A non-local algorithm for imagedenoising,” in Proc. IEEE Comput. Soc. Conf. Comput. Vis. PatternRecognit., vol. 2, Jun. 2005, pp. 60–65.

[25] L. Zhang, W. Dong, D. Zhang, and G. Shi, “Two-stage image denoisingby principal component analysis with local pixel grouping,” PatternRecognit., vol. 43, no. 4, pp. 1531–1549, 2010.

[26] W. Dong, L. Zhang, G. Shi, and X. Li, “Nonlocally centralized sparserepresentation for image restoration,” IEEE Trans. Image Process.,vol. 22, no. 4, pp. 1620–1630, Apr. 2013.

[27] S.-C. Pei and C.-M. Cheng, “Color image processing by using binaryquaternion-moment-preserving thresholding technique,” IEEE Trans.Image Process., vol. 8, no. 5, pp. 614–628, May 1999.

[28] F. Zhang, “Quaternions and matrices of quaternions,” Linear AlgebraAppl., vol. 251, pp. 21–57, Jan. 1997.

[29] S. J. Sangwine and N. L. Bihan, “Quaternion singular value decomposi-tion based on bidiagonalization to a real or complex matrix using quater-nion Householder transformations,” Appl. Math. Comput., vol. 182,no. 1, pp. 727–738, 2006.

[30] X. Xiao and Y. Zhou, “Two-dimensional quaternion PCA and sparsePCA,” IEEE Trans. Neural Netw. Learn. Syst., vol. 30, no. 7,pp. 2028–2042, Jul. 2018.

[31] R. Lan and Y. Zhou, “Quaternion-michelson descriptor for colorimage classification,” IEEE Trans. Image Process., vol. 25, no. 11,pp. 5281–5292, Nov. 2016.

[32] N. Le Bihan and S. J. Sangwine, “Quaternion principal componentanalysis of color images,” in Proc. IEEE Int. Conf. Image Process.,vol. 1, Sep. 2003, p. I-809.

[33] C. Zou, K. Kou, and Y. Wang, “Quaternion collaborative and sparserepresentation with application to color face recognition,” IEEE Trans.Image Process., vol. 25, no. 7, pp. 3287–3302, Jul. 2016.

[34] R. Zeng, J. Wu, Z. Shao, Y. Chen, B. Chen, L. Senhadji, andH. Shu, “Color image classification via quaternion principal com-ponent analysis network,” Neurocomputing, vol. 216, pp. 416–428,Dec. 2016.

[35] X. Zhu, Y. Xu, H. Xu, and C. Chen, “Quaternion convolutional neuralnetworks,” in Proc. Eur. Conf. Comput. Vis., 2018, pp. 631–647.

[36] R. Lan, Y. Zhou, and Y. Y. Tang, “Quaternionic weber local descriptorof color images,” IEEE Trans. Circuits Syst. Video Technol., vol. 27,no. 2, pp. 261–274, Feb. 2017.

[37] J. Wright, A. Y. Yang, A. Ganesh, S. S. Sastry, and Y. Ma, “Robust facerecognition via sparse representation,” IEEE Trans. Pattern Anal. Mach.Intell., vol. 31, no. 2, pp. 210–227, Feb. 2009.

[38] T. Parcollet, M. Morchid, and G. Linarès, “Quaternion convolutionalneural networks for heterogeneous image processing,” in Proc. IEEEInt. Conf. Acoust., Speech Signal Process., May 2019, pp. 8514–8518.

[39] E. J. Candès, X. Li, Y. Ma, and J. Wright, “Robust principal componentanalysis?” J. ACM, vol. 58, no. 3, p. 11, May 2011.

[40] J.-F. Cai, E. J. Candès, and Z. Shen, “A singular value thresholdingalgorithm for matrix completion,” SIAM J. Optim., vol. 20, no. 4,pp. 1956–1982, 2010.

[41] Y. Chen, X. Xiao, and Y. Zhou, “Multi-view clustering via simulta-neously learning graph regularized low-rank tensor representation andaffinity matrix,” in Proc. IEEE Int. Conf. Multimedia Expo, Jul. 2019,pp. 1348–1353.

[42] A. E. Waters, A. C. Sankaranarayanan, and R. Baraniuk, “SpaRCS:Recovering low-rank and sparse matrices from compressive measure-ments,” in Proc. Adv. Neural Inf. Process. Syst., 2011, pp. 1089–1097.

[43] Q. Sun, S. Xiang, and J. Ye, “Robust principal component analysis viacapped norms,” in Proc. ACM SIGKDD Int. Conf. Knowl. DiscoveryData Mining, 2013, pp. 311–319.

[44] T.-H. Oh, Y.-W. Tai, J.-C. Bazin, H. Kim, and I. S. Kweon, “Partialsum minimization of singular values in robust PCA: Algorithm andapplications,” IEEE Trans. Pattern Anal. Mach. Intell., vol. 38, no. 4,pp. 744–758, Apr. 2016.

[45] M. Fazel, H. Hindi, and S. P. Boyd, “Log-det heuristic for matrixrank minimization with applications to Hankel and Euclidean distancematrices,” in Proc. IEEE Amer. Control Conf., vol. 3, Jun. 2003,pp. 2156–2162.

[46] Q. Feng et al., “Superimposed sparse parameter classifiers for facerecognition,” IEEE Trans. Cybern., vol. 47, no. 2, pp. 378–390,Feb. 2017.

[47] W. R. Hamilton, Elements of Quaternions. Harlow, U.K.: Longmans,1866.

[48] R. D. Schafer, “On the algebras formed by the Cayley–Dickson process,”Amer. J. Math., vol. 76, no. 2, pp. 435–446, 1954.

[49] P. R. Girard, Quaternions, Clifford Algebras and Relativistic Physics.Springer, 2007.

[50] M. E. Kilmer and C. D. Martin, “Factorization strategies for third-ordertensors,” Linear Algebra Appl., vol. 435, no. 3, pp. 641–658, 2011.

[51] P. D. Tao and L. T. H. An, “Convex analysis approach to DC pro-gramming: Theory, algorithms and applications,” Acta Math. Vietnamica,vol. 22, no. 1, pp. 289–355, 1997.

[52] M. Rizkinia, T. Baba, K. Shirai, and M. Okuda, “Local spectral com-ponent decomposition for multi-channel image denoising,” IEEE Trans.Image Process., vol. 25, no. 7, pp. 3208–3218, Jul. 2016.

[53] S. Gu, L. Zhang, W. Zuo, and X. Feng, “Weighted nuclear normminimization with application to image denoising,” in Proc. IEEE Conf.Comput. Vis. Pattern Recognit., Jun. 2014, pp. 2862–2869.

[54] W. Dong, G. Shi, and X. Li, “Nonlocal image restoration with bilateralvariance estimation: A low-rank approach,” IEEE Trans. Image Process.,vol. 22, no. 2, pp. 700–711, Feb. 2013.

[55] Y. Chen, Y. Wang, M. Li, and G. He, “Augmented Lagrangian alternatingdirection method for low-rank minimization via non-convex approxi-mation,” Signal, Image Video Process., vol. 11, no. 7, pp. 1271–1278,2017.

[56] L. Li, X. Wang, and G. Wang, “Alternating direction method ofmultipliers for separable convex optimization of real functions incomplex variables,” Math. Problems Eng., vol. 2015, Nov. 2015,Art. no. 104531.

[57] Z. Wang, A. C. Bovik, H. R. Sheikh, and E. P. Simoncelli,“Image quality assessment: From error visibility to structural simi-larity,” IEEE Trans. Image Process., vol. 13, no. 4, pp. 600–612,Apr. 2004.

[58] Y. Shen, Z. Wen, and Y. Zhang, “Augmented Lagrangian alternatingdirection method for matrix separation based on low-rank factorization,”Optim. Methods Softw., vol. 29, no. 2, pp. 239–263, 2014.

[59] D. L. Donoho, “De-noising by soft-thresholding,” IEEE Trans. Inf.Theory, vol. 41, no. 3, pp. 613–627, May 1995.

[60] M. Lebrun, M. Colom, and J.-M. Morel, “Multiscale image blinddenoising,” IEEE Trans. Image Process., vol. 24, no. 10, pp. 3149–3161,Oct. 2015.

Page 14: Low-Rank Quaternion Approximation for Color Image Processing · 2020. 4. 17. · fits color image processing tasks. Using QR, a color image is represented by a pure quaternion matrix

CHEN et al.: LOW-RANK QUATERNION APPROXIMATION FOR COLOR IMAGE PROCESSING 1439

Yongyong Chen received the B.S. and M.S. degreesfrom the College of Mathematics and Systems Sci-ence, Shandong University of Science and Tech-nology, Qingdao, China. He is currently pursuingthe Ph.D. degree with the Department of Com-puter and Information Science, University of Macau,Macau, China. He visited the National Key Labfor Novel Software Technology, Nanjing Univer-sity, as an Exchange Student in 2017. His researchinterests include (non-convex) low-rank and sparsematrix/tensor decomposition models, with applica-

tions to image processing, data mining, and computer vision.

Xiaolin Xiao received the B.E. degree from WuhanUniversity, China, in 2013, and the Ph.D. degreefrom the University of Macau, Macau, China,in 2019. She is currently a Postdoctoral Fellow withthe School of Computer Science and Engineering,South China University of Technology, China. Herresearch interests include superpixel segmentation,saliency detection, and color image processing andunderstanding.

Yicong Zhou (M’07–SM’14) received the B.S.degree from Hunan University, Changsha, China,and the M.S. and Ph.D. degrees from TuftsUniversity, Massachusetts, USA, all in electricalengineering.

He is currently an Associate Professor and theDirector of the Vision and Image Processing Lab-oratory, Department of Computer and InformationScience, University of Macau. His research inter-ests include image processing and understanding,computer vision, machine learning, and multimediasecurity.

Dr. Zhou is a Senior Member of the International Society for OpticalEngineering (SPIE). He was a recipient of the Third Price of Macau NaturalScience Award in 2014. He is a Co-Chair of Technical Committee onCognitive Computing in the IEEE Systems, Man, and Cybernetics Society. Heserves as an Associate Editor for the IEEE TRANSACTIONS ON CIRCUITSAND SYSTEMS FOR VIDEO TECHNOLOGY, the IEEE TRANSACTIONS ON

GEOSCIENCE AND REMOTE SENSING, and four other journals.