4
8 IEEE Sensor Array and Multichannel Signal Processing Workshop (SAM), June 22-25, 2014, A Coru˜ na, Spain 1 A Sparse Bayesian Learning Algorithm With Dictionary Parameter Estimation Thomas L. Hansen * , Mihai A. Badiu * , Bernard H. Fleury * and Bhaskar D. Rao * Dept. of Electronic Systems, Aalborg University, Fr. Bajers Vej 7, DK-9220, Aalborg, Denmark – {tlh,in mba, bfl}@es.aau.dk Dept. of Electrical and Computer Engineering, University of California at San Diego, 9500 Gilman Drive, La Jolla, CA 92093, USA – [email protected] Abstract—This paper concerns sparse decomposition of a noisy signal into atoms which are specified by unknown continuous- valued parameters. An example could be estimation of the model order, frequencies and amplitudes of a superposition of complex sinusoids. The common approach is to reduce the continuous parameter space to a fixed grid of points, thus restricting the solution space. In this work, we avoid discretization by working directly with the signal model containing parameterized atoms. Inspired by the ”fast inference scheme” by Tipping and Faul we develop a novel sparse Bayesian learning (SBL) algorithm, which estimates the atom parameters along with the model order and weighting coefficients. Numerical experiments for spectral estimation with closely-spaced frequency components, show that the proposed SBL algorithm outperforms state-of-the- art subspace and compressed sensing methods. I. I NTRODUCTION Suppose we have a length-N signal x given by a weighted sum of K N elementary functions, so-called signal atoms: x = K i=1 ψ(θ i )α i , (1) where θ =[θ 1 , ··· K ] T is a vector of atom (dictionary) parameters and α =[α 1 , ··· K ] T is a vector of weight- ing coefficients. We specify the atoms by the vector-valued function ψ : [0, 1) C N . 1 We take noisy measurements y C N as y = x + w, where w C N is a zero-mean complex Gaussian noise vector with independent and identical distributed (i.i.d.) entries of variance λ -1 . The problem of estimating K, θ and α given the atom specification ψ(·) and observation y is ubiquitous in sig- nal processing. When the Fourier atom is considered, i.e. ψ(θ i ) = [e j2πθi0 , ··· , e j2πθi(N-1) ] T , the problem reduces to the line spectral estimation problem, which has a wide range of applications, e.g. direction of arrival estimation. The problem also arises in the context of compressed sensing (CS) reconstruction: When a sensing matrix Φ is employed, the atoms in our model become ψ(·)= Φ ˜ ψ(·), where ˜ ψ(·) is the atom specified without a sensing matrix. A common approach, particularly in CS [1], is to discretize the parameter space into a (uniform) grid of M N values on [0, 1). Then, a sparse representation of x is sought in the This work was partly supported by the European Commission within the FP7 Network of Excellence in Wireless COMmunications NEWCOM# (Grant agreement no. 318306). 1 Any function ψ(·) having a connected and bounded domain in R can be expressed in this form. dictionary obtained by evaluating ψ(·) at the M grid points. However, the mismatch between the atoms in the dictionary and the true atoms limits the accuracy of the fixed dictionary approach. To combat the severity of the atom mismatch a finer grid can be employed, leading to two undesired effects: a) the dictionary becomes increasingly coherent, rendering the estimation problem ill-posed [1] and b) the larger size of the dictionary results in higher computational cost of estimation. Model-based CS can be applied to mitigate the coherence issue occurring with fixed dictionaries, see for example [2], [3]. If the objective is to reconstruct the signal vector x, and not to find a decomposition into atoms, “analysis sparsity” [4], [5] alleviates the need for dictionary incoherence. All these coherence-controlling methods suffer from high computational complexity when a fine grid is used. Recent works [6] formulate our estimation problem as a total variation norm minimization problem. The strongly related parallel development [7] uses a similar atomic norm minimization. Minimizing the 1 -norm promotes sparse es- timates. The idea in [6], [7] is to generalize the 1 -norm from vectors to the real line, such that minimizing the norm promotes a sparse signal on the real line, i.e. a sum of spikes. For Fourier atoms, [6], [7] rewrite the norm minimization as a semi-definite program. The theoretical analysis of these works requires the frequencies in θ to be well-separated, e.g. by at least 2/N in [6]. Another recently-proposed approach is to introduce a com- plementary dictionary which characterizes the basis mismatch, e.g. [8], [9]. For Fourier atoms (1) reduces to the line spectral estimation problem and many methods have been proposed; see [10] for a list of references. The most prominent of these methods are the so-called subspace methods [11], e.g. ESPRIT [12]. In this paper, we devise a sparse Bayesian learning (SBL) algorithm for estimating K, θ and α. Since most SBL methods (e.g. [13]–[16]) are developed for discrete dictionaries, they suffer from the mentioned drawbacks when applied to our problem. Therefore, we instead use the parameterized model (1) to devise our algorithm. Specifically, we extend the sparse prior model proposed in [13] and devise an inference scheme, inspired by [14], which estimates K, θ and α. A parallel de- velopment is found in [17], which uses a variational Bayesian method for estimation. c 2014 IEEE. Personal use of this material is permitted. Permission from IEEE must be obtained for all other uses, in any current or future media, including reprinting/republishing this material for advertising or promotional purposes, creating new collective works, for resale or redistribution to servers or lists, or reuse of any copyrighted component of this work in other works. brought to you by CORE View metadata, citation and similar papers at core.ac.uk provided by VBN

A Sparse Bayesian Learning Algorithm With Dictionary

  • Upload
    others

  • View
    4

  • Download
    0

Embed Size (px)

Citation preview

Page 1: A Sparse Bayesian Learning Algorithm With Dictionary

8th IEEE Sensor Array and Multichannel Signal Processing Workshop (SAM), June 22-25, 2014, A Coruna, Spain 1

A Sparse Bayesian Learning AlgorithmWith Dictionary Parameter EstimationThomas L. Hansen∗, Mihai A. Badiu∗, Bernard H. Fleury∗ and Bhaskar D. Rao†

∗ Dept. of Electronic Systems, Aalborg University, Fr. Bajers Vej 7,DK-9220, Aalborg, Denmark – {tlh,in mba, bfl}@es.aau.dk

† Dept. of Electrical and Computer Engineering, University of California at San Diego,9500 Gilman Drive, La Jolla, CA 92093, USA – [email protected]

Abstract—This paper concerns sparse decomposition of a noisysignal into atoms which are specified by unknown continuous-valued parameters. An example could be estimation of the modelorder, frequencies and amplitudes of a superposition of complexsinusoids. The common approach is to reduce the continuousparameter space to a fixed grid of points, thus restricting thesolution space. In this work, we avoid discretization by workingdirectly with the signal model containing parameterized atoms.Inspired by the ”fast inference scheme” by Tipping and Faulwe develop a novel sparse Bayesian learning (SBL) algorithm,which estimates the atom parameters along with the modelorder and weighting coefficients. Numerical experiments forspectral estimation with closely-spaced frequency components,show that the proposed SBL algorithm outperforms state-of-the-art subspace and compressed sensing methods.

I. INTRODUCTION

Suppose we have a length-N signal x given by a weightedsum of K � N elementary functions, so-called signal atoms:

x =∑Ki=1ψ(θi)αi, (1)

where θ = [θ1, · · · , θK ]T is a vector of atom (dictionary)parameters and α = [α1, · · · , αK ]T is a vector of weight-ing coefficients. We specify the atoms by the vector-valuedfunction ψ : [0, 1) → CN .1 We take noisy measurementsy ∈ CN as y = x + w, where w ∈ CN is a zero-meancomplex Gaussian noise vector with independent and identicaldistributed (i.i.d.) entries of variance λ−1.

The problem of estimating K, θ and α given the atomspecification ψ(·) and observation y is ubiquitous in sig-nal processing. When the Fourier atom is considered, i.e.ψ(θi) = [ej2πθi0, · · · , ej2πθi(N−1)]T, the problem reducesto the line spectral estimation problem, which has a widerange of applications, e.g. direction of arrival estimation. Theproblem also arises in the context of compressed sensing (CS)reconstruction: When a sensing matrix Φ is employed, theatoms in our model become ψ(·) = Φψ(·), where ψ(·) is theatom specified without a sensing matrix.

A common approach, particularly in CS [1], is to discretizethe parameter space into a (uniform) grid of M ≥ N valueson [0, 1). Then, a sparse representation of x is sought in the

This work was partly supported by the European Commission within theFP7 Network of Excellence in Wireless COMmunications NEWCOM# (Grantagreement no. 318306).

1Any function ψ(·) having a connected and bounded domain in R can beexpressed in this form.

dictionary obtained by evaluating ψ(·) at the M grid points.However, the mismatch between the atoms in the dictionaryand the true atoms limits the accuracy of the fixed dictionaryapproach. To combat the severity of the atom mismatch a finergrid can be employed, leading to two undesired effects: a)the dictionary becomes increasingly coherent, rendering theestimation problem ill-posed [1] and b) the larger size of thedictionary results in higher computational cost of estimation.

Model-based CS can be applied to mitigate the coherenceissue occurring with fixed dictionaries, see for example [2],[3]. If the objective is to reconstruct the signal vector x, andnot to find a decomposition into atoms, “analysis sparsity” [4],[5] alleviates the need for dictionary incoherence. All thesecoherence-controlling methods suffer from high computationalcomplexity when a fine grid is used.

Recent works [6] formulate our estimation problem asa total variation norm minimization problem. The stronglyrelated parallel development [7] uses a similar atomic normminimization. Minimizing the `1-norm promotes sparse es-timates. The idea in [6], [7] is to generalize the `1-normfrom vectors to the real line, such that minimizing the normpromotes a sparse signal on the real line, i.e. a sum of spikes.For Fourier atoms, [6], [7] rewrite the norm minimization as asemi-definite program. The theoretical analysis of these worksrequires the frequencies in θ to be well-separated, e.g. by atleast 2/N in [6].

Another recently-proposed approach is to introduce a com-plementary dictionary which characterizes the basis mismatch,e.g. [8], [9].

For Fourier atoms (1) reduces to the line spectral estimationproblem and many methods have been proposed; see [10] fora list of references. The most prominent of these methods arethe so-called subspace methods [11], e.g. ESPRIT [12].

In this paper, we devise a sparse Bayesian learning (SBL)algorithm for estimating K, θ and α. Since most SBL methods(e.g. [13]–[16]) are developed for discrete dictionaries, theysuffer from the mentioned drawbacks when applied to ourproblem. Therefore, we instead use the parameterized model(1) to devise our algorithm. Specifically, we extend the sparseprior model proposed in [13] and devise an inference scheme,inspired by [14], which estimates K, θ and α. A parallel de-velopment is found in [17], which uses a variational Bayesianmethod for estimation.

c© 2014 IEEE. Personal use of this material is permitted. Permission from IEEE must be obtained for all other uses, in any current orfuture media, including reprinting/republishing this material for advertising or promotional purposes, creating new collective works, for

resale or redistribution to servers or lists, or reuse of any copyrighted component of this work in other works.

brought to you by COREView metadata, citation and similar papers at core.ac.uk

provided by VBN

Page 2: A Sparse Bayesian Learning Algorithm With Dictionary

2

↵i�i �" a

Observed variablesUnobserved variables Parameters

✓ii = 1, . . . , K n = 1, . . . , N

b

yn

Fig. 1. Bayesian network of the probabilistic model.

II. PROBABILISTIC MODELLING

Our probabilistic model is an extension of that in [13]to include modelling of the atom parameter vector θ. TheBayesian network representation of our model is shown inFig. 1. We note that the model in [13] is itself obtained bymodifying the model in [15] such that α depends on λ. Thismodification is employed since it gives certain benefits incomputational complexity as discussed in [13].2

The following treatment is valid for both the real andcomplex cases.3 The observation y is taken in white Gaussiannoise with variance λ−1:

p(y|θ,α, λ) = N(y|Ψ(θ)α, λ−1I), (2)

where the dictionary matrix Ψ(θ) = [ψ(θ1), · · · , ψ(θK)]contains the atoms as columns. The noise precision λ is mod-elled as gamma distributed with shape a and rate b: p(λ) =Ga(λ|a, b). The atom parameter vector θ ∈ [0, 1)K has i.i.d.uniformly distributed entries: p(θ) =

∏Ki=1 unif(θi|0, 1). The

coefficients α are modelled through a two-layer hierarchicalspecification. The first layer is a zero-mean Gaussian distri-bution, i.e. p(α|γ, λ) = N(α|0, λ−1Γ), where Γ = diag(γ).The vector γ constitute the second layer. Its entries are mod-elled as i.i.d. gamma distributed: p(γ) =

∏Ki=1 Ga(γi|ε, η).

III. BAYESIAN INFERENCE

To infer on (K,θ,α), we proceed by applying Type-IIestimation [18], i.e. we use Bayesian inference to find pointestimates (K, θ, γ, λ) of the parameters (K,θ,γ, λ) and findthe estimate α of α as the mode of p(α|y, θ, γ, λ). FromBayes rule, p(α|y,θ,γ, λ) turns out to be a normal density:

p(α|y,θ,γ, λ) = N(α|µ, λ−1Σ), (3)

where

µ = ΣΨH(θ)y, (4)

Σ =(ΨH(θ)Ψ(θ) + Γ−1

)−1. (5)

We find (θ, γ, λ) as an approximation of the maximum aposteriori (MAP) estimate

(θ, γ, λ)MAP = arg max(θ,γ,λ)

ln p(θ,γ, λ|y). (6)

2Note that with this modification, the entries of γ represent the ratio of theweight variances to the noise variance, i.e. signal-to-noise ratios.

3 The multivariate normal density is parameterized to encompass both thereal (ρ = 1

2) and complex (ρ = 1) cases:

N(x|µ,Σ) =( ρπ

)ρ dim(x)|Σ|−ρ exp

(−ρ(x−µ)HΣ−1(x−µ)

)

Following the steps of [13], we proceed by writing the log-posterior by its factors:

ln p(θ,γ, λ|y) ∝e ln p(y|θ,γ, λ)p(θ)p(γ)p(λ), (7)

where x ∝e y denotes x = y+ const. The marginal likelihoodcan be found by marginalizing the coefficient vector α out:

p(y|θ,γ, λ) = N(y|0, λ−1B), (8)

where B =(I−Ψ(θ)ΣΨH(θ)

)−1= I + Ψ(θ)ΓΨH(θ).

Using this result, we can rewrite the log-posterior (7) suchthat one set of parameters (θi, γi, λ) appear explicitly:

ln p(θ,γ, λ|y) ∝e ρN lnλ− ρ ln |B−i|

− ρ ln (1 + γisi)− ρλyHB−1−iy +

ρλ|qi|2

γ−1i + si

+

K∑k=1

{(ε− 1) ln γk − ηγk}+ (a− 1) lnλ− bλ, (9)

where B−i = I+Ψ(θ−i)Γ−iΨH(θ−i) and Γ−i = diag(γ−i).

The notation a−i denotes a vector a with the ith componentremoved. We have further defined the quantities

si , ψH(θi)B

−1−iψ(θi) and qi , ψ

H(θi)B−1−iy. (10)

To estimate the parameters {θi, γi}i=1,...,K and λ we it-eratively maximize (9) with respect to one parameter, whilekeeping all other parameters fixed at their current estimate. Todo so, we take partial derivatives of (9) w.r.t λ and γi andsolve for the roots. Following a procedure similar to that in[15], to analyse the stationary points, we get the updates

λ =ρN + a− 1

ρyHB−1y − b(11)

γi =

{−(2ε−2−ρ)si−ρλ|qi|2−

√∆

2(ε−1−ρ)s2iρλ|qi|2 > δ,

0 otherwise,(12)

where ∆ =(

(2ε− 2− ρ)si + ρλ|qi|2)2

−4(ε−1)(ε−1−ρ)s2i

and δ =(

2 + ρ− 2ε+ 2√

(1− ε)(1 + ρ− ε))si.

It is not tractable to obtain similar closed-from expressionfor updating θi. We instead use Newton’s method for uncon-strained optimization, which iteratively updates θi as

θnewi = θold

i − l′(θoldi

)/l′′(θoldi

), (13)

where l′(θi) and l′′(θi) are the first and second partial deriva-tives of (9) with respect to θi as given in the Appendix.

Algorithm 1 combines update equations (11), (12) and(13) into a constructive sequential algorithm for estimating(K,θ,α). The algorithm follows the idea of the fast inferencescheme [14] and starts with an empty model and iterativelyadds atoms into the model. Due to the highly-nonlinear formof (9) as a function of θi, the search for candidate atoms toinclude in the model is aided by a grid search. We have foundthat using the Jeffreys prior for γi and λ, obtained as thespecial case ε = η = a = b = 0, yields good performance.

Page 3: A Sparse Bayesian Learning Algorithm With Dictionary

3

Algorithm 1: SBL with Dictionary Parameter EstimationInput: Signal measurement y.Output: Estimates of the model order K, atom

parameters in θ ∈ [0, 1)K and coefficients inα ∈ CK .

Parameters: Prior parameters ε, a, b (η = 0 assumed.)1 (θ, γ)← Empty vectors; λ← 1002 while Stopping criterion not met do3 if Iteration number ∈ {0, 5, 10, 20, 30, . . .} then4 for i = 0, . . . , 3N − 1 do5 θcandidate ← Calc. from (13), with θold = i

3N .6 γcandidate ← Calculate from (12).7 if γcandidate > 0 then8 Append (θcandidate, γcandidate) to (θ, γ).9 end

10 end11 end12 for i = 1, . . . , ||θ||0 do13 θi ← Update from (13).14 γi ← Update from (12).15 if γi == 0 then16 Remove component i from θ and γ.17 end18 end19 λ← Update from (11).20 end21 α← µ µ calculated from (4) based on (θ, γ).

We have observed that applying the update (13) morethan once in lines 5 and 13, does not improve the overallconvergence rate significantly for Fourier atoms. As a stoppingcriterion in the kth iteration we use ||θk−θk−1||∞ < 10−6/N .To reduce computational complexity, the matrix Σ can beiteratively updated as atoms are added and removed from themodel, by using the expressions in [16]. With these updates,the algorithm has computational complexity per iteration ofO(N2K) when searching for new atoms to add to the modeland O(NK2) otherwise.

IV. NUMERICAL EXPERIMENTS

To investigate the performance of our algorithm, we conducta numerical experiment with Fourier atoms. The simulationcode is available online at http://doi.org/srp. We use N = 100measurements and K = 10 signal components. We generate θconsisting of 5 pairs of frequencies. The distance d(θi, θi+1)between two paired frequencies is i.i.d. uniform random on theinterval [0.7/N, 1/N ]. The distance metric d(·, ·) is the wrap-around distance on the interval [0, 1). The pairs are locatedrandomly such that any set of frequencies (θi, θj) whichare not paired are separated by at least d(θi, θj) ≥ 1.5/N .The complex amplitudes of the signal components in α aregenerated i.i.d. with uniform random phase on [0, 2π) andamplitudes drawn from a normal density of mean 1 andvariance 0.1.

We measure performance in terms of normalized mean-squared error (MSE) of the reconstructed signal x = Ψ(θ)αand the following performance metric for θ:

β(θ, θ) ,1

K

K∑i=1

(minθ∈θ

d(θ, θi)

)2

, (14)

where the notation θ ∈ θ signifies that θ is given as one of theentries in θ. Further, we show histograms of the model orderestimation error at 15 dB SNR for some of the algorithms. Allreported values are averaged over 100 trials.

We compare our algorithm with the fixed dictionary SBLalgorithm [13], which our algorithm is based on, also usingJeffreys priors for γi and λ (ε = η = a = b = 0). Further, wecompare with a selection of state-of-the-art estimators: totalvariation norm minimization via semi-definite programming(SDP) [6]; a complex reformulation of continuous basis pursuit(CCBP) from [9]; the integer program employed as part ofspectral iterative hard thresholding (SIHT IP) [2]; “analysis”basis-pursuit denoising (BPDN) [5]; and ESPRIT [12].

Fixed dictionary SBL, SIHT IP and analysis BPDN aregiven a fixed dictionary of size M = 10N . Due to com-putational feasibility, CCBP is given a smaller dictionary ofsize M = 3N . ESPRIT is given the true model order, i.e.K = K, and SIHT IP is given K = 3K as the model order.The regularization parameters of SDP and CCBP are set basedon the true noise variance λ. The above choices of K andregularization parameters are based on preliminary numericalinvestigations to obtain the best performance of the algorithms.

For the reconstruction MSE, we compare with an oracleestimator which computes a least-squares solution to find αfrom the true set of frequencies. The Cramer-Rao bound for θ(CRB) is calculated as in [19] and averaged over all simulatedsignal realizations. The CRB is a lower bound on β(θ, θ),assuming that the matching of an estimate θ to each of thetrue frequencies θi in the minimization of (14) is one-to-one.

The simulation results are shown in Fig. 2. We see that ourproposed SBL algorithm is superior to all other algorithms atSNR ≥ 5 dB. At SNR ≥ 15 dB ESPRIT performs similarly toour algorithm in terms of reconstruction MSE. The proposedSBL algorithm attains the CRB of β(θ, θ) at SNR ≥ 5 dB,while ESPRIT only almost does so at SNR ≥ 20 dB.

All algorithms which rely on a fixed dictionary (fixed dict.SBL, CCBP and SIHT IP) show a saturation effect in β(θ, θ);beyond a certain SNR threshold the performance does notimprove since the resolution limit of the dictionary is reached.Notice how CCBP performs similar to fixed dict. SBL in termsof β(θ, θ), even though it uses a much smaller dictionary.

For fixed dict. SBL, SDP and SIHT IP the saturation effectin terms of β(θ, θ) does not give rise to a correspondingsaturation in reconstruction MSE. These three estimators are,however, significantly overestimating K in high SNR (see Fig.2c and remember that K = 3K for SIHT IP). Our algorithmhas small model order estimation error for all SNR valuesgreater than 0 dB and still out-performs ESPRIT which is giventhe true model order.

Page 4: A Sparse Bayesian Learning Algorithm With Dictionary

4

Proposed SBL Fixed Dict. SBL SDP CCBP SIHT IP Analysis BPDN ESPRIT Oracle CRB

0 10 20 30

10−4

10−2

100

SNR [dB]

Avg[ ||x−x||2 2/||x||2 2

]

(a) Normalized reconstruction MSE.

0 10 20 3010−9

10−6

10−3

SNR [dB]

Avg[ β

(θ,θ

)](b) Performance metric for θ.

0 5 10 15 20 25 30 35

0

10

20

SNR [dB]

K−K

(c) Mean model order estimation error withshaded regions giving the 20−80% quantiles.

Fig. 2. Simulation results.

V. CONCLUSION

In this paper we addressed the so-called off-grid sparsedecomposition problem which consists in decomposing a noisysignal into atoms specified by unknown continuous-valuedparameters. We found that a convenient way to avoid theundesired effects of a discretized atom parameter space isto consider the atom parameters as unknown variables to beestimated along with the model order and atom coefficients.

Thus, we proposed a novel SBL algorithm based on anextension of the probabilistic model in [13]. Inspired by theconstructive scheme of [14], we devised update expressionswhich provide a simple criterion for inclusion of an atominto the estimated model; candidate atoms for inclusion areidentified via a grid search. Unlike for the rest of the variables,the updates of the atom parameters cannot be computed inclosed-form, so we resorted to Newton’s method to updatethese estimates.

The numerical results show that our algorithm is superior tothe reference algorithms for spectral estimation with closely-spaced frequency components. This is remarkable since ouralgorithm estimates the model order, while many of thereference algorithms are given the true value. An interestingaspect for further research is to reduce the computationaldemands of the algorithm, in particular that connected withthe search for new atoms to include in the model estimate.

APPENDIX: PARTIAL DERIVATIVES OF (9) W.R.T. θi

l′(θi) =

ρλγi

1 + γisi

∂|qi|2

∂θi−(

ργi

1 + γisi+

ρλγ2i |qi|

2

(1 + γisi)2

)∂si

∂θi

l′′

(θi) =

(λ∂2|qi|2

∂θ2i−∂2si

∂θ2i

)ργi

1 + γisi+

(∂si

∂θi

)2 2ρλγ3i |qi|

2

(1 + γisi)3

+

((∂si

∂θi

)2

− 2λ∂si

∂θi

∂|qi|2

∂θi− λ|qi|2

∂2si

∂θ2i

)ργ2i

(1 + γisi)2

∂si

∂θi= 2 Re

H(θi)B

−1−i∂ψ(θi)

∂θi

}∂|qi|2

∂θi= 2 Re

{qiy

HB−1−i∂ψ(θi)

∂θi

}∂2si

∂θ2i= 2 Re

H(θi)B

−1−i∂2ψ(θi)

∂θ2i+∂ψH(θi)

∂θiB−1−i∂ψ(θi)

∂θi

}∂2|qi|2

∂θ2i= 2 Re

{yHB−1−i

[∂ψ(θi)

∂θi

∂ψH(θi)

∂θi+∂2ψ(θi)

∂θ2iψ

H(θi)

]B−1−iy

}.

For Fourier atoms we have ∂ψ(θi)

∂θi= Dψ(θi) and ∂2ψ(θi)

∂θ2i

= D2ψ(θi), where

D = diag([0, j2π, · · · , j2π(N − 1)]).

REFERENCES

[1] T. Strohmer, “Measure what should be measured: Progress and chal-lenges in compressive sensing,” IEEE Signal Process. Lett., vol. 19, pp.887–893, Dec. 2012.

[2] M. Duarte and R. Baraniuk, “Spectral compressive sensing,” Appl. andComputational Harmonic Anal., vol. 35, pp. 111–129, Jul. 2013.

[3] A. Fannjiang and W. Liao, “Coherence pattern-guided compressivesensing with unresolved grids,” SIAM J. Imaging Sci., vol. 5, pp. 179–202, 2012.

[4] M. Elad, P. Milanfar, and R. Rubinstein, “Analysis versus synthesis insignal priors,” Inverse problems, vol. 23, no. 3, p. 947, Jun. 2007.

[5] E. J. Candes, Y. C. Eldar, D. Needell, and P. Randall, “Compressed sens-ing with coherent and redundant dictionaries,” Appl. and ComputationalHarmonic Anal., vol. 31, pp. 59–73, Jul. 2011.

[6] E. J. Candes and C. Fernandez-Granda, “Super-resolution from noisydata,” J. of Fourier Anal. and Applicat., vol. 19, no. 6, pp. 1229–1254,Dec. 2013.

[7] B. N. Bhaskar, G. Tang, and B. Recht, “Atomic norm denoising withapplications to line spectral estimation,” IEEE Trans. Signal Process.,vol. 61, no. 23, pp. 5987–5999, Dec. 2013.

[8] C. Ekanadham, D. Tranchina, and E. P. Simoncelli, “Recovery of sparsetranslation-invariant signals with continuous basis pursuit,” IEEE Trans.Signal Process., vol. 59, no. 10, pp. 4735–4744, Oct. 2011.

[9] K. Fyhn, M. F. Duarte, and S. H. Jensen, “Compressive parameter esti-mation for sparse translation-invariant signals using polar interpolation,”2013, submitted to IEEE Trans. Signal Process., arXiv:1305.3483.

[10] P. Stoica, “List of references on spectral line analysis,” Signal Process.,vol. 31, no. 3, pp. 329–340, Apr. 1993.

[11] S.-Y. Kung, K. S. Arun, and B. D. V. Rao, “State-space and singular-value decomposition-based approximation methods for the harmonicretrieval problem,” J. of the Optical Soc. of America, vol. 73, pp. 1799–1811, Dec. 1983.

[12] R. Roy and T. Kailath, “ESPRIT - estimation of signal parameters viarotational invariance techniques,” IEEE Trans. Acoust., Speech, SignalProcess., vol. 37, pp. 984–995, Jul. 1989.

[13] T. L. Hansen, P. B. Jørgensen, N. L. Pedersen, C. N. Manchon, and B. H.Fleury, “Bayesian compressed sensing with unknown measurement noiselevel,” in Asilomar Conf. Signals, Syst., and Computers, Nov. 2013.

[14] M. E. Tipping and A. Faul, “Fast marginal likelihood maximisation forsparse Bayesian models,” in Proc. 9th Int. Workshop Artificial Intell.and Stat., Jan. 2003, pp. 3–6.

[15] N. L. Pedersen, C. N. Manchon, M.-A. Badiu, D. Shutin, and B. H.Fleury, “Sparse estimation using Bayesian hierarchical prior modelingfor real and complex models,” 2013, submitted to IEEE Trans. SignalProcess., arXiv:1108.4324.

[16] S. Ji, D. Dunson, and L. Carin, “Multitask compressive sensing,” IEEETrans. Signal Process., vol. 57, no. 1, pp. 92–106, Jan. 2009.

[17] D. Shutin, W. Wang, and T. Jost, “Incremental sparse Bayesian learningfor parameter estimation of superimposed signals,” in Proc. 10th Int.Conf. Sampling Theory and Applicat. (SampTA), Jul. 2013.

[18] D. P. Wipf, B. D. Rao, and S. Nagarajan, “Latent variable Bayesianmodels for promoting sparsity,” IEEE Trans. Inf. Theory, vol. 57, no. 9,pp. 6236–6255, Sep. 2011.

[19] P. Stoica and N. Arye, “MUSIC, maximum likelihood, and Cramer-Raobound,” IEEE Trans. Acoust., Speech, Signal Process., vol. 37, no. 5,pp. 720–741, May 1989.