14
Bayesian Estimation of the Multifractality Parameter for Image Texture Using a Whittle Approximation S. Combrexelle, H. Wendt * , Member, IEEE, N. Dobigeon, Senior Member, IEEE, J.-Y. Tourneret, Senior Member, IEEE, S. McLaughlin, Fellow, IEEE, P. Abry, Fellow, IEEE *†‡§ Abstract Texture characterization is a central element in many image processing applications. Multifractal analysis is a useful signal and image processing tool, yet, the accurate estimation of multifractal param- eters for image texture remains a challenge. This is due in the main to the fact that current estima- tion procedures consist of performing linear regres- sions across frequency scales of the two-dimensional (2D) dyadic wavelet transform, for which only a few such scales are computable for images. The strongly non-Gaussian nature of multifractal processes, com- bined with their complicated dependence structure, makes it difficult to develop suitable models for pa- rameter estimation. Here, we propose a Bayesian procedure that addresses the difficulties in the esti- mation of the multifractality parameter. The orig- inality of the procedure is threefold: The construc- tion of a generic semi-parametric statistical model for the logarithm of wavelet leaders; the formula- tion of Bayesian estimators that are associated with this model and the set of parameter values admitted by multifractal theory; the exploitation of a suitable Whittle approximation within the Bayesian model which enables the otherwise infeasible evaluation of the posterior distribution associated with the model. Performance is assessed numerically for several 2D multifractal processes, for several image sizes and a large range of process parameters. The proce- dure yields significant benefits over current bench- mark estimators in terms of estimation performance * This work was supported by ANR BLANC 2011 AMATIS BS0101102 and ANR Project Hypanema ANR-12-BS03-003. S. Com- brexelle was supported by the Direction G´ en´ erale de l’Armement (DGA). S. McLaughlin acknowledges the support of EPSRC grant number EP/J015180/1. S. Combrexelle, N. Dobigeon, J.-Y. Tourneret and H. Wendt are with IRIT Laboratory, INP-ENSEEIHT, University of Toulouse, CNRS, Toulouse, France (email: fi[email protected]) S. McLaughlin is with the School of Engineering and Phys- ical Sciences, Heriot-Watt University, Edinburgh, UK (email: [email protected]) § P. Abry is with the Physics Dept., Ecole Normale Sup´ erieure de Lyon, CNRS, France (email: [email protected]) and ability to discriminate between the two most commonly used classes of multifractal process mod- els. The gains in performance are particularly pro- nounced for small image sizes, notably enabling for the first time the analysis of image patches as small as 64 × 64 pixels. Index terms— Texture characterization, Multifractal analysis, Wavelet leaders, Bayesian estimation, Whittle ap- proximation, Multiplicative cascades, Fractional Brownian motion 1 Introduction 1.1 Context and motivation Since the early origins of digital image processing, texture has been recognized as one of the central characteristic fea- tures in images. There is no common definition for tex- ture, and different paradigms have been introduced in the literature [24]. Several authors have proposed to model tex- ture using random fractals, scale invariance or self-similarity [29, 43]. Indeed, it has been reported in the literature that scale invariant processes are relevant and effective models for textures associated with a large class of natural images, see, e.g., [16, 59, 63]. The concepts of scale invariance and self-similarity are deeply tied to the degree of pointwise singular behavior or local regularity of the image amplitudes [39, 45]. It has long been recognized that multiscale and wavelet analyzes consti- tute ideal tools to study data regularity [1, 25, 41, 45, 58]. It is therefore not surprising that these tools play a central role not only for the study of image contours (edges), but also for texture characterization [18,37,51]. Yet, while contours are essentially isolated singularities, the texture models con- sist of densely interwoven sets of singularities of different regularity strength. Multifractal analysis provides a mathe- matical framework for the study of such spatial fluctuations of local regularity and texture characterization is therefore nowadays often conducted using this tool [7,31]. Multifractal analysis. The local regularity of an image X is commonly measured using the so-called H¨olderex- ponent h(t) [25, 45]. Qualitatively, the smaller h(t 0 ), the rougher X is at spatial location t 0 and the larger h(t 0 ), the 1 arXiv:1410.4871v2 [physics.data-an] 9 Apr 2015

Bayesian Estimation of the Multifractality Parameter for ... · ML and Bayesian estimation. Examples of ML and Bayesian estimators for 1D fBm formulated in the spectral or wavelet

  • Upload
    others

  • View
    14

  • Download
    0

Embed Size (px)

Citation preview

Page 1: Bayesian Estimation of the Multifractality Parameter for ... · ML and Bayesian estimation. Examples of ML and Bayesian estimators for 1D fBm formulated in the spectral or wavelet

Bayesian Estimation of the Multifractality Parameter for

Image Texture Using a Whittle Approximation

S. Combrexelle, H. Wendt∗, Member, IEEE, N. Dobigeon, Senior Member, IEEE,J.-Y. Tourneret, Senior Member, IEEE, S. McLaughlin, Fellow, IEEE, P. Abry, Fellow, IEEE ∗†‡§

Abstract

Texture characterization is a central element inmany image processing applications. Multifractalanalysis is a useful signal and image processing tool,yet, the accurate estimation of multifractal param-eters for image texture remains a challenge. Thisis due in the main to the fact that current estima-tion procedures consist of performing linear regres-sions across frequency scales of the two-dimensional(2D) dyadic wavelet transform, for which only a fewsuch scales are computable for images. The stronglynon-Gaussian nature of multifractal processes, com-bined with their complicated dependence structure,makes it difficult to develop suitable models for pa-rameter estimation. Here, we propose a Bayesianprocedure that addresses the difficulties in the esti-mation of the multifractality parameter. The orig-inality of the procedure is threefold: The construc-tion of a generic semi-parametric statistical modelfor the logarithm of wavelet leaders; the formula-tion of Bayesian estimators that are associated withthis model and the set of parameter values admittedby multifractal theory; the exploitation of a suitableWhittle approximation within the Bayesian modelwhich enables the otherwise infeasible evaluation ofthe posterior distribution associated with the model.Performance is assessed numerically for several 2Dmultifractal processes, for several image sizes anda large range of process parameters. The proce-dure yields significant benefits over current bench-mark estimators in terms of estimation performance

∗This work was supported by ANR BLANC 2011 AMATISBS0101102 and ANR Project Hypanema ANR-12-BS03-003. S. Com-brexelle was supported by the Direction Generale de l’Armement(DGA). S. McLaughlin acknowledges the support of EPSRC grantnumber EP/J015180/1.†S. Combrexelle, N. Dobigeon, J.-Y. Tourneret and H. Wendt

are with IRIT Laboratory, INP-ENSEEIHT, University of Toulouse,CNRS, Toulouse, France (email: [email protected])‡S. McLaughlin is with the School of Engineering and Phys-

ical Sciences, Heriot-Watt University, Edinburgh, UK (email:[email protected])§P. Abry is with the Physics Dept., Ecole Normale Superieure de

Lyon, CNRS, France (email: [email protected])

and ability to discriminate between the two mostcommonly used classes of multifractal process mod-els. The gains in performance are particularly pro-nounced for small image sizes, notably enabling forthe first time the analysis of image patches as smallas 64 × 64 pixels.

Index terms— Texture characterization, Multifractalanalysis, Wavelet leaders, Bayesian estimation, Whittle ap-proximation, Multiplicative cascades, Fractional Brownianmotion

1 Introduction

1.1 Context and motivation

Since the early origins of digital image processing, texturehas been recognized as one of the central characteristic fea-tures in images. There is no common definition for tex-ture, and different paradigms have been introduced in theliterature [24]. Several authors have proposed to model tex-ture using random fractals, scale invariance or self-similarity[29, 43]. Indeed, it has been reported in the literature thatscale invariant processes are relevant and effective modelsfor textures associated with a large class of natural images,see, e.g., [16, 59,63].The concepts of scale invariance and self-similarity aredeeply tied to the degree of pointwise singular behavior orlocal regularity of the image amplitudes [39,45]. It has longbeen recognized that multiscale and wavelet analyzes consti-tute ideal tools to study data regularity [1, 25, 41, 45, 58]. Itis therefore not surprising that these tools play a central rolenot only for the study of image contours (edges), but alsofor texture characterization [18, 37, 51]. Yet, while contoursare essentially isolated singularities, the texture models con-sist of densely interwoven sets of singularities of differentregularity strength. Multifractal analysis provides a mathe-matical framework for the study of such spatial fluctuationsof local regularity and texture characterization is thereforenowadays often conducted using this tool [7, 31].Multifractal analysis. The local regularity of an imageX is commonly measured using the so-called Holder ex-ponent h(t) [25, 45]. Qualitatively, the smaller h(t0), therougher X is at spatial location t0 and the larger h(t0), the

1

arX

iv:1

410.

4871

v2 [

phys

ics.

data

-an]

9 A

pr 2

015

Page 2: Bayesian Estimation of the Multifractality Parameter for ... · ML and Bayesian estimation. Examples of ML and Bayesian estimators for 1D fBm formulated in the spectral or wavelet

smoother X is at t0. The goal of multifractal analysis is theestimation of the multifractal spectrum D(h), which pro-vides a global description of the spatial fluctuations of h(t).It is defined as the collection of the fractal dimensions ofthe sets of points for which the Holder exponent takes thesame value [25, 45], cf., Section 2.1. Multifractal analysishas recently matured into a standard image processing tooland has been successfully used in a large number of appli-cations including texture classification [59, 63], biomedicalapplications [11, 30], physics [44, 48] and art investigation[2, 19,27,28].In most applications, the estimation of D(h) cannot bebased directly on its definition [25]. Instead, a so-called mul-tifractal formalism is constructed based on multiresolutioncoefficients TX(a,k), essentially capturing the content of theimage X around the discrete spatial location k for a givenfrequency scale a = 2j . Examples are given by increments,wavelet coefficients and more recently wavelet leaders `(j,k)[25] (defined in Section 2.2), which yield the current bench-mark multifractal formalism. The multifractal formalismprovides an expansion of the multifractal spectrum of theimage X in terms of the so-called log-cumulants cp, p ≥ 1[26,58]

D(h) = 2 +c22!

(h− c1c2

)2

+−c33!

(h− c1c2

)3

+−c4 + 3c3

2/c24!

(h− c1c2

)4

+ . . . (1)

when c2 < 0, while D(h) = δ(h−c1) when c2 ≡ 0 (c2 cannotbe positive theoretically [15,25,45]).Estimation of c2. The leading order coefficients cp pro-vide a relevant summary of the multifractal properties ofX in applications where it would often not be convenientto handle an entire function D(h) [15, 26, 55, 58]. The firstlog-cumulant c1, for instance, is the mode of D(h) and canbe read as a measure for the “average” smoothness of X.More importantly, the coefficient c2, referred to as the mul-tifractality or intermittency parameter, is directly related tothe width of D(h) and captures the multifractal signature(i.e., the fluctuations of the local regularity) of the image X.Its primary importance stems from the fact that it enablesthe identification of the two major classes of multifractalstochastic processes: self-similar processes for which c2 = 0and multifractal multiplicative cascade (MMC) based pro-cesses for which c2 is strictly negative [57]. While the formerclass is tied deeply to additive constructions, the latter isbased on multiplicative constructions and is hence linked tofundamentally different physical principles [22,39,45]. More-over, the magnitude of c2 quantifies the degree of multi-fractality of an image for the latter class. For an overviewand details on scale invariant and multifractal processes, thereader is referred to, e.g., [33, 45] and references therein.

In the seminal contribution [15], it has been shownthat the log-cumulants cp are tied to the quantities `(j,k)

through the key relation Cump[log `(j,k)] = c0p + cp log 2j ,where Cump[·] is the p-th order cumulant. In particular

C2(j) , Var [log `(j,k)] = c02 + c2 log 2j . (2)

Relation (2) leads to the definition of the current standardand benchmark estimator for the parameter c2, based onlinear regression of the sample variance, denoted by Var, oflog `(j,k) over a range of scales j ∈ [j1, j2]

c2 =1

log 2

j2∑

j=j1

wjVar[log `(j, ·)] (3)

where wj are suitably defined regression weights [55,58].Limitations. The use of multifractal analysis remains re-stricted to images of relatively large size (of order 5122 pix-els) because a sufficient number of scales j must be availableto perform the linear regression (3). While a similar issue isencountered for the analysis of 1D signals, it is significantlymore severe for images: indeed, modulo border effects ofthe wavelet transform, the number of available scales is pro-portional to the logarithm of the number of samples for 1Dsignals and to the logarithm of the square root of the num-ber of pixels for an image. For instance, for a 1D signal with256×256 = 65536 samples, j2 = 13 or 14 scales can be com-puted, while j2 = 4 or 5 for an image of N ×N = 256× 256pixels. In addition, the finest scale, j = 1, should not beused in (3), see, e.g., [53]. The practical consequences forthe multifractal analysis of images are severe: First, imagesof small size and thus image patches cannot be analyzed inpractice. Second, (3) yields modest performance for imageswhen compared with 1D signals of equivalent sample size[58], making it difficult to discriminate between c2 ≡ 0 andvalues c2 < 0 that are encountered in applications (typically,c2 lies between −0.01 and −0.08). The goal of this work isto propose and validate a novel procedure for the estimationof c2 for images that addresses these difficulties.

1.2 Related works

There are a limited number of reports in the literature thatattempt to overcome the limitations of multifractal analy-sis for images described above. The generalized method ofmoments has been proposed and studied in, e.g., [9, 35, 36]and formulates parameter inference as the solution (in theleast squares sense) of an over-determined system of equa-tions that are derived from the moments of the data. Themethod depends strongly on fully parametric models andyields, to the best of our knowledge, only limited benefits inpractical applications.

Although classical in parameter inference, maximum like-lihood (ML) and Bayesian estimation methods have mostlybeen formulated for a few specific self-similar and multi-fractal processes [12, 62]. The main reason for this lies inthe complex statistical properties of most of these processes,

2

Page 3: Bayesian Estimation of the Multifractality Parameter for ... · ML and Bayesian estimation. Examples of ML and Bayesian estimators for 1D fBm formulated in the spectral or wavelet

which exhibit marginal distributions that are strongly non-Gaussian as well as intricate algebraically decaying depen-dence structures that remain poorly studied to date. Thesame remark is true for their wavelet coefficients and waveletleaders, see, e.g., [42, 52].

One exception is given by the fractional Brownian mo-tion (in 1D) and fractional Brownian fields (in 2D) (fBm),that are jointly Gaussian self-similar (i.e., c2 ≡ 0) processeswith fully parametric covariance structure appropriate forML and Bayesian estimation. Examples of ML and Bayesianestimators for 1D fBm formulated in the spectral or waveletdomains can be found in [12, 17, 40, 62]. For images, an MLestimator has been proposed in [34] (note, however, that theestimation problem is reduced to a univariate formulationfor the rows / columns of the image there).

As far as MMC processes are concerned, [32] proposesan ML approach in the time domain for one specific pro-cess. However, the method relies strongly on the particularconstruction of this process and cannot easily accommodatemore general model classes. Moreover, the method is for-mulated for 1D signals only. Finally, a Bayesian estimationprocedure for the parameter c2 of multifractal time serieshas recently been proposed in [56]. Unlike the methodsmentioned above, it does not rely on specific assumptionsbut instead employs a heuristic semi-parametric model forthe statistics of the logarithm of wavelet leaders associatedwith univariate MMC processes. Yet, it is designed for andcan only be applied to univariate time series of small samplesize.

1.3 Goals and contributions

The use of fully parametric models for the data can be veryrestrictive in many real-world applications. Therefore, thegoal and the main contribution of this work is to study aBayesian estimation procedure for the multifractality pa-rameter c2 with as few as possible assumptions on the data(essentially, the relation (2)) that can actually be applied toreal-world images of small as well as large sizes. To this end,we adopt a strategy that is inspired by [56] and develop thekey elements that are required for its formulation for images.

First, we show by means of numerical simulations that thedistribution of the logarithm of wavelet leaders log `(j,k)of 2D MMC processes can, at each scale j, be well ap-proximated by a multivariate Gaussian distribution. In-spired by the covariance properties induced by the multi-plicative nature of cascade constructions, we propose a newgeneric radial symmetric model for the variance-covarianceof this distribution. This second-order statistical model isparametrized only by the two parameters c2 and c02 in (2)and enables us to formulate estimation in a Bayesian frame-work.

Second, we formulate a Bayesian estimation procedure forthe parameter c2 of images that permits to take into account

the constraints that are associated with the proposed statis-tical model. To this end, an appropriate prior distributionis assigned to the parameter vector (c2, c

02) which essentially

ensures that the variance (2) is positive. Additional priorinformation, if available, can easily be incorporated. TheBayesian estimators of c2 associated with the posterior dis-tribution of interest cannot be evaluated directly becauseof the constraints that the parameter vector (c2, c

02) has to

satisfy. Therefore, we design a suitable Markov chain MonteCarlo (MCMC) algorithm that generates samples that areasymptotically distributed according to the posterior dis-tribution of interest. These samples are in turn used toapproximate the Bayesian estimators. More precisely, wepropose a random-walk Metropolis-Hastings scheme to ex-plore efficiently the posterior distribution according to theadmissible set of values for c2 and c02.

Finally, the exact evaluation of the likelihood associatedwith the proposed model for the log-wavelet leaders requiresthe computation of the inverse and the determinant of largedense matrices, which is numerically and computationallytoo demanding for practical applications. To obtain a sta-ble and efficient algorithm that can actually be applied toimages, following intuitions developed in the univariate case,cf. e.g., [12], we approximate the exact likelihood witha Whittle-type expansion that is adapted to the proposedmodel and can be efficiently evaluated in the spectral do-main.

The proposed algorithm for the estimation of the multi-fractality parameter c2 is effective both for small and largeimage sizes. Its performance is assessed numerically bymeans of Monte Carlo simulations for two classical and rep-resentative 2D MMC constructions, the canonical Mandel-brot cascades (CMC) [39] and compound Poisson cascades(CPC) [10], using the most common multipliers, and a largerange of process parameters and sample sizes from 64 × 64to 512×512 pixels. Complementary results are provided for2D fBms (that are self-similar but not MMC). Our resultsindicate that the proposed estimation procedure is robustwith respect to different choices of process constructions andgreatly outperforms (2), in particular for small images andfor identifying a value c2 ≡ 0. It enables, for the first time,a multifractal analysis of images (or image patches) whosesizes are as small as 64× 64 pixels.

The remainder of this work is organized as follows. Sec-tion 2 summarizes the main concepts of multifractal analysisand the wavelet leader multifractal formalism. Section 3 in-troduces the statistical model and the Bayesian frameworkunderlying the estimation procedure for the parameter c2 ofimages, which is formulated in Section 4. Numerical resultsare given in Section 5. In Section 6, the proposed procedureis applied to the patch-wise analysis of a real-world image,illustrating its potential benefits for practical applications.Finally, Section 7 concludes this paper and presents somefuture work.

3

Page 4: Bayesian Estimation of the Multifractality Parameter for ... · ML and Bayesian estimation. Examples of ML and Bayesian estimators for 1D fBm formulated in the spectral or wavelet

2 Multifractal analysis of Images

Let X : R2 → R denote the 2D function (image) to beanalyzed. The image X is assumed to be locally boundedin what follows (see Section 2.2 for a practical solution tocircumvent this prerequisite).

2.1 Multifractal analysis

Holder exponent. Multifractal analysis aims at charac-terizing the image X in terms of the fluctuations of its localregularity, characterized by the so-called Holder exponent,which is defined as follows [25, 45]. The image X is said tobelong to Cα(t0) if there exists α > 0 and a polynomial Pt0of degree smaller than α such that

||X(t)− Pt0(t)|| ≤ C||t− t0||α

where || · || is the Euclidian norm. The Holder exponent atposition t0 is the largest value of α such that this inequalityholds, i.e.,

h(t0) , sup{α : X ∈ Cα(t0)}. (4)

Multifractal spectrum. For large classes of stochas-tic processes, the Holder exponents h(t) can be theoreti-cally shown to behave in an extremely erratic way [25, 26].Therefore, multifractal analysis provides a global descriptionof the spatial fluctuations of h(t) in terms of the multifrac-tal spectrum D(h). It is defined as the Hausdorff dimension(denoted dimH) of the sets of points at which the Holderexponent takes the same value, i.e.,

D(h) , dimH

(Eh = {t : h(t = h}

). (5)

For more details on multifractal analysis and a precise defi-nition of the Hausdorff dimension, see, e.g., [25, 26].

2.2 Wavelet leader multifractal formalism

Historically, multifractal formalisms have been proposedbased on increments or wavelet coefficients. These choicesof multiresolution quantities lead to both theoretical andpractical limitations, see [55, 58] for a discussion. Recently,it has been shown that a relevant multifractal formalism canbe constructed from the wavelet leaders [25, 26, 55], whichare specifically tailored for this purpose.Wavelet coefficients. We assume that the image is givenin form of discrete sample values X(k), k = (k1, k2). Atwo-dimensional (2D) orthonormal discrete wavelet trans-form (DWT) can be obtained as the tensor product ofone-dimensional (1D) DWT as follows. Let G0(k) andG1(k) denote the low-pass and high-pass filters defininga 1D DWT. These filters are associated with a motherwavelet ψ, characterized by its number of vanishing mo-ments Nψ > 0. Four 2D filters G(m)(k), m = 0, . . . , 3 are

defined by tensor products of Gi, i = 1, 2. The 2D low-pass filter G(0)(k) , G0(k1)G0(k2) yields the approximation

coefficients D(0)X (j,k), whereas the high-pass filters defined

by G(1)(k) , G0(k1)G1(k2), G(2)(k) , G1(k1)G0(k2) andG(3)(k) , G1(k1)G1(k2) yield the wavelet (detail) coeffi-

cients D(m)X (j,k), m = 1, 2, 3 as follows: at the finest scale

j = 1, the D(m)X (j,k), m = 0, . . . , 3 are obtained by convolv-

ing the image X with G(m), m = 0, . . . , 3, and decimation;for the coarser scales j ≥ 2 they are obtained iteratively

by convolving G(m), m = 0, . . . , 3, with D(0)X (j − 1, ·) and

decimation. For scaling and multifractal analysis purposes,

the approximation coefficients D(0)X are discarded and it is

common to normalize the wavelet coefficients according tothe L1-norm

d(m)X (j,k) , 2−jD(m)

X (j,k), m = 1, 2, 3 (6)

so that they reproduce the self-similarity exponent for self-similar processes [7]. For a formal definition and details on(2D) wavelet transforms, the reader is referred to [5, 37].Wavelet leaders. Denote as

λj,k = {[k12j , (k1 + 1)2j), [k22j , (k2 + 1)2j)}

the dyadic cube of side length 2j centered at k2j and

3λj,k =⋃

n1,n2∈{−1,0,1}λj,k1+n1,k2+n2

the union of this cube with its eight neighbors. The waveletleaders are defined as the largest wavelet coefficient magni-tude within this neighborhood over all finer scales [25]

`(j,k) ≡ `(λj,k) , supm∈(1,2,3),λ′⊂3λj,k

|d(m)X (λ′)|. (7)

Wavelet leaders reproduce the Holder exponent as follows

h(t0) = lim infj→−∞

(log `(λj,k(t0))

/log 2j

)(8)

where λj,k(t0) denotes the cube at scale j including the spa-tial location t0 [25]. It has been shown that (8) is the theo-retical key property required for constructing a multifractalformalism, see [25] for details. In particular, it can be shownthat the wavelet leader multifractal formalism (WLMF), i.e.,the use of (1) with coefficients cp estimated using waveletleaders, is valid for large classes of multifractal model pro-cesses, see [55, 58] for details and discussions. The WLMFhas been extensively studied both theoretically and in termsof estimation performance and constitutes the benchmarktool for performing multifractal analysis, cf. e.g., [55, 58].Negative regularity. The WLMF can be applied tolocally bounded images (equivalently, to images with strictlypositive uniform regularity) only, see [3, 55, 58] for precisedefinitions and for procedures for assessing this condition inpractice. However, it has been reported that a large number

4

Page 5: Bayesian Estimation of the Multifractality Parameter for ... · ML and Bayesian estimation. Examples of ML and Bayesian estimators for 1D fBm formulated in the spectral or wavelet

of real-world images do not satisfy this prerequisite [58,59].In these cases, a practical solution consists of constructingthe WLMF using the modified wavelet coefficients

d(m),αX (j,k) , 2αjd

(m)X (j,k), α > 0 (9)

instead of d(m)X in (7). When α is chosen sufficiently large,

the WLMF holds (see [58] for details about the theoreticaland practical consequences implied by this modification).

Finally, note that the above analysis as well as the WLMFare meaningful for homogeneous multifractal functions X,for which the multifractal spectra D(h) of different subsetsof t are identical. This excludes the class of multifractionalmodels [8, 64], for which the function h(t) is given by asmooth non-stationary evolution. Such models, also of in-terest in other application contexts, are not considered here,as the focus is on multifractality parameter c2 which is notrelevant to characterize multifractional processes.

3 Bayesian framework

In this section, a novel empirical second-order statisticalmodel for the logarithm of wavelet leaders for 2D MMC pro-cesses is proposed. This model is the key tool for estimatingthe multifractality parameter c2 in a Bayesian framework.

3.1 Modeling the statistics of log-waveletleaders

Marginal distribution model. It has recently been ob-served that for 1D signals the distribution of the log-waveletleaders

l(j,k) , log `(j,k) (10)

can be reasonably well approximated by a Gaussian distri-bution [56]. Here, we numerically investigate the marginaldistributions of l(j, ·) for 2D images. To this end, a repre-sentative selection of scaling processes (the MMC processesCMC-LN, CMC-LP, CPC-LN and CPC-LP, as well as fBm,where LN stands for log-Normal and LP for log-Poisson, re-spectively) have been analyzed for a wide range of processparameters (see Section 5.1 for a description of these pro-cesses). Representative examples of quantile-quantile plotsof the standard Normal distribution against empirical distri-butions of log-wavelet leaders (scale j = 2) associated withCPC-LN, CPC-LP and fBm are plotted in Fig. 1 (upperrow).

Clearly, the normal distribution provides, within ±3standard deviations, a reasonable approximation for themarginal distribution of log-wavelet leaders of images forboth members of the MMC class. It is also the case for thefBm, a Gaussian self-similar process that is not a memberof MMC. Note that the fact that the marginal distributionsof the log-wavelet leaders are approximately Gaussian forscale invariant processes confirms the intuitions formulated

of estimation performance and constitutes the benchmarktool for performing multifractal analysis, cf. e.g., [55, 58].Negative regularity. The WLMF can be applied tolocally bounded images (equivalently, to images with strictlypositive uniform regularity) only, see [3, 55, 58] for precisedefinitions and for procedures for assessing this condition inpractice. However, it has been reported that a large numberof real-world images do not satisfy this prerequisite [58,59].In these cases, a practical solution consists of constructingthe WLMF using the modified wavelet coe�cients

d(m),↵X (j, k) , 2↵jd

(m)X (j, k), ↵ > 0 (9)

instead of d(m)X in (7). When ↵ is chosen su�ciently large,

the WLMF holds (see [58] for details about the theoreticaland practical consequences implied by this modification).

Finally, note that the above analysis as well as the WLMFare meaningful for homogeneous multifractal functions X,for which the multifractal spectra D(h) of di↵erent subsetsof t are identical. This excludes the class of multifractionalmodels [8, 64], for which the function h(t) is given by asmooth non-stationary evolution. Such models, also of in-terest in other application contexts, are not considered here,as the focus is on multifractality parameter c2 which is notrelevant to characterize multifractional processes.

3 Bayesian framework

In this section, a novel empirical second-order statisticalmodel for the logarithm of wavelet leaders for 2D MMC pro-cesses is proposed. This model is the key tool for estimatingthe multifractality parameter c2 in a Bayesian framework.

3.1 Modeling the statistics of log-waveletleaders

Marginal distribution model. It has recently been ob-served that for 1D signals the distribution of the log-waveletleaders

l(j, k) , log `(j, k) (10)

can be reasonably well approximated by a Gaussian distri-bution [56]. Here, we numerically investigate the marginaldistributions of l(j, ·) for 2D images. To this end, a repre-sentative selection of scaling processes (the MMC processesCMC-LN, CMC-LP, CPC-LN and CPC-LP, as well as fBm,where LN stands for log-Normal and LP for log-Poisson, re-spectively) have been analyzed for a wide range of processparameters (see Section 5.1 for a description of these pro-cesses). Representative examples of quantile-quantile plotsof the standard Normal distribution against empirical distri-butions of log-wavelet leaders (scale j = 2) associated withCPC-LN, CPC-LP and fBm are plotted in Fig. 1 (upperrow).

−4

−3

−2

−1

0

1

l(2,k)

−5 0 5−20

−15

−10

−5

0

5

ln|d

(3)

X(2,k

)|

−4

−3

−2

−1

0

1

2

−5 0 5−20

−15

−10

−5

0

5−5

−4.5

−4

−3.5

−3

−5 0 5−20

−15

−10

−5

0

Standard Normal Standard Normal Standard Normal

CPC - LN CPC - LP fBm

Figure 1: Quantile-quantile plots of the empirical distribu-tions of the log-wavelet leaders l(2, k) (top) and wavelet co-

e�cients log d(3)X (2, k) (bottom) against standard normal for

CPC-LN (left column) and CPC-LP (center column) withc2 = �0.04, respectively, and for fBm (right column).

Clearly, the normal distribution provides, within ±3standard deviations, a reasonable approximation for themarginal distribution of log-wavelet leaders of images forboth members of the MMC class. It is also the case for thefBm, a Gaussian self-similar process that is not a memberof MMC. Note that the fact that the marginal distributionsof the log-wavelet leaders are approximately Gaussian forscale invariant processes confirms the intuitions formulatedby Mandelbrot [38]. However, it is not a trivial finding:There is no a priori reason for this property even if theanalyzed stochastic process has log-normal marginals (as isthe case for CMC-LN, for instance). Indeed, it is not thecase for the logarithm of the absolute value of wavelet coef-ficients whose marginal distributions are significantly morecomplicated and strongly depart from Gaussian, cf., Fig. 1(bottom row).

Variance-covariance model. We introduce a modelfor the covariance of the logarithm of 2D wavelet lead-ers for MMC processes at fixed scale j denotated asCov[l(j, k), l(j, k + �k)]. It is motivated by the asymp-totic covariance of the logarithm of multiscale quantitiesgenerically associated with multiplicative construction (c.f.[39]), studied in detail for wavelet coe�cients of 1D randomwavelet cascades in [6], and also by recent numerical resultsobtained for the covariance of the logarithm of 1D waveletleaders for MMC processes [56]. These results suggest alinear decay of Cov[l(j, k), l(j, k + �k)] in log coordinateslog�k, with slope given by the parameter c2. Numericalsimulations with 2D MMC processes for a wide range of pro-cess parameters (detailed in Section 5.1) indicate that theempirical intra-scale covariance is radially symmetric anddecays as c2 log�r with �r , ||�k|| for an intermediary

5

Figure 1: Quantile-quantile plots of the empirical distribu-tions of the log-wavelet leaders l(2,k) (top) and wavelet co-

efficients log d(3)X (2,k) (bottom) against standard normal for

CPC-LN (left column) and CPC-LP (center column) withc2 = −0.04, respectively, and for fBm (right column).

0

00

0.05

0.1

0.15

0.2

∆k1∆k2

Sample covariance

Cov[l(2,k

),l(2,k

+∆k)]

(a) Sample covariance

0

0 ∆k1∆k2

Model ρj(∆r;θ)

(b) Radial model

0 2 4 60

0.1

0.2

0.3

log2 (∆r + 1)

∆k2 = ∆k1

Cov j = 2Model j = 2Cov j = 3Model j = 3

(c) Slice �k2 = �k1

0 2 4 6log2 (∆r + 1)

∆k2 = 0

Cov j = 2Model j = 2Cov j = 3Model j = 3

(d) Slice �k2 = 0

Figure 2: Fitting between the sample covariance (a), aver-aged on 100 realizations of CMC-LN ([N, c2] = [29,�0.04]),and the parametric covariance (b); (c) and (d) compare themodel (blue) and the sample covariance (red) for two slices.

range of values �r given by 3 < �r �rmaxj

Cov[l(j, k), l(j, k + �k)] ⇡%(1)j (�r; c2) , � + c2(log2 �r + j) log 2 (11)

where �rmaxj =

p2(p

nj � 1) and nj ⇡ bN2/22jc denotes

the number of wavelet leaders at scale 2j of an N⇥N image.The constant � is found to be well approximated by using

the heuristic condition %(1)j (bpnj/4c; c2)) = 0, where the

operator b c truncates to integer values.

The theoretical variance of the log-wavelet leaders is givenby C2(j) = C2(j; c2, c

02) defined in (2). Finally, the short-

term covariance is modeled as a line connecting C2(j; c2, c02)

at �r = 0 and %(1)j (�r; c2) at �r = 3 as follows

%(0)j (�r; c2, c

02) ,

log(�r + 1)

log 4[%

(1)j (3; c2) � C2(j; c2, c

02)] + C2(j; c2, c

02). (12)

Combining (2), (11) and (12) yields the following full modelfor the covariance, parametrized by two parameters ✓ =[c2, c

02]

T only

%j(�r;✓) =

8><>:

C2(j; c2, c02) �r = 0

%(0)j (�r; c2, c

02) 0 �r 3

max(0, %(1)j (�r; c2)) 3 �r �rmax

j .

(13)

Here, only the positive portions of %(1)j are considered for

numerical reasons (conditioning of the covariance matrix).The proposed covariance model is illustrated in Fig. 2 forCMC-LN.

The joint Gaussian model with covariance model (13) as-sumes limited information on the dependence between di↵er-ent scales, essentially the variance (2). The correspondingcovariance matrix model for log-wavelet leaders at severalscales j 2 [j1, j2] has thus block-diagonal structure. Forconvenience and without loss of generality, the formulationsgiven below and in Section 4 will be stated in block-diagonalform yet could be extended without di�culty to any othervalid covariance matrix model.

3.2 Likelihood, prior and posterior distri-butions

We focus on the estimation of the parameter c2 and thereforework with centered log-wavelet leaders below. Let lj denote

the vector of the nj centered coe�cients l(j, k) � bE[lX(j, .)]at scale j 2 [j1, j2], organized in lexicographic order, wherebE[·] stands for the sample mean. Let ⌃j(✓) denote thecorresponding nj ⇥ nj covariance matrix whose entries aregiven by the 2D parametric covariance function model (13).For convenience of notation, all coe�cients are stacked in aunique zero-mean vector L = [lTj1 , ..., l

Tj2 ]

T .Likelihood. With the above notation and assumptions,the likelihood of lj is given by

p(lj |✓) ,exp

⇣� 1

2 lTj ⌃j(✓)�1lj

p(2⇡)nj det⌃j(✓)

. (14)

Using the independence between l(j, k) and l(j0, k0) for j 6=j0, the likelihood of L is given by

p(L|✓) =

j2Y

j=j1

p(lj |✓). (15)

Prior distribution. The parameter vector ✓ must be cho-sen such that the variances of l(j, k) are positive, C2(j) � 0.We define the admissible set

A = (A+ [ A�) \ Am (16)

where

A� = {(c2, c02) 2 R2 | c2 < 0 and c0

2 + c2 j2ln 2 > 0},A+ = {(c2, c

02) 2 R2 | c2 > 0 and c0

2 + c2 j1ln 2 > 0},

Am = {(c2, c02) 2 R2 | |c0

2| < c0,m2 , |c2| < cm

2 }

and cm2 , c0,m

2 quantify the largest admissible values for c2 andc02, parameters that need to be tuned by practitioners and

may depend on the application considered. When no addi-tional prior information is available regarding ✓, a uniformprior distribution on the set A is assigned to ✓

⇡(✓) = UA(✓) / 1A(✓). (17)

6

Figure 2: Fitting between the sample covariance (a), aver-aged on 100 realizations of CMC-LN ([N, c2] = [29,−0.04]),and the parametric covariance (b); (c) and (d) compare themodel (blue) and the sample covariance (red) for two slices.

by Mandelbrot [38]. However, it is not a trivial finding:There is no a priori reason for this property even if theanalyzed stochastic process has log-normal marginals (as isthe case for CMC-LN, for instance). Indeed, it is not thecase for the logarithm of the absolute value of wavelet coef-ficients whose marginal distributions are significantly morecomplicated and strongly depart from Gaussian, cf., Fig. 1(bottom row).

Variance-covariance model. We introduce a modelfor the covariance of the logarithm of 2D wavelet lead-

5

Page 6: Bayesian Estimation of the Multifractality Parameter for ... · ML and Bayesian estimation. Examples of ML and Bayesian estimators for 1D fBm formulated in the spectral or wavelet

ers for MMC processes at fixed scale j denotated asCov[l(j,k), l(j,k + ∆k)]. It is motivated by the asymp-totic covariance of the logarithm of multiscale quantitiesgenerically associated with multiplicative construction (c.f.[39]), studied in detail for wavelet coefficients of 1D randomwavelet cascades in [6], and also by recent numerical resultsobtained for the covariance of the logarithm of 1D waveletleaders for MMC processes [56]. These results suggest alinear decay of Cov[l(j, k), l(j, k + ∆k)] in log coordinateslog ∆k, with slope given by the parameter c2. Numericalsimulations with 2D MMC processes for a wide range of pro-cess parameters (detailed in Section 5.1) indicate that theempirical intra-scale covariance is radially symmetric anddecays as c2 log ∆r with ∆r , ||∆k|| for an intermediaryrange of values ∆r given by 3 < ∆r ≤ ∆rmax

j

Cov[l(j,k), l(j,k + ∆k)] ≈%(1)j (∆r; c2) , γ + c2(log2 ∆r + j) log 2 (11)

where ∆rmaxj =

√2(√nj − 1) and nj ≈ bN2/22jc denotes

the number of wavelet leaders at scale 2j of an N×N image.The constant γ is found to be well approximated by using

the heuristic condition %(1)j (b√nj/4c; c2)) = 0, where the

operator b c truncates to integer values.The theoretical variance of the log-wavelet leaders is given

by C2(j) = C2(j; c2, c02) defined in (2). Finally, the short-

term covariance is modeled as a line connecting C2(j; c2, c02)

at ∆r = 0 and %(1)j (∆r; c2) at ∆r = 3 as follows

%(0)j (∆r; c2, c

02) ,

log(∆r + 1)

log 4[%

(1)j (3; c2)− C2(j; c2, c

02)] + C2(j; c2, c

02). (12)

Combining (2), (11) and (12) yields the following full modelfor the covariance, parametrized by two parameters θ =[c2, c

02]T only

%j(∆r;θ) =

C2(j; c2, c02) ∆r = 0

%(0)j (∆r; c2, c

02) 0 ≤ ∆r ≤ 3

max(0, %(1)j (∆r; c2)) 3 ≤ ∆r ≤ ∆rmax

j .

(13)

Here, only the positive portions of %(1)j are considered for

numerical reasons (conditioning of the covariance matrix).The proposed covariance model is illustrated in Fig. 2 forCMC-LN.

The joint Gaussian model with covariance model (13) as-sumes limited information on the dependence between differ-ent scales, essentially the variance (2). The correspondingcovariance matrix model for log-wavelet leaders at severalscales j ∈ [j1, j2] has thus block-diagonal structure. Forconvenience and without loss of generality, the formulationsgiven below and in Section 4 will be stated in block-diagonalform yet could be extended without difficulty to any othervalid covariance matrix model.

3.2 Likelihood, prior and posterior distri-butions

We focus on the estimation of the parameter c2 and thereforework with centered log-wavelet leaders below. Let lj denote

the vector of the nj centered coefficients l(j,k)− E[lX(j, .)]at scale j ∈ [j1, j2], organized in lexicographic order, where

E[·] stands for the sample mean. Let Σj(θ) denote thecorresponding nj × nj covariance matrix whose entries aregiven by the 2D parametric covariance function model (13).For convenience of notation, all coefficients are stacked in aunique zero-mean vector L = [lTj1 , ..., l

Tj2 ]T .

Likelihood. With the above notation and assumptions,the likelihood of lj is given by

p(lj |θ) ,exp

(− 1

2 lTj Σj(θ)−1lj

)

√(2π)nj det Σj(θ)

. (14)

Using the independence between l(j,k) and l(j′,k′) for j 6=j′, the likelihood of L is given by

p(L|θ) =

j2∏

j=j1

p(lj |θ). (15)

Prior distribution. The parameter vector θ must be cho-sen such that the variances of l(j,k) are positive, C2(j) ≥ 0.We define the admissible set

A = (A+ ∪ A−) ∩ Am (16)

where

A− = {(c2, c02) ∈ R2 | c2 < 0 and c02 + c2 j2ln 2 > 0},A+ = {(c2, c02) ∈ R2 | c2 > 0 and c02 + c2 j1ln 2 > 0},Am = {(c2, c02) ∈ R2 | |c02| < c0,m2 , |c2| < cm2 }

and cm2 , c0,m2 quantify the largest admissible values for c2 and

c02, parameters that need to be tuned by practitioners andmay depend on the application considered. When no addi-tional prior information is available regarding θ, a uniformprior distribution on the set A is assigned to θ

π(θ) = UA(θ) ∝ 1A(θ). (17)

Posterior distribution and Bayesian estimators.The posterior distribution of θ is obtained from the Bayesrule

p(θ|L) ∝ p(L|θ) π(θ) (18)

and can be used to define the Bayesian maximum a pos-teriori (MAP) and minimum mean squared error (MMSE)estimators given in (20) and (21) below.

6

Page 7: Bayesian Estimation of the Multifractality Parameter for ... · ML and Bayesian estimation. Examples of ML and Bayesian estimators for 1D fBm formulated in the spectral or wavelet

4 Estimation procedure

The computation of the Bayesian estimators is not straight-forward because of the complicated dependence of the pos-terior distribution (18) on the parameters θ. Specifically,the inverse and determinant of Σj in the expression of thelikelihood (14) do not have a parametric form and hence(18) can not be optimized with respect to the parametersθ. In such situations, it is common to use a Markov ChainMonte Carlo (MCMC) algorithm generating samples thatare distributed according to p(θ|L). These samples are usedin turn to approximate the Bayesian estimators.

4.1 Gibbs sampler

The following Gibbs sampler enables the generation of sam-ples {θ(t)}Nmc

1 that are distributed according to the pos-terior distribution (18). This sampler consists of succes-sively sampling according to the conditional distributionsp(c2|c02,L) and p(c02|c2,L) associated with p(θ|L). To gener-ate the samples according to the conditional distributions,a Metropolis-within-Gibbs procedure is used. The instru-mental distributions for the random walks are Gaussian andhave variances σ2

c2 and σ2c02

, respectively, which are adjusted

to ensure an acceptance rate between 0.4 and 0.6 (to ensuregood mixing properties). For details on MCMC methods,the reader is referred to, e.g., [46].Sampling according to p(c2|c02, L). At iteration t, de-

note as θ(t) = [c(t)2 , c

0,(t)2 ]T the current state vector. A

candidate c(?)2 is drawn according to the proposal distri-

bution p1(c(?)2 |c

(t)2 ) = N (c

(t)2 , σ2

c2). The candidate state

vector θ(?) = [c(?)2 , c

0,(t)2 ]T is accepted with probability

πc2 = min(1, rc2) (i.e., θ(t+12 ) = θ(?)) and rejected with

probability 1 − πc2 (i.e., θ(t+12 ) = θ(t)). Here, rc2 is the

Metropolis-Hastings acceptance ratio, given by

rc2 =p(θ(?)|L) p1(c

(t)2 |c

(?)2 )

p(θ(t)|L) p1(c(?)2 |c

(t)2 )

=1A(θ(?))

j2∏

j=j1

√det Σj(θ

(t))

det Σj(θ(?))

× exp

(−1

2lTj

(Σj(θ

(?))−1 −Σj(θ(t))−1

)lj

). (19)

Sampling according to p(c02|c2,L). Similarly, at itera-

tion t+ 12 , a candidate c

0,(?)2 is proposed according to the in-

strumental distribution p2(c0,(?)2 |c0,(t)2 ) = N (c

0,(t)2 , σ2

c02). The

candidate state vector θ(?) = [c(t+ 1

2 )2 , c

0,(?)2 ]T is accepted

with probability πc02 = min(1, rc02) (i.e., θ(t+1) = θ(?)) and

rejected with probability 1−πc02 (i.e., θ(t+1) = θ(t+12 )). The

Metropolis-Hastings acceptance ratio rc02 is given by (19)

with p1 replaced by p2, c2 replaced by c02 and t replaced byt+ 1

2 .Approximation of the Bayesian estimators. Aftera burn-in period defined by t = 1, . . . , Nbi, the proposed

Gibbs sampler generates samples {θ(t)}Nmc

t=Nbi+1 that are dis-tributed according to the posterior distribution (18). Thesesamples are used to approximate the MAP and MMSE es-timators

θMMSE , E[θ|L] ≈ 1

Nmc −Nbi

Nmc∑

t=Nbi+1

θ(t) (20)

θMAP , argmax

θp(θ|L) ≈ argmax

t>Nbi

p(θ(t)|L). (21)

4.2 Whittle approximation

The Gibbs sampler defined in subsection 4.1 requires the in-version of the nj × nj matrices Σj(θ) in (19) for each sam-pling step in order to obtain rc2 and rc02 . These inversionsteps are computationally prohibitive even for very modestimage sizes (for instance, a 64× 64 image would require theinversion of a dense matrix of size ∼ 1000 × 1000 at scalej = 1 at each sampling step). In addition, it is numericallyinstable for larger images (due to growing condition num-ber). To alleviate this difficulty, we propose to replace theexact likelihood (15) with an asymptotic approximation dueto Whittle [60, 61]. With the above assumptions, the col-lection of log-leaders {l(j, ·)} are realizations of a Gaussianrandom field on a regular lattice Pj = {1, ..,mj}2, wheremj =

√nj . Up to an additive constant, the Whittle approxi-

mation for the negative logarithm of the Gaussian likelihood(14) reads [4, 12,23,61]

− log p(lj |θ) ≈

pW (lj |θ) =1

2

ω∈Dj

log φj(ω;θ) +Ij(ω)

njφj(ω;θ)(22)

where the summation is taken over the spectral grid Dj ={ 2πmjb(−mj − 1)/2c,−1, 1,mj −bmj/2c}2. Here Ij(ω) is the

2D standard periodogram of {l(j,k)}k∈Pj

Ij(ω) =

∣∣∣∣∑

k∈Pj

l(j,k) exp(−ikTω)

∣∣∣∣2

(23)

and φj(ω;θ) is the spectral density associated with the co-variance function %j(∆r;θ), respectively. Without a closed-form expression for φj(ω;θ), it can be evaluated numericallyby discrete Fourier transform (DFT)

φj(ω;θ) =

∣∣∣∣∑

k∈Pj

%j(√k21 + k22;θ) exp(−ikTω)

∣∣∣∣. (24)

Frequency range. It is commonly reported in the liter-ature that the range of frequencies used in (22) can be re-stricted. This is notably the case for the periodogram-basedestimation of the memory coefficient of long-range depen-dent time series for which only the low frequencies can be

7

Page 8: Bayesian Estimation of the Multifractality Parameter for ... · ML and Bayesian estimation. Examples of ML and Bayesian estimators for 1D fBm formulated in the spectral or wavelet

0

0

−10

−5

0

ω1ω2

Periodogram ln Ij(ω)

(a) Periodogram

0

0

−10

−5

0

ω1ω2

Model ln φj(ω; θ)

(b) Parametric spectral density

0 //2 /

<10

<5

0

ω1

j = 2ω2 = ω1

ln Ij(ω)ln φj(ω;θ)

(c) Slice !2 = !1

0 //2 /

<10

<5

0

ω1

j = 2ω2 = 0

ln Ij(ω)ln φj(ω;θ)

(d) Slice !2 = 0

Figure 3: Fitting between the periodogram (a), averagedon 100 realizations of CMC-LN ([N, c2, j] = [29,�0.04, 2]),and the model �j(!;✓), obtained from a DFT of %j(�r;✓).(c) and (d) compare the model (blue) and the periodogram(red) for two slices.

by discrete Fourier transform (DFT)

�j(!;✓) =

����X

k2Pj

%j(q

k21 + k2

2;✓) exp(�ikT!)

����. (24)

Frequency range. It is commonly reported in the liter-ature that the range of frequencies used in (22) can be re-stricted. This is notably the case for the periodogram-basedestimation of the memory coe�cient of long-range depen-dent time series for which only the low frequencies can beused, see, e.g., [12,47,54]. Similarly, in the present context,the proposed spectral density model �j(!;✓) yields an excel-lent fit at low frequencies and degrades at higher frequencies.This is illustrated in Fig. 3 where the average periodogramsof l(j, k) for CPC-LN are plotted together with the model�j(!;✓). We therefore restrict the summation in (22) to

low frequencies D†j(⌘) =

n! 2 Dj |k!k2 p

⌘ 2⇡mj

bmj/2co

where the fixed parameter ⌘ approximately corresponds tothe fraction of the spectral grid Dj that is actually used. Wedenote the approximation of pW (lj |✓) obtained by replacing

Dj with D†j in (22) by p†

W (lj |✓). The final Whittle approx-imation of the likelihood (15) is then given by the following

0 0.2 0.4 0.6 0.8 1−0.01

0

0.01

0.02

0.03

0.04

0.05

N=

29

0 0.2 0.4 0.6 0.8 1

c2 = −0.02c2 = −0.08

0 0.2 0.4 0.6 0.8 1

−0.01

0

0.01

0.02

0.03

0.04

0.05

N=

28

η η η

m-c2 s rms

Figure 4: Influence of the bandwidth parameter, ⌘, on esti-mation performance for two di↵erent sizes of LN-CMC andtwo di↵erent values of c2 values (�0.02 in red and �0.08 inblue).

equation

p(L|✓) ⇡ exp

0@�

j2X

j=j1

p†W (lj |✓)

1A

= exp

0B@�

j2X

j=j1

1

2

X

!2D†j (⌘)

log �j(!;✓) +Ij(!)

nj�j(!;✓)

1CA (25)

up to a multiplicative constant. It is important to note thatthe restriction to low frequencies in (25) is not a restrictionto a limited frequency content of the image (indeed, scalesj 2 [j1, j2] are used) but only concerns the numerical evalu-ations of the likelihood (15).

5 Numerical experiments

The proposed algorithm is numerically validated for severaltypes of scale invariant and multifractal stochastic processesfor di↵erent sample sizes and a large range of values for c2.

5.1 Stochastic multifractal model processes

Canonical Mandelbrot Cascade (CMC). CMCs [39]are the historical archetypes of multifractal measures. Theirconstruction is based on an iterative split-and-multiply pro-cedure on an interval; we use a 2D binary cascade for twodi↵erent multipliers: First, log-normal multipliers W =2�U , where U ⇠ N (m, 2m/ log 2) is a Gaussian randomvariable (CMC-LN); Second, log-Poisson multipliers W =2� exp (log(�)⇡�), where ⇡� is a Poisson random variablewith parameter � = �� log 2

(��1) (CMC-LP). For CMC-LN, the

log-cumulants are given by c1 = m + ↵, c2 = �2m and

cp = 0 for all p � 3. For CMC-LP, c1 = ↵ + �⇣

log(�)��1 � 1

and all higher-order log-cumulants are non-zero with cp =

8

Figure 3: Fitting between the periodogram (a), averagedon 100 realizations of CMC-LN ([N, c2, j] = [29,−0.04, 2]),and the model φj(ω;θ), obtained from a DFT of %j(∆r;θ).(c) and (d) compare the model (blue) and the periodogram(red) for two slices.

used, see, e.g., [12,47,54]. Similarly, in the present context,the proposed spectral density model φj(ω;θ) yields an excel-lent fit at low frequencies and degrades at higher frequencies.This is illustrated in Fig. 3 where the average periodogramsof l(j,k) for CPC-LN are plotted together with the modelφj(ω;θ). We therefore restrict the summation in (22) to

low frequencies D†j(η) ={ω ∈ Dj |‖ω‖2 ≤ √η 2π

mjbmj/2c

}

where the fixed parameter η approximately corresponds tothe fraction of the spectral grid Dj that is actually used. Wedenote the approximation of pW (lj |θ) obtained by replacing

Dj with D†j in (22) by p†W (lj |θ). The final Whittle approx-imation of the likelihood (15) is then given by the followingequation

p(L|θ) ≈ exp

j2∑

j=j1

p†W (lj |θ)

= exp

j2∑

j=j1

1

2

ω∈D†j (η)

log φj(ω;θ) +Ij(ω)

njφj(ω;θ)

(25)

up to a multiplicative constant. It is important to note thatthe restriction to low frequencies in (25) is not a restrictionto a limited frequency content of the image (indeed, scalesj ∈ [j1, j2] are used) but only concerns the numerical evalu-ations of the likelihood (15).

0 0.2 0.4 0.6 0.8 1−0.01

0

0.01

0.02

0.03

0.04

0.05

N=

29

0 0.2 0.4 0.6 0.8 1

c2 = −0.02c2 = −0.08

0 0.2 0.4 0.6 0.8 1

−0.01

0

0.01

0.02

0.03

0.04

0.05

N=

28

η η η

m-c2 s rms

Figure 4: Influence of the bandwidth parameter, η, on esti-mation performance for two different sizes of LN-CMC andtwo different values of c2 values (−0.02 in red and −0.08 inblue).

5 Numerical experiments

The proposed algorithm is numerically validated for severaltypes of scale invariant and multifractal stochastic processesfor different sample sizes and a large range of values for c2.

5.1 Stochastic multifractal model processes

Canonical Mandelbrot Cascade (CMC). CMCs [39]are the historical archetypes of multifractal measures. Theirconstruction is based on an iterative split-and-multiply pro-cedure on an interval; we use a 2D binary cascade for twodifferent multipliers: First, log-normal multipliers W =2−U , where U ∼ N (m, 2m/ log 2) is a Gaussian randomvariable (CMC-LN); Second, log-Poisson multipliers W =2γ exp (log(β)πλ), where πλ is a Poisson random variablewith parameter λ = −γ log 2

(β−1) (CMC-LP). For CMC-LN, the

log-cumulants are given by c1 = m + α, c2 = −2m and

cp = 0 for all p ≥ 3. For CMC-LP, c1 = α + γ(

log(β)β−1 − 1

)

and all higher-order log-cumulants are non-zero with cp =− γβ−1 (− log(β))

p, p ≥ 2. Below, γ = 1.05 and β is varied

according to the value of c2.Compound Poisson Cascade (CPC). CPCs were in-troduced to overcome certain limitations of the CMCs thatare caused by their discrete split-and-multiply construc-tion [10, 16]. In the construction of CPCs, the localiza-tion of the multipliers in the space-scale volume follows aPoisson random process with specific prescribed density.We use CPCs with log-normal multipliers W = exp(Y ),where Y ∼ N (µ, σ) is a Gaussian random variable (CPC-LN), or log-Poisson CPCs for which multipliers W are re-duced to a constant w (CPC-LP). The first log-cumulants ofCPC-LN are given by c1 = −

(µ+ 1− exp

(µ+ σ2/2

))+α,

c2 = −(µ2 + σ2), and cp 6= 0 for p ≥ 3. Here, we fixµ = −0.1. For CPC-LP, c2 = − log(w)2.Fractional Brownian motion (fBm). We use 2D fBmsas defined in [50]. FBm is not a CMC process and is based

8

Page 9: Bayesian Estimation of the Multifractality Parameter for ... · ML and Bayesian estimation. Examples of ML and Bayesian estimators for 1D fBm formulated in the spectral or wavelet

Table 1: Estimation performance for CMC-LN (a), CPC-LN(b) CMC-LP (c), CPC-LP (d) for sample sizes N = {28, 29}and j1 = 2, j2 = {4, 5}. Best results are marked in bold.

(a) CMC-LNc2 −0.01 −0.02 −0.04 −0.06 −0.08

N=

28m

LF −0.015 −0.027 −0.053 −0.066 −0.087MMSE −0.014 −0.023 −0.042 −0.060 −0.078

s

LF 0.010 0.011 0.015 0.018 0.030MMSE 0.005 0.006 0.013 0.014 0.020

rms LF 0.011 0.014 0.019 0.019 0.030

MMSE 0.007 0.007 0.013 0.014 0.020

N=

29m

LF −0.016 −0.027 −0.049 −0.070 −0.087MMSE −0.014 −0.025 −0.047 −0.067 −0.087

s

LF 0.005 0.006 0.008 0.011 0.016MMSE 0.002 0.004 0.006 0.008 0.012

rms LF 0.008 0.010 0.012 0.015 0.018

MMSE 0.005 0.007 0.009 0.011 0.014

(b) CPC-LNc2 −0.01 −0.02 −0.04 −0.06 −0.08

N=

28m

LF −0.013 −0.025 −0.049 −0.066 −0.089MMSE −0.007 −0.013 −0.035 −0.054 −0.074

s

LF 0.009 0.011 0.017 0.024 0.029MMSE 0.003 0.005 0.011 0.016 0.022

rms LF 0.010 0.012 0.020 0.025 0.031

MMSE 0.004 0.008 0.012 0.018 0.023

N=

29m

LF −0.013 −0.025 −0.045 −0.066 −0.089MMSE −0.008 −0.015 −0.035 −0.057 −0.079

s

LF 0.004 0.005 0.010 0.014 0.015MMSE 0.002 0.003 0.006 0.009 0.013

rms LF 0.005 0.007 0.011 0.015 0.017

MMSE 0.003 0.005 0.008 0.009 0.013

(c) CMC-LP (d) CPC-LPc2 −0.02 −0.04 −0.08 −0.02 −0.04 −0.08

N=

28m

LF −0.019−0.038−0.076 −0.043 −0.065 −0.120MMSE −0.017 −0.032 −0.063 −0.029−0.055−0.100

s

LF 0.010 0.014 0.023 0.016 0.035 0.035MMSE 0.005 0.009 0.016 0.010 0.012 0.027

rms LF 0.010 0.014 0.023 0.028 0.043 0.050

MMSE 0.006 0.012 0.023 0.013 0.020 0.036

N=

29m

LF −0.020−0.040−0.075 −0.037 −0.063−0.100MMSE −0.019 −0.036 −0.070 −0.031−0.060 −0.120

s

LF 0.006 0.009 0.014 0.010 0.014 0.020MMSE 0.004 0.006 0.010 0.004 0.007 0.013

rms LF 0.006 0.009 0.015 0.020 0.027 0.032

MMSE 0.004 0.007 0.015 0.012 0.021 0.038

on an additive construction instead. Its multifractal andstatistical properties are entirely determined by a single pa-rameter H such that c1 = H, c2 = 0 and cp = 0 for allp > 2, and below we set H = 0.7.

5.2 Numerical simulations

Wavelet transform. A Daubechies’s mother waveletwith Nψ = 2 vanishing moments is used, and α = 1 in(9), which is sufficient to ensure positive uniform regularityfor all processes considered.Estimation. The linear regression weights wj in the stan-dard estimator (3) have to satisfy the usual constraints∑j2j1jwj = 1 and

∑j2j1wj = 0 and can be chosen to reflect

the confidence granted to each Varnj[log `(j, ·)], see [55,58].

Here, they are chosen proportional to nj as suggested in[55]. The linear regression based standard estimator (3) willbe denoted LF (for “linear fit”) in what follows. The Gibbssampler is run with Nmc = 7000 iterations and a burn-in pe-riod of Nbi = 3000 samples. The bandwidth parameter η in(25) has been set to η = 0.3 following preliminary numericalsimulations; these are illustrated in Fig. 4 where estima-tion performance is plotted as a function of η for LN-CMC(N = 28 top, N = 29 bottom) with two different values ofc2 (−0.02 in red, −0.08 in blue). As expected, η tunes aclassical bias-variance tradeoff: a large value of η leads to alarge bias and small standard deviation and vice versa. Thechoice η = 0.3 yields a robust practical compromise.Performance assessment. We apply the LF estimator(3) and the proposed MAP and MMSE estimators (20) and(21) to R = 100 independent realizations of size N × Neach for the above described multifractal processes. Arange of weak to strong multifractality parameter valuesc2 ∈ {−0.01,−0.02,−0.04,−0.06,−0.08} and sample sizesN ∈ {26, 27, 28, 29} are used. The coarsest scale j2 usedfor estimation is set such that nj2 ≥ 100 (i.e., the coarsestavailable scale is discarded), yielding j2 = {2, 3, 4, 5}, re-spectively, for the considered sample sizes. The finest scalej1 is commonly set to j1 = 2 in order to avoid pollution fromimproper initialization of the wavelet transform, see [53] fordetails. Performance is evaluated using the sample mean,the sample standard deviation and the root mean squarederror (RMSE) of the estimates averaged across realizations

m = E[c2], s =

√Var[c2], rms =

√(m− c2)2 + s2.

5.3 Results

Estimation performance. Tab. 1 summarizes the esti-mation performance of LF and MMSE estimators for CMC-LN, CPC-LN, CMC-LP, CPC-LP (subtables (a)-(d), respec-tively) and for sample sizes N = {28, 29}. The performanceof the MAP estimator was found to be similar to the MMSEestimator and therefore is not reproduced here due to spaceconstraints. Note, however, that different (application de-pendent) priors for θ may lead to different results (here, thenon-informative prior (17) is used).

First, it is observed that the proposed algorithm slightlybut systematically outperforms LF in terms of bias. This

9

Page 10: Bayesian Estimation of the Multifractality Parameter for ... · ML and Bayesian estimation. Examples of ML and Bayesian estimators for 1D fBm formulated in the spectral or wavelet

(a)16 17 18 19 20 21 22 23 24

11

12

13

14

15

16

17

18

19

(b)

(c) (d)

−0.2 −0.1 0 0.1

histogramMMSE

c2−0.2 −0.1 0 0.1

histogramLF

c2

−0.25 −0.2 −0.15 −0.1 −0.05 0 0.05 0.10

0.01

0.02

0.03

c2 threshold

Fisher linear discriminant criterion

MMSELF

(e)

Figure 5: Band #20 of a hyperspectral datacube (a); estimates of c2 for overlapping 64 ⇥ 64 pixel patches obtained byMMSE (c) and LF (d); zooms on the patches indicated by a red frame (b); the centers of the image patches are indicatedby white dots in the original image, the distance between two of the dots corresponds to one half of the patch size, axislabels indicate patch numbers. Histograms and Fisher linear discriminant criteria for estimates of c2 obtained by MMSEand LF (e).

computational cost, with computation times of the order of8s (N = 64) to 50s (N = 512) per image, respectively, on astandard desktop computer, which is two orders of magni-tude larger than the computational cost of the LF estimator.

Performance for fractional Brownian motion. Self-similar fBms with c2 = 0 do not belong to the class of MMCprocesses for which the proposed estimation procedure wasdesigned. The correlation structure of the wavelet coe�-cients of fBms has been studied in, e.g., [21]. This corre-lation is weak, i.e., it goes to zero fast with the distancebetween wavelet coe�cients in the time-scale plane. FBmresults are summarized in Tab. 3. They indicate that theperformance of the LF estimator is comparable to the casec2 = �0.01 reported in Tab. 1. In contrast, the proposedBayesian estimators are practically unbiased and have stan-dard deviations and RMSE values that significantly outper-form those of LF by up to a factor 10. Therefore, it is muchmore likely to be able to identify a model for which c2 = 0when using the proposed Bayesian procedure instead of theclassical LF.

6 Illustration for real-world data

We illustrate the proposed Bayesian estimation procedurefor the multifractal analysis of a real-world image in Fig.5(a). The image of size 960 ⇥ 1952 pixels is the channel#20 of a hyperspectral datacube corresponding to a forestedarea near a city that was acquired by the Hyspex hyperspec-tral scanner over Villelongue, France, during the Madonna

project [49]. Estimates of c2 are computed for 29⇥ 60 over-lapping patches of size 64 ⇥ 64 pixels.

The estimates are plotted in Fig. 5 for MMSE (c) and LF(d), subfigure (b) provides a magnification (indicated by ared frame) on the square of patches of rows 11-19 / columns16-24. Visual inspection indicates that the Bayesian esti-mates are much better reproducing the spatial structure ofthe image texture than the classical LF (cf., Fig. 5(a), (c)and (d)). Specifically, the zoom in Fig. 5(b) (equivalently,the corresponding textures in Fig. 5(a) and (c)), shows thatthe Bayesian estimates are spatially strongly homogeneousfor the forested regions with visually homogeneous texture(e.g., upper right portions in Fig. 5(b)), indicating a weakyet non-zero multifractality for these regions. Similar ob-servations are obtained for other homogeneous vegetationpatches (e.g., bottom left corners in Fig. 5(b)). Moreover,the zones of mixed vegetation (e.g., upper left corner in Fig.5(b)) also yield spatially coherent and consistent estimatesof c2, with more negative values (stronger multifractality).The LF based estimates display a strong variability through-out the image. Indeed, even for the homogeneous texturein the forested regions, LF yields strongly spatially varyingestimates. Finally, note that the strongly negative valuesof c2 observed for both MMSE and LF in the bottom leftcorner of Fig. 5(c) correspond to regions consisting of both(textured) vegetations and of roofs of buildings (with closeto zero amplitudes and no texture).

Although no ground truth is available for this illustra-tion, a more quantitative analysis of the relative quality of

10

Figure 5: Band #20 of a hyperspectral datacube (a); estimates of c2 for overlapping 64 × 64 pixel patches obtained byMMSE (c) and LF (d); zooms on the patches indicated by a red frame (b); the centers of the image patches are indicatedby white dots in the original image, the distance between two of the dots corresponds to one half of the patch size, axislabels indicate patch numbers. Histograms and Fisher linear discriminant criteria for estimates of c2 obtained by MMSEand LF (e).

reduction of bias does not depend on a specific choice of themultifractal process or its parameters, or on the sample size.Second, and most strikingly, the proposed Bayesian estima-tors yield significantly reduced standard deviations, with areduction of up to a factor of 3 as compared to linear regres-sions. The standard deviation reduction is more importantfor small values of |c2| yet remains above a factor of 1.5 forlarge values of |c2|.

These performance gains are directly reflected in the over-all RMSE values, which remain up to a factor of 2.5 belowthose of linear fits. Finally, note that the estimation perfor-mance for CMCs and CPCs with log-Poisson multipliers arefound to be slightly inferior to those with log-normal mul-tipliers. This may be due to an arguably slightly strongerdeparture from Gaussian for the former, cf. Fig. 2.

Performance for small sample size. For small samplesizes N ≤ 27, the limited number of available scales forcesthe choice j1 = 1. Results for N = {26, 27} (for whichj2 = {2, 3}, respectively) are reported in Tab. 2. They in-dicate that the performance gains of the proposed Bayesianestimators with respect to LF estimators are even more pro-nounced for small sample size, both in terms of bias andstandard deviations, yielding a reduction of RMSE valuesof up to a factor of 4. In particular, note that LF yieldsbiases that are prohibitively large to be useful in real-worldapplications due to the use of the finest scale j = 1, cf., [58].Notably, values c2 = 0 cannot be reliably detected with LF.In contrast, the proposed Bayesian procedure yields suffi-ciently small bias and standard deviations to enable the es-

timation of the multifractality parameter c2 even for verysmall images (or image patches) of size 64 × 64. The re-ported performance gains come at the price of an increasedcomputational cost, with computation times of the order of8s (N = 64) to 50s (N = 512) per image, respectively, on astandard desktop computer, which is two orders of magni-tude larger than the computational cost of the LF estimator.Performance for fractional Brownian motion. Self-similar fBms with c2 = 0 do not belong to the class of MMCprocesses for which the proposed estimation procedure wasdesigned. The correlation structure of the wavelet coeffi-cients of fBms has been studied in, e.g., [21]. This corre-lation is weak, i.e., it goes to zero fast with the distancebetween wavelet coefficients in the time-scale plane. FBmresults are summarized in Tab. 3. They indicate that theperformance of the LF estimator is comparable to the casec2 = −0.01 reported in Tab. 1. In contrast, the proposedBayesian estimators are practically unbiased and have stan-dard deviations and RMSE values that significantly outper-form those of LF by up to a factor 10. Therefore, it is muchmore likely to be able to identify a model for which c2 = 0when using the proposed Bayesian procedure instead of theclassical LF.

6 Illustration for real-world data

We illustrate the proposed Bayesian estimation procedurefor the multifractal analysis of a real-world image in Fig.5(a). The image of size 960 × 1952 pixels is the channel

10

Page 11: Bayesian Estimation of the Multifractality Parameter for ... · ML and Bayesian estimation. Examples of ML and Bayesian estimators for 1D fBm formulated in the spectral or wavelet

Table 2: Estimation performance for CMC-LN (a) and CPC-LN (b) for sample sizes N = {26, 27} and j1 = 1, j2 = {2, 3}.Best results are marked in bold.

(a) CMC-LNc2 −0.01 −0.02 −0.04 −0.06 −0.08

N=

26m

LF −0.042 −0.051 −0.067 −0.082 −0.110MMSE −0.014 −0.022 −0.038 −0.059 −0.078

s

LF 0.024 0.030 0.042 0.042 0.070MMSE 0.010 0.014 0.018 0.026 0.038

rms LF 0.040 0.043 0.050 0.047 0.076

MMSE 0.010 0.014 0.018 0.026 0.038

N=

27m

LF −0.035 −0.044 −0.064 −0.082 −0.100MMSE −0.013 −0.023 −0.044 −0.064 −0.082

s

LF 0.010 0.013 0.019 0.024 0.026MMSE 0.005 0.008 0.013 0.017 0.018

rms LF 0.027 0.027 0.031 0.033 0.033

MMSE 0.006 0.009 0.014 0.017 0.018

(b) CPC-LNc2 −0.01 −0.02 −0.04 −0.06 −0.08

N=

26m

LF −0.026 −0.054 −0.076 −0.085 −0.100MMSE −0.0082 −0.017 −0.030 −0.050 −0.065

s

LF 0.024 0.029 0.045 0.068 0.067MMSE 0.005 0.011 0.018 0.028 0.033

rms LF 0.029 0.044 0.058 0.073 0.070

MMSE 0.006 0.011 0.021 0.030 0.036

N=

27m

LF −0.021 −0.047 −0.064 −0.082 −0.110MMSE −0.008 −0.017 −0.035 −0.057 −0.079

s

LF 0.0091 0.013 0.020 0.024 0.032MMSE 0.0032 0.0082 0.012 0.018 0.021

rms LF 0.014 0.030 0.031 0.033 0.042

MMSE 0.004 0.0087 0.013 0.019 0.021

#20 of a hyperspectral datacube corresponding to a forestedarea near a city that was acquired by the Hyspex hyperspec-tral scanner over Villelongue, France, during the Madonnaproject [49]. Estimates of c2 are computed for 29× 60 over-lapping patches of size 64× 64 pixels.

The estimates are plotted in Fig. 5 for MMSE (c) and LF(d), subfigure (b) provides a magnification (indicated by ared frame) on the square of patches of rows 11-19 / columns16-24. Visual inspection indicates that the Bayesian esti-mates are much better reproducing the spatial structure ofthe image texture than the classical LF (cf., Fig. 5(a), (c)and (d)). Specifically, the zoom in Fig. 5(b) (equivalently,the corresponding textures in Fig. 5(a) and (c)), shows thatthe Bayesian estimates are spatially strongly homogeneousfor the forested regions with visually homogeneous texture(e.g., upper right portions in Fig. 5(b)), indicating a weakyet non-zero multifractality for these regions. Similar ob-servations are obtained for other homogeneous vegetationpatches (e.g., bottom left corners in Fig. 5(b)). Moreover,the zones of mixed vegetation (e.g., upper left corner in Fig.5(b)) also yield spatially coherent and consistent estimatesof c2, with more negative values (stronger multifractality).

Table 3: FBm estimation performance for sample sizesN = {27, 28, 29} and j1 = 2, j2 = {3, 4, 5}. Best resultsare marked in bold.

N 27 28 29

m

LF 0.0034 0.0047 0.0037MMSE −0.0020 −0.0008 −0.0003

s

LF 0.0170 0.0089 0.0056MMSE 0.0093 0.0010 0.0002

rms LF 0.0180 0.0100 0.0067

MMSE 0.0095 0.0012 0.0004

The LF based estimates display a strong variability through-out the image. Indeed, even for the homogeneous texturein the forested regions, LF yields strongly spatially varyingestimates. Finally, note that the strongly negative valuesof c2 observed for both MMSE and LF in the bottom leftcorner of Fig. 5(c) correspond to regions consisting of both(textured) vegetations and of roofs of buildings (with closeto zero amplitudes and no texture).

Although no ground truth is available for this illustra-tion, a more quantitative analysis of the relative quality ofestimates of c2 obtained with MMSE and LF is proposedhere. First, the reference-free image quality indicator of [13],which quantifies the image sharpness by approximating acontrast-invariant measure of its phase coherence [14], is cal-culated for the maps of c2 in Fig. 5(c) and (d). These sharp-ness indexes are 10.8 for MMSE and a considerably smallervalue of 4.6 for LF, hence reinforcing the visual inspection-based conclusions of improved spatial coherence for MMSEdescribed above. Second, Fig. 5(e) (top) shows histogramsof the estimates of c2 obtained with MMSE and LF, whichconfirm the above conclusions of significantly larger variabil-ity (variance) of LF as compared to MMSE. Moreover, LFyields a large portion of estimates with positive values, whichare not coherent with multifractal theory since necessarilyc2 < 0, while MMSE estimates are consistently negative. Fi-nally, the Fisher linear discriminant criterion [20, Ch. 3.8]is calculated for c2 obtained with MMSE and with LF, as afunction of a threshold for c2 separating two classes of tex-tures. The results, plotted in Fig. 5(e) (bottom), indicatethat the estimates obtained with MMSE have a far superiordiscriminative power than those obtained with LF.

7 Conclusions

This paper proposed a Bayesian estimation procedure forthe multifractality parameter of images. The procedure re-lies on the use of novel multiresolution quantities that haverecently been introduced for regularity characterization andmultifractal analysis, i.e., wavelet leaders. A Bayesian in-ference scheme was enabled through the formulation of anempirical yet generic semi-parametric statistical model forthe logarithm of wavelet leaders. This model accounts for

11

Page 12: Bayesian Estimation of the Multifractality Parameter for ... · ML and Bayesian estimation. Examples of ML and Bayesian estimators for 1D fBm formulated in the spectral or wavelet

the constraints imposed by multifractal theory and is de-signed for a large class of multifractal model processes. TheBayesian estimators associated with the posterior distribu-tion of this model were approximated by means of sam-ples generated by a Metropolis-within-Gibbs sampling pro-cedure, wherein the practically infeasible evaluation of theexact likelihood was replaced by a suitable Whittle approx-imation. The proposed procedure constitutes, to the best ofour knowledge, the first operational Bayesian estimator forthe multifractality parameter that is applicable to real-worldimages and effective both for small and large sample sizes.Its performance was assessed numerically using a large num-ber of multifractal processes for several sample sizes. Theprocedure yields improvements in RMSE of up to a factorof 4 for multifractal processes, and up to a factor of 10 forfBms when compared to the current benchmark estimator.The procedure therefore enables, for the first time, the reli-able estimation of the multifractality parameter for imagesor image patches of size equal to 64 × 64 pixels. It is in-teresting to note that the Bayesian framework introducedin this paper could be generalized to hierarchical models,for instance, using spatial regularization for patch-wise esti-mates. In a similar vein, future work will include the studyof appropriate models for the analysis of multivariate data,notably for hyperspectral imaging applications.

References

[1] P. Abry, R. Baraniuk, P. Flandrin, R. Riedi, andD. Veitch. Multiscale network traffic analysis, mod-eling, and inference using wavelets, multifractals, andcascades. IEEE Signal Processing Magazine, 3(19):28–46, May 2002.

[2] P. Abry, S. Jaffard, and H. Wendt. When Van Goghmeets Mandelbrot: Multifractal classification of paint-ing’s texture. Signal Proces., 93(3):554–572, 2013.

[3] P. Abry, S. Jaffard, and H. Wendt. A bridge betweengeometric measure theory and signal processing: Mul-tifractal analysis. In K. Grochenig, Y. Lyubarskii, andK. Seip, editors, Operator-Related Function Theory andTime-Frequency Analysis, The Abel Symposium 2012,pages 1–56. Springer, 2015.

[4] V.V. Anh and K.E. Lunney. Parameter estimation ofrandom fields with long-range dependence. Math. Com-put. Model., 21(9):67–77, 1995.

[5] J.-P. Antoine, R. Murenzi, P. Vandergheynst, and S. T.Ali. Two-Dimensional Wavelets and their Relatives.Cambridge University Press, 2004.

[6] A. Arneodo, E. Bacry, and J.F. Muzy. Randomcascades on wavelet dyadic trees. J. Math. Phys.,39(8):4142–4164, 1998.

[7] A. Arneodo, N. Decoster, P. Kestener, and S. G. Roux.A wavelet-based method for multifractal image analy-sis: from theoretical concepts to experimental applica-tions. In Adv. Imag. Electr. Phys., volume 126, pages1–98. Academic Press, 2003.

[8] A. Ayache, S. Cohen, and J.L. Vehel. The covariancestructure of multifractional brownian motion, with ap-plication to long range dependence. In Proc. IEEE Int.Conf. Acoust., Speech, and Signal Process. (ICASSP),volume 6, pages 3810–3813, Istanbul, Turkey, 2000.

[9] E. Bacry, A. Kozhemyak, and Jean-Francois Muzy.Continuous cascade models for asset returns. J. Eco-nomic Dynamics and Control, 32(1):156–199, 2008.

[10] J. Barral and B. Mandelbrot. Multifractal productsof cylindrical pulses. Probab. Theory Relat. Fields,124:409–430, 2002.

[11] C.L. Benhamou, S. Poupon, E. Lespessailles,S. Loiseau, R. Jennane, V. Siroux, W. J. Ohley,and L. Pothuaud. Fractal analysis of radiographictrabecular bone texture and bone mineral density:two complementary parameters related to osteoporoticfractures. J. Bone Miner. Res., 16(4):697–704, 2001.

[12] J. Beran. Statistics for Long-Memory Processes, vol-ume 61 of Monographs on Statistics and Applied Prob-ability. Chapman & Hall, New York, 1994.

[13] G. Blanchet and L. Moisan. An explicit sharpness indexrelated to global phase coherence. In Proc. IEEE Int.Conf. Acoust., Speech, and Signal Process. (ICASSP),pages 1065–1068, Kyoto, Japan, 2012.

[14] G. Blanchet, L. Moisan, and B. Rouge. Measuring theglobal phase coherence of an image. In Proc. Int. Conf.on Image Processing (ICIP), pages 1176–1179, 2008.

[15] B. Castaing, Y. Gagne, and M. Marchand. Log-similarity for turbulent flows? Physica D, 68(34):387 –400, 1993.

[16] P. Chainais. Infinitely divisible cascades to model thestatistics of natural images. IEEE Trans. Pattern Anal.Mach. Intell., 29(12):2105–2119, 2007.

[17] N. H. Chan and W. Palma. Estimation of long-memorytime series models: A survey of different likelihood-based methods. Adv. Econom., 20:89–121, 2006.

[18] T. Chang and C.-C. J. Kuo. Texture analysis and clas-sification with tree-structured wavelet transform. IEEETrans. Image Process., 2(4):429–441, 1993.

[19] J. Coddington, J. Elton, D. Rockmore, and Y. Wang.Multifractal analysis and authentication of Jackson Pol-lock paintings. In Proc. SPIE 6810, page 68100F, 2008.

12

Page 13: Bayesian Estimation of the Multifractality Parameter for ... · ML and Bayesian estimation. Examples of ML and Bayesian estimators for 1D fBm formulated in the spectral or wavelet

[20] R.O. Duda, P.E. Hart, and D.G. Stork. Pattern classi-fication. John Wiley & Sons, 2012.

[21] P. Flandrin. Wavelet analysis and synthesis of frac-tional Brownian motions. IEEE Trans. Inform. Theory,38:910–917, 1992.

[22] U. Frisch. Turbulence: the legacy of A.N. Kolmogorov.Cambridge University Press, 1995.

[23] M. Fuentese. Approximate likelihood for large irreg-ularly spaced spatial data. J. Am. Statist. Assoc.,102:321–331, 2007.

[24] R. M. Haralick. Statistical and structural approachesto texture. Proc. of the IEEE, 67(5):786–804, 1979.

[25] S. Jaffard. Wavelet techniques in multifractal analy-sis. In M. Lapidus and M. van Frankenhuijsen, edi-tors, Fractal Geometry and Applications: A Jubilee ofBenoıt Mandelbrot, Proc. Symp. Pure Math., volume72(2), pages 91–152. AMS, 2004.

[26] S. Jaffard, P. Abry, and H. Wendt. Irregularities andscaling in signal and image processing: Multifractalanalysis. In Michael Frame, editor, Benoit Mandelbrot:A Life in Many Dimensions. World scientific publish-ing, 2015. to appear.

[27] C.R. Johnson, P. Messier, W.A. Sethares, A.G. Klein,C. Brown, A.H. Do, P. Klausmeyer, P. Abry, S. Jaf-fard, H. Wendt, S. Roux, N. Pustelnik, N. van No-ord, L. van der Maaten, E. Potsma, J. Coddington,L.A. Daffner, H. Murata, H. Wilhelm, S. Wood, andM. Messier. Pursuing automated classification of his-toric photographic papers from raking light photomi-crographs. J. Amer. Inst. Conserv., 53(3):159–170,2014.

[28] K. Jones-Smith and H. Mathur. Fractal analysis: Re-visiting Pollock’s drip paintings. Nature, 444(7119):E9–E10, 2006.

[29] J. M. Keller, S. Chen, and R.M. Crownover. Texturedescription and segmentation through fractal geometry.Comp. Vis., Graphics, and Image Process., 45(2):150–166, 1989.

[30] P. Kestener, J. Lina, P. Saint-Jean, and A. Arneodo.Wavelet-based multifractal formalism to assist in diag-nosis in digitized mammograms. Image Analysis andStereology, 20(3):169–175, 2001.

[31] J. Levy-Vehel, P. Mignot, and J. Berroir. Multifrac-tals, texture and image analysis. In Proc. IEEE Conf.Comp. Vis. Pattern Recognition (CVPR), pages 661–664, Champaign, IL, USA, June 1992.

[32] O. Løvsletten and M. Rypdal. Approximated maximumlikelihood estimation in multifractal random walks.Phys. Rev. E, 85:046705, 2012.

[33] S.B. Lowen and M.C. Teich. Fractal-Based Point Pro-cesses. Wiley, 2005.

[34] Torbjorn Lundahl, William J. Ohley, S.M. Kay, andRobert Siffert. Fractional Brownian motion: A maxi-mum likelihood estimator and its application to imagetexture. IEEE Trans. Med. Imaging, 5(3):152–161, Sept1986.

[35] T. Lux. Higher dimensional multifractal processes: AGMM approach. J. Business & Economic Stat., 26:194–210, 2007.

[36] Thomas Lux. The Markov-switching multifractal modelof asset returns. J. Business & Economic Stat.,26(2):194–210, 2008.

[37] S. Mallat. A Wavelet Tour of Signal Processing. Aca-demic Press, 3rd edition, 2008.

[38] B. Mandelbrot. Limit lognormal multifractal measures.In E.A. Gotsman, Y. Ne’eman, and A. Voronel, editors,Frontiers of Physics, Proc. Landau Memorial Conf., TelAviv, 1988, pages 309–340. Pergamon Press, 1990.

[39] B.B. Mandelbrot. Intermittent turbulence in self-similar cascades: divergence of high moments and di-mension of the carrier. J. Fluid Mech., 62:331–358,1974.

[40] E. Moulines, F. Roueff, and M.S. Taqqu. A waveletWhittle estimator of the memory parameter of a non-stationary Gaussian time series. Ann. Stat., pages1925–1956, 2008.

[41] J.F. Muzy, E. Bacry, and A. Arneodo. The multifractalformalism revisited with wavelets. Int. J. of Bifurcationand Chaos, 4:245–302, 1994.

[42] M. Ossiander and E.C. Waymire. Statistical estima-tion for multiplicative cascades. Ann. Stat., 28(6):1533–1560, 2000.

[43] B. Pesquet-Popescu and J. Levy-Vehel. Stochastic frac-tal models for image processing. IEEE Signal Process.Mag., 19(5):48–62, 2002.

[44] L. Ponson, D. Bonamy, H. Auradou, G. Mourot,S. Morel, E. Bouchaud, C. Guillot, and J. Hulin.Anisotropic self-affine properties of experimental frac-ture surface. J. Fracture, 140(1–4):27–36, 2006.

[45] R.H. Riedi. Multifractal processes. In P. Doukhan,G. Oppenheim, and M.S. Taqqu, editors, Theory andapplications of long range dependence, pages 625–717.Birkhauser, 2003.

13

Page 14: Bayesian Estimation of the Multifractality Parameter for ... · ML and Bayesian estimation. Examples of ML and Bayesian estimators for 1D fBm formulated in the spectral or wavelet

[46] Christian P. Robert and George Casella. Monte CarloStatistical Methods. Springer, New York, USA, 2005.

[47] P. M. Robinson. Gaussian semiparametric estimationof long range dependence. Ann. Stat., 23(5):1630–1661,1995.

[48] S. G. Roux, A. Arneodo, and N. Decoster. A wavelet-based method for multifractal image analysis. III. Ap-plications to high-resolution satellite images of cloudstructure. Eur. Phys. J. B, 15(4):765–786, 2000.

[49] D. Sheeren, M. Fauvel, S. Ladet, A. Jacquin,G. Bertoni, and A. Gibon. Mapping ash tree coloniza-tion in an agricultural mountain landscape: Investigat-ing the potential of hyperspectral imagery. In Proc.IEEE Int. Conf. Geosci. Remote Sens. (IGARSS),pages 3672–3675, Vancouver, Canada, July 2011.

[50] M.L. Stein. Fast and exact simulation of frac-tional Brownian surfaces. J. Comput. Graph. Statist.,11(3):587–599, 2002.

[51] M. Unser. Texture classification and segmentationusing wavelet frames. IEEE Trans. Image Process.,4(11):1549–1560, 1995.

[52] B. Vedel, H. Wendt, P. Abry, and S. Jaffard. On theimpact of the number of vanishing moments on the de-pendence structures of compound Poisson motion andfractional Brownian motion in multifractal time. InP. Doukhan, G. Lang, D. Surgailis, and G. Teyssiere,editors, Dependence in Probability and Statistics, Lec-ture Notes in Statistics, pages 71–101. Springer BerlinHeidelberg, 2010.

[53] D. Veitch, M. Taqqu, and P. Abry. Meaningful MRAinitialization for discrete time series. Signal Process.,80(9):1971–1983, 2000.

[54] C. Velasco and P.M. Robinson. Whittle pseudo-maximum likelihood estimation for nonstationary timeseries. J. Am. Statist. Assoc., 95(452):1229–1243, 2000.

[55] H. Wendt, P. Abry, and S. Jaffard. Bootstrap for empir-ical multifractal analysis. IEEE Signal Process. Mag.,24(4):38–48, 2007.

[56] H. Wendt, N. Dobigeon, J.-Y. Tourneret, and P. Abry.Bayesian estimation for the multifractality parameter.In Proc. IEEE Int. Conf. Acoust., Speech, and SignalProcess. (ICASSP), Vancouver, Canada, May 2013.

[57] H. Wendt, S. Jaffard, and P. Abry. Multifractal analysisof self-similar processes. In Proc. IEEE Workshop onstatistical signal processing (SSP), pages 69–72, AnnArbor, MI, USA, Aug. 2012.

[58] H. Wendt, S. G. Roux, S. Jaffard, and P. Abry. Waveletleaders and bootstrap for multifractal analysis of im-ages. Signal Process., 89(6):1100 – 1114, 2009.

[59] Herwig Wendt, Patrice Abry, Stephane Jaffard, Hui Ji,and Zuowei Shen. Wavelet leader multifractal analy-sis for texture classification. In Proc. IEEE Int. Conf.Image Process. (ICIP), Cairo, Egypt, Nov. 2009.

[60] P. Whittle. Estimation and information in stationarytime series. Arkiv for matematik, 2(5):423–434, 1953.

[61] P. Whittle. On stationary processes in the plane.Biometrika, 41:434–449, 1954.

[62] G.W. Wornell and A. V. Oppenheim. Estimation offractal signals from noisy measurements using wavelets.IEEE Trans. Signal Process., 40(3):611–623, 1992.

[63] Y. Xu, X. Yang, H. Ling, and H. Ji. A new texture de-scriptor using multifractal analysis in multi-orientationwavelet pyramid. In Proc. IEEE Conf. Comp. Vis. Pat-tern Recognition (CVPR), pages 161–168, San Fran-cisco, CA, USA, June 2010.

[64] A. Atto Z., Tan, O. Alata, and M. Moreaud. Non-stationary texture synthesis from random field model-ing. In Proc. IEEE Int. Conf. Image Processing (ICIP),pages 4266–4270, Paris, France, 2014.

14