26
Functional Data Analysis for Big Data: A case study on California temperature trends Pantelis Zenon Hadjipantelis and Hans-Georg M¨ uller Abstract In recent years, detailed historical records, remote sensing, genomics and medical imaging applications as well as the raise of the Internet-of-Things present novel data streams. Many of these data are instances where functions are more suit- able data atoms than traditional multivariate vectors. Applied functional data analy- sis (FDA) presents a potentially fruitful but largely unexplored alternative analytics framework that can be incorporated directly into a general Big Data analytics suite. As an example, we present a modeling approach for the dynamics of a functional data set of climatic data. By decomposing functions via a functional principal com- ponent analysis and functional variance process analysis, a robust and informative characterization of the data can be derived; this provides insights into the relation- ship between the different modes of variation, their inherit variance process as well as their dependencies over time. The model is applied to historical data from the Global Historical Climatology Network in California, USA. The analysis reveals that climatic time-dependent information is jointly carried by the original processes as well as their noise/variance decomposition. 1 Introduction Functional Data Analysis (FDA) is the statistical framework for the analysis of function-valued data. Under this framework each sample instance is considered to be an individual realization of an (often smooth) underlying stochastic process [76]. Common examples are curves (one-dimensional functions) or images (two- dimensional functions). Pantelis Zenon Hadjipantelis University of California, Davis, Department of Statistics e-mail: [email protected] Hans-Georg M¨ uller University of California, Davis, Department of Statistics e-mail: [email protected] 1

Functional Data Analysis for Big Data: A case study on ...anson.ucdavis.edu/~mueller/pant.pdfThis latter point emphasizes the connection of FDA with big data. Big data are assumed

  • Upload
    others

  • View
    1

  • Download
    0

Embed Size (px)

Citation preview

Page 1: Functional Data Analysis for Big Data: A case study on ...anson.ucdavis.edu/~mueller/pant.pdfThis latter point emphasizes the connection of FDA with big data. Big data are assumed

Functional Data Analysis for Big Data: A casestudy on California temperature trends

Pantelis Zenon Hadjipantelis and Hans-Georg Muller

Abstract In recent years, detailed historical records, remote sensing, genomics andmedical imaging applications as well as the raise of the Internet-of-Things presentnovel data streams. Many of these data are instances where functions are more suit-able data atoms than traditional multivariate vectors. Applied functional data analy-sis (FDA) presents a potentially fruitful but largely unexplored alternative analyticsframework that can be incorporated directly into a general Big Data analytics suite.As an example, we present a modeling approach for the dynamics of a functionaldata set of climatic data. By decomposing functions via a functional principal com-ponent analysis and functional variance process analysis, a robust and informativecharacterization of the data can be derived; this provides insights into the relation-ship between the different modes of variation, their inherit variance process as wellas their dependencies over time. The model is applied to historical data from theGlobal Historical Climatology Network in California, USA. The analysis revealsthat climatic time-dependent information is jointly carried by the original processesas well as their noise/variance decomposition.

1 Introduction

Functional Data Analysis (FDA) is the statistical framework for the analysis offunction-valued data. Under this framework each sample instance is consideredto be an individual realization of an (often smooth) underlying stochastic process[76]. Common examples are curves (one-dimensional functions) or images (two-dimensional functions).

Pantelis Zenon HadjipantelisUniversity of California, Davis, Department of Statistics e-mail: [email protected]

Hans-Georg MullerUniversity of California, Davis, Department of Statistics e-mail: [email protected]

1

Page 2: Functional Data Analysis for Big Data: A case study on ...anson.ucdavis.edu/~mueller/pant.pdfThis latter point emphasizes the connection of FDA with big data. Big data are assumed

2 Pantelis Zenon Hadjipantelis and Hans-Georg Muller

A typical real-life dataset is a set of growth curves; a size/development-relatedprocess Y is measured in a sample of N individuals along a time domain T [68].The observed process Y is therefore a height process of which a sample of individ-ual growth trajectories yi(t), i = 1, . . . ,n, are observed for n individuals. Anotherexample is auction data [58]; the bid amounts from N individuals from the time anauction starts until it ends. In this case the observed process Y is a price process.The continuum over which a process is observed is not exclusively time; for exam-ple spatial coordinates [17] or mass [40] and mass-charge spectra [52] are measuredover various continua that occur in FDA applications. The fact that one has repeatedobservations from the process at hand distinguishes FDA from standard time-seriesanalysis techniques for which certain stationarity assumptions are made. Similarlythe continuum needs not be one-dimensional; in shape analysis [18] and medicalimaging [19, 47] functional data are often two- [50, 79] or even three-dimensional[54, 72].

As expressed by Valderrama: “[extensions towards] FDA have been done [in]two main ways: by extending multivariate techniques from vectors to curves andby descending from stochastic processes to [the] real world” [88]. FDA has beendeveloped under diverse frameworks: eg. Dirichlet [71], Gaussian [37] or Poisson[43] data. Furthermore the data themselves can be represented using splines [75],wavelets [31] or principal components as a basis; any basis (eg. Fourier [75] orLegendre [29] polynomials) could be used depending on the suitability of the basisin question to represent the data.

A major distinction from multivariate data is the smoothness and differentiabilityof functional data. If the functional nature of a functional dataset is ignored, resultstypically are sub-optimal as valuable information related to the functional nature ofthe dataset is not properly utilized.

This latter point emphasizes the connection of FDA with big data. Big dataare assumed to be primarily characterized in terms of volume, variety, velocity[25] and complexity. This can be true for multivariate and functional data alike;what is specific for functional data is that they are infinite-dimensional objects andtherefore even highly complex multivariate analysis techniques cannot account forthe complexity reflected in a stochastic process. In addition traditional multivari-ate techniques will suffer as the number of components gets bigger. Their directapplicability diminishes due to increased computational demands as well as the-oretical shortcomings as the dimensions increase. Standard principal componentanalysis (PCA) does not account for smoothness or continuity. One can considerpenalization/sparsity-inducing [86, 6] or re-sampling techniques [4, 20] to intelli-gently reduce the number of variables required to work with and then potentiallypost-process the data in a heuristic way to enforce certain modeling assumptions(eg. the non-negativeness of the generated factors in the case of non-negative matrixfactorization [55]).

These are valid approaches that are however inapplicable when the data have asmooth functional structure and are in this sense infinite dimensional. For exam-ple differentiation and convolution operations are natural steps when working withfunctions. They allow practitioners to quantify changes in the underlying dynamics

Page 3: Functional Data Analysis for Big Data: A case study on ...anson.ucdavis.edu/~mueller/pant.pdfThis latter point emphasizes the connection of FDA with big data. Big data are assumed

Functional Data Analysis for Big Data: A case study on California temperature trends 3

of time evolving patterns (eg. [74, 89]) or to exploit structural notions such as con-vexity (eg. [59]) or to apply appropriate transformations (eg. [33]). Finally, it shouldbe noted that FDA can tackle cases where the data are recorded over an irregularor even sparse design of points where measurements are taken, while multivariatetechniques assume that the dimensions of all data vectors are the same.

2 Basic Statistics with functional data

When working with functional data the unit of observation is a smooth functionY (t), t ∈ T . For simplicity we assume that T ⊂R. One then observes a sample of sizeN realizations of Y ; we will use Yi j to describe the observed value of i-th realizationat the j-th time-point. Importantly, we need to also assume that Y ∈ L2(T ). For thesmooth random processes Y we define the mean function:

µY (t) = E[Y (t)] (1)

and the symmetric positive definite auto-covariance function:

CYY (s, t) = Cov[Y (s),Y (t)] (2)=E[(Y (t)−µY (t))(Y (s)−µY (s))]. (3)

Noting that the auto-covariance function CYY (s, t) is by definition symmetric andpositive semi-definite, its spectral decomposition is given by Mercer’s theorem [64]:

CYY (s, t) =∞

∑k=1

λkφk(s)φk(t), (4)

where λ1 ≥ λ2 ≥ ...≥ 0 are the ordered eigenvalues of the Hilbert-Schmidt operatorwith kernel CYY with ∑

∞k=1 λk < ∞. The φk are the orthonormal eigenfunctions of

this operator. The Karhunen-Loeve expansion of the observations Y [36] is givenas:

Yi(t) = µY (t)+∞

∑k=1

ξikφk(t), (5)

where the ξik are the zero-mean functional principal component scores with vari-ance equal to the corresponding eigenvalues λk and the eigenfunction φk act as theorthonormal basis for the space spanned by the process Y .

Assuming a second process X similar to Y , in a similar manner we define thekernel CY X of the cross-covariance operator CY X as:

CY X (s, t) = E(Yi(s)−µY (s))(Xi(t)−µX (t)) (6)

which has the following functional singular decomposition [94]:

Page 4: Functional Data Analysis for Big Data: A case study on ...anson.ucdavis.edu/~mueller/pant.pdfThis latter point emphasizes the connection of FDA with big data. Big data are assumed

4 Pantelis Zenon Hadjipantelis and Hans-Georg Muller

CY X (s, t) =∞

∑p=1

σpφY,p(s)φX ,p(t), (7)

where now the σp denote the singular values associated with the decomposition ofthe covariance between the processes Y and X [94] and φY,p are the “left” and φX ,

the “right” eigenfunctions. Notably the cross-covariance surface, while adjoint, isnot symmetric and not self-adjoint.

In real-life applications statistical analysts are increasingly presented with mas-sive datasets stemming from continuous multi-channel recordings: from anatomi-cal brain regions [93] to climatic variables across different geographic regions [96]and multiple sensors tracking humans or animals [13]. The functional data analysisframework is essential to analyze such data.

3 Dimension Reduction for Functional Data

As stated by Delicado [16], functional data can be viewed as augmenting multi-variate techniques to a functional domain. One is therefore presented with infinitedimensional objects that require extension of traditional inferential procedures. Ad-ditionally, an infinite dimensional object cannot be fully represented when using afinite amount of memory.

Aside the obvious size-constraints, it is possible that we are encoding redundant,unrelated or even misleading information in a high-dimensional dataset. It is there-fore to a modeler’s benefit to extract features or modes of variation that are infor-mative representations. In that regard, even higher dimensional datasets might havean adequate, in terms of variation explained, representation in two or three dimen-sions enabling their visualization. Classical methods to obtain such representationsare PCA and Multi-Dimensional Scaling (MDS). Ideally, the reduced dimensionrepresentation of the dataset is useful as a surrogate dataset for the original high-dimensional data analyzed, both for dimension reduction and to filter unstructuredinformation out of the original dataset.

Dimension reduction is based on the notion that we can produce a compactlow-dimensional encoding of a given high-dimensional dataset. The currently mainmethodology to achieve this is Functional Principal Components Analysis (FPCA)[36, 51].

FPCA is an inherently linear and unsupervised method for dimension reduction[27] that has been applied for diverse applications of FDA. Linear refers to the factthat the resulting representation is linear in the random functions that are repre-sented. In the case of FPCA the original zero-mean process Y is approximated byforming projections of the form:

ξν ,n =∫ T

0φν(t)(Y f (t)−µy(t))dt (8)

Page 5: Functional Data Analysis for Big Data: A case study on ...anson.ucdavis.edu/~mueller/pant.pdfThis latter point emphasizes the connection of FDA with big data. Big data are assumed

Functional Data Analysis for Big Data: A case study on California temperature trends 5

where EY (t) = 0, φν is the ν-th eigenfunction and ξν is the corresponding FPCscore; Var(ξν) = λν as in Eq. 5. Alternative non-linear (but still unsupervised) di-mension reduction algorithms include kernel PCA [80], locally linear embedding(LLE) [78] and semi-definite embedding (SDE) [90]); the latter can also be castedas kernel PCA [27]. Chen and Muller demonstrate that some functional data canbe successfully represented as non-linear manifolds [9] using Isomap [84]. Fisher’sLinear Discriminant analysis is a typical example of an alternative supervised lineardimensionality reduction algorithm [3] that also has been successfully applied tofunctional data [45].

As a final note, a problem that is common to FPCA applications is the selectionof the number of components to retain. This is still an open problem that is effec-tively a model selection problem [42, 65]. One can formulate the selection problemas an optimization problem of an equivalent information criterion; for example Yaoet al. [95] propose pseudo-AIC and BIC criteria for the determination of the rel-evant k; other heuristic methods (eg. the broken stick model) [7] are popular. Afinal important note on dimension reduction is that, as Ma & Zhu emphasize: “be-cause dimension reduction is generally viewed as a problem of estimating a space,inference is strikingly left out of the main-stream research” [60]. Sliced inverse re-gression can be a fruitful alternative approach [57] for dimension reduction. Ferre& Yao [23] have formulated a relevant application framework for applying slicedinverse regression to functional data; Jiang et al. [48] have moved this frameworkeven further with results on irregularly sampled data.

4 Functional Principal Component Analysis

Tucker [87] introduced the idea of a function as the fundamental unit of statisticalanalysis in the context of Factor Analysis (FA). Nevertheless it was the later workof Castro et al. [8] that popularized the concept of dimensionality reduction viacovariance function eigendecomposition for functional data as in Eq. 4. Returningto the original notion of a stochastic process Y , the “best” K-dimensional linearapproximation in the sense of minimizing the variance of the residuals is:

Yi(t) = µ(t)+K

∑k=1

ξi,kφk(t), (9)

where the ξi,k are the k-th principal component scores and φk (k ≥ 1) are the eigen-functions that form a basis of orthonormal functions. In this functional settingthe usual mean squared error optimality condition found in standard multivariatePCA minimizing ∑

Ki=1 ||yi− y||2 can now be restated as the integrated square error∫

[y(t)− y(t)]2dt. The fraction of the sample variance explained is maximized alongthe modes of variation represented by the eigenfunctions φk.

Empirically finding the eigenfunctions φk requires first the estimation of the sam-ple auto-covariance function CY (s, t); see Eq. 3:

Page 6: Functional Data Analysis for Big Data: A case study on ...anson.ucdavis.edu/~mueller/pant.pdfThis latter point emphasizes the connection of FDA with big data. Big data are assumed

6 Pantelis Zenon Hadjipantelis and Hans-Georg Muller

CYY (s, t) =1N

N

∑i=1

(Yi(s)− µY (s))(Yi(t)− µY (t)), (10)

where µY (t) = 1N ∑

Ni=1 Yi(t). Then the subsequent eigendecomposition of CYY (s, t)

for the zero-meaned sample Y is approximated by the first K components as:

CYY (s, t) =K

∑ν=1

λν φν(s)φν(t). (11)

This is in direct analogy with the principal axis theorem [49] for multivariate datawhere the sample covariance matrix CYY is decomposed as CYY = ΦΛ ΦT , whereΦ is the orthogonal matrix compounded of the eigenvectors for the multivariatesample Y and Λ is the diagonal matrix composed of estimated eigenvalues. Ulti-mately, because of the optimality of the FPCs in terms of variance explained, thesenew axes explain the maximum amount of variance in the original sample. The esti-mated eigenfunctions φk are restricted to be orthonormal. The corresponding modeof variation has the form µ(t)+ζ φk(t) for ζ ∈ R. For a more detailed introductionto FPCA, Horvath and Kokoszka provide an excellent resource [41].

FPCA serves not only as a convenient method for dimensionality reduction butalso provides a way to build characterizations of the sample trajectories around anoverall mean trend function [95]. An additional geometric perspective is that theeigenfunctions represent a rotation of the original functional data that diagonalizesthe covariance matrix of the data. The resulting scores ξk are uncorrelated [3], asin the case of multivariate PCA. Table 1 lists the analogies between standard multi-variate and functional PCA.

Common assumptions are that the mean curve and the first few eigenfunctionsare smooth and that the eigenvalues λk tend to zero rapidly; the faster this conver-gence happens the smoother are the trajectories. The smoothness of the underlyingtrajectories is critical so that the discrete sample data can be considered functional[41], even when there is usually additive measurement noise. As seen in Chen &Muller [10] for the case of two-dimensional functional data, the discretization andthe subsequent interpolation can have significant implications.

FPCA can be implemented in many different ways. A number of regularized orsmoothed functional PCA approaches have been developed over the years. Smooth-ness is imposed in two main ways. The first approach is by penalizing the roughnessof the candidate eigenfunctions φk directly, where their roughness is measured bythe integrated squared second derivative over the interval of interest [73] or by ap-proximating φk with a family of smooth functions (eg. B-splines [46] or Fourierpolynomials [30]). The second approach is by smoothing the data or their equiva-lent representation (their autocovariance) directly and then projecting the dataset toa lower dimensional domain where smoothness is ensured. The actual smoothingin these cases can be done using usual smoothing procedures (eg. [95]). The basicqualitative difference between these two smoothing approaches is that in the firstcase smoothing occurs directly on the FPCA eigendecomposition step, while in the

Page 7: Functional Data Analysis for Big Data: A case study on ...anson.ucdavis.edu/~mueller/pant.pdfThis latter point emphasizes the connection of FDA with big data. Big data are assumed

Functional Data Analysis for Big Data: A case study on California temperature trends 7

second case we smooth the raw data or their higher order moments. Both approacheshave been used extensively in the literature.

Concept MVPCA FPCA

Data Y ∈ Rp Y (t) ∈ L2(T )Dimension p < ∞ ∞

Mean µY = E(Y ) µY (t) = E(Y (t))Inner Product 〈x,y〉 ∑x jy j

∫x(t)y(t)dt

Covariance ΣY,Y ( j, l) (matrix) CY,Y (s, t) (kernel)Eigenvalues λ1 ≥ λ2 ≥ ·· · ≥ λp ≥ 0 λ1 ≥ λ2 ≥ ·· · ≥ 0

Eigen-vectors/functions φ1 ⊥ φ2 ⊥ ·· · ⊥ φp φ1(t)⊥ φ2(t)⊥ . . .

Principal Component Scores ξk = ∑(y j−µy)φk ξ =∫(Y (t)−µY (t)φ(t)dt

Table 1 The analogies between multivariate PCA and functional PCA.

4.1 Smoothing and Interpolation

Common design assumptions in FDA are that the sample consists either of sparse(and possible irregularly sampled) observations [95] or of densely/perfectly ob-served discretized instances of smooth varying functions [2]. Situations in betweenthese two have also been considered. In all cases, we assume there is an under-lying smooth generating process such that the observed data result from possiblynoisy measurements of this process. Due to the assumed smoothness of the underly-ing process the following smoothing techniques can be employed: localized kernelsmoothing [12], smoothing splines [31] or wavelet smoothing [1].

A local smoother based on kernel weights is analogous to the windowing of a dig-ital signal to conduct short-time analysis. It corresponds roughly to the windowingidea in Digital Signal Processing that the signal/sample within that window is sta-tionary and the model’s inferred parameters are reasonable estimates for the overallbehavior within the window. In this setting, a smoothing kernel is a non-negative,symmetric, continuous (on its support), real-valued integrable function K satisfy-ing:

∫ +∞

−∞K(t)dt = 1. The bandwidth (or tuning parameter) b plays a key role in

kernel smoothing; it can be viewed as analogous to the characteristic length scale ofa Gaussian process [77].

A basic kernel smoother is the Nadaraya-Watson estimator which is effec-tively a weighted mean µNW , where the weights are determined by the kernelfunction K and bandwidth b used. For data (ti,yi), i = 1, . . . ,s, generated by yi =m(ti)+ εi, assuming a smooth function m, the Nadaraya-Watson estimator is given

by µNW (t) =∑

si=1 K

(t−ti

b

)yi(t)

∑sj=1 K

( t−t jb

) [15]; alternative kernel smoothers have been proposed

[26]. A closely related method is weighted least squares smoothing where one fitslocal linear lines to the data within the windows [t− b, t + b] and produces smooth

Page 8: Functional Data Analysis for Big Data: A case study on ...anson.ucdavis.edu/~mueller/pant.pdfThis latter point emphasizes the connection of FDA with big data. Big data are assumed

8 Pantelis Zenon Hadjipantelis and Hans-Georg Muller

curves. Using the notation SL for the smoother, evaluating the non-parametric re-gression function at point t and considering a sample (ti,yi), i = 1, . . . ,s, as inputs,the smoother SL takes the form:

SLt;b,(ti,yi)i=1,...,s= (12)

argminα0

minα1

(s

∑i=1

K(t− ti

b)[yi−α0 +α1(t− ti)]2)).

Kernels commonly used include the uniform (or rectangular) kernel function:K(x) = 0.5I|x|≤1 and the Epanechnikov kernel function: K(x) = 3

4 (1− x2)I|x|≤1,

where I|x|≤1 is an indicator function; the Gaussian kernel function: K(x) = 1√2π

e−x22

and the triangular kernel function: K(x) = (1−|x|)I|x|≤1; see [44]. The bandwidth bis commonly estimated using cross-validation [44] or generalized cross-validation[82]. The uniform, Epanechnikov and the triangular kernels have a finite supportwhich ensures that the size of the linear system solved by Eq. 12 always remainsfinite. Qualitatively, smaller values of b are associated with large-variability whilelarger values of b are associated with larger bias that results from broader smoothingwindows.

It is important to note that by employing a locally weighted least squaressmoother we can adjust for irregular sampling and/or data to some extent. Whileideally all curves are sampled over the same dense grid of points, in practice thisis often not the case. Smoothing can be used to transform the original irregularlyspaced observations to smooth curves that can then be sampled on a regular outputgrid. The number of grid points for the output grid is usually determined empiri-cally; in practice L is set equal to the expected number of readings per case for thesample at hand.

For example, when yearly measurements of a reasonably smooth varying processare available, 53 equidistant points are an obvious choice for the output grid; thusensuring that there are smooth values on weekly support points. As Levitin et al.mentions: “this [the choice] is largely a matter of experience, experimenter intu-ition and trial-and-error”[56]. The choice of L strongly relates to the binning of thedata and has a bearing on the computational load. When analyzing datasets with fewmissing values (ie. not significantly sparse datasets) often the use of the Epanech-nikov kernel function is a good choice. As an alternative to the finite support ker-nels, the Gaussian kernel, which enjoys an infinite support, can always “interpolateover holes” in situations with extremely sparse designs. One must be careful thoughto avoid situations where the dimensions of the actual systems solved (Eq. 12) be-come prohibitively large; explicit solutions are available [22]. Implementation-wise,based on simulations we recommend truncating a Gaussian kernel window at threebandwidth lengths from each side to ensure a reasonable maximal length-scale.

As an alternative to kernel smoothing, smoothing splines are a popular choice[21]. For FDA applications Guo presented a test-case where the smoothing frame-work employed a generalization of smoothing splines [31]. For standard splinesmoothing, the number of knots, the number of basis functions and the order of

Page 9: Functional Data Analysis for Big Data: A case study on ...anson.ucdavis.edu/~mueller/pant.pdfThis latter point emphasizes the connection of FDA with big data. Big data are assumed

Functional Data Analysis for Big Data: A case study on California temperature trends 9

the spline employed affect the final result. A third alternative is based on the notionof multi-resolution signal decomposition as implemented by wavelets [62]. Waveletapproximations have been widely employed; for example the popularly used JPEGstandard is based on wavelets.Wavelet estimators use projections onto subspaces ofL2[0,1] to represent successive approximations. In contrast to a standard orthogonalbasis that are localized only in a single domain (eg. Fourier polynomials are local-ized in frequency), wavelets can be localized both in time and frequency. Waveletsallow for fast implementations based on the Discrete Wavelet Transformation -DWT and can be used to smooth original noisy functional data [66]. Under thisparadigm, smoothing is achieved by thresholding; certain wavelet coefficients are“thresholded” in order to exclude parts considered to be noise. Often however, thesmoothed curves resulting from wavelets are not as smooth as desired.

All techniques require a tuning parameter which in the case of wavelets is theoriginal mother wavelet and the threshold used, in the case of spline smoothing, thetype of splines and the penalty parameter used, and in the case of kernel smoothingthe choice of kernel type and bandwidth.

Kernel, splines or wavelets estimators are linear in the data y:

ysmooth = Hy, (13)

where H is a projection or hat matrix. On a conceptual level one might argue thatsmoothing leads to dimension enhancement as it generates readings across a wholecontinuum that was previously unpopulated as this is obvious in the case of irreg-ularly sampled data. Numerically, both spline and wavelet smoothing have beenreported to be more computationally efficient than kernel smoothing [91, 5, 85] inits naive form. Binning and computational parallelizing overcome this issue and of-fer significant speed-ups compared to naive kernel smoothing, while providing astraightforward implementation that is easily adaptable to specific restrictions andmodifications one may wish to adopt. An additional advantage of this approach isthat it is easily amenable to mathematical analysis.

4.2 Sample size considerations

When working with a large sampe of dense functional data, sample size considera-tions come into play not only because the number of design points may increase. Toaddress the computational challenges we propose two basic remedies, binning andthe use of the Gram matrix.

Binning of data is a basic data pre-processing procedure. For the analysis oflarge data, binning has significant utilities; it has an important place in neuroscienceapplications [14].

When sampling a signal, the notion of the Nyquist-Shannon sampling theoremis relevant when choosing the sampling frequency, as this corresponds to the binwidth. There are a number of heuristics available to choose the number of bins, eg.

Page 10: Functional Data Analysis for Big Data: A case study on ...anson.ucdavis.edu/~mueller/pant.pdfThis latter point emphasizes the connection of FDA with big data. Big data are assumed

10 Pantelis Zenon Hadjipantelis and Hans-Georg Muller

Sturges [83], Scott [81] and Freedman-Diaconis [24] as well information-theoreticapproaches (eg. [53]); in any case using some prior knowledge as well as visualizingthe data is essential. The number of bins must be large enough to allow the processdynamics to be expressed without aliasing. This is clearly a common problem insituations where densely sampled signals are presented. When implementing FDA,one may face situations with “small n, large p”, sometimes abbreviated as “n<< p”.

For small n, large p, the eigendecompositions become non-trivial in terms ofspace as well computations required. A common computational trick that supportsbinning is to recognise that for the zero-meaned sample matrix Y0 the covariancematrix is:

C =1

n−1Y T

0 Y0, (14)

with corresponding Gram matrix:

G =1

n−1Y0Y T

0 (15)

where C and G share the same singular values. Additionally and equally impor-tantly the principal component scores of the covariance matrix C coincide with theeigenvectors of the Gram matrix G. The proof is straightforward: it entails express-ing Y0 as USV T and expressing the eigendecomposition of C as 1

n−1V S2V T and G as1

n−1US2UT [28]. Using this property, if either n << p or n >> p, one can computethe principal modes of variation of the original sample stored in the matrix Y0 byconsidering only min(n, p)×min(n, p) matrices.

Clearly there are situations where the number of observed functions n is large. Inthese situations the most common solutions fall under the category of randomizedalgorithms [69]. The basic idea is that one uses a random projection matrix Ω tomultiply the original large sample A to get the matrix B = AΩ . Subsequently onecomputes the matrix Q that defines an orthonormal basis for the range of B. A caveatis that, as B is an approximation to the range of A and thus projecting A by usingQT Q, we will get A ≈ QT QA [92]; this is natural as given that Q is approximatelyorthonormal then QT QapproxI. There is extensive work on how to sample theserandom projections. Randomized algorithms have been shown to be promising forvery large datasets [61]. For example, even in cases where less than a hundredth ofthe original dataset can be stored in the computer’s RAM, efficient algorithms havebeen derived to obtain nearly optimal decompositions [34]. Another option is simplyto work with a random subsample of size n∗ < n, for a suitably small n∗ at the costof increased variability, that needs to be quantified to obtain viable selections for n∗.

Page 11: Functional Data Analysis for Big Data: A case study on ...anson.ucdavis.edu/~mueller/pant.pdfThis latter point emphasizes the connection of FDA with big data. Big data are assumed

Functional Data Analysis for Big Data: A case study on California temperature trends 11

5 Functional Variance Process

When analyzing functional data, there is sometimes valuable information in smalloscillations that can be captured by the functional variance process [67] (FVP).FVP is a non-parametric tool to analyze stochastic time-trends in the noise vari-ance of functional data. In contrast to parametric techniques that impose particularassumptions regarding the form of potential heteroskedastic trends, non-parametricvariance function estimation techniques allow for data-driven detection of locallyvarying patterns of variations that may contain relevant information to quantify thevariation observed in the data.

More formally, we assume that Y1, . . . ,Yn denote n continuous smooth randomfunctions defined over a real interval [0,T ]. Here, we also assume that these func-tions are observed at a grid of dense time-points t j =

j−1m−1 T , j = 1, . . . ,m, with

measurements:

Yi j = Yi(t j)+Ri j, i = 1, . . . ,n, j = 1, . . . ,m, (16)

and Ri, j is additive noise such that ERi, jRi′k = 0 for all i 6= i′, ER = 0 andER2< ∞.

Importantly, the noise process R is assumed to generate the squared errors R2i, j

which themselves are assumed to be the product of two non-negative componentsV (ti, j) and Wi, j such that R2

i, j = exp[V (ti, j)]exp[Wi, j]; V (ti, j) and Wi, j representing theunderlying smooth functional variance process and a noise component, respectively.To draw conclusions about the underlying variance process of the original data Y wethen focus on the examination of its two components V (ti, j) and Wi, j.

The initial step is to work on the log-domain to transform the multiplicative rela-tion between V and W to an additive relation. This leads to the transformed errors:

Zi, j = log(Ri, j)2 =V (ti, j)+Wi, j. (17)

We assume that:

EZi, j= EV (ti, j)= µV (ti, j), (18)Cov(Zi, j,Zi, j′) = Cov(V (ti, j),V (ti, j′)). (19)

Furthermore we assume that the V (ti, j) are derived from a smooth variance processV (ti, j) with mean:

EV (t)= µV (t), (20)

and smooth symmetric auto-covariance:

CVV (s, t) =∞

∑k=1

ρkψk(s)ψk(t) (21)

Page 12: Functional Data Analysis for Big Data: A case study on ...anson.ucdavis.edu/~mueller/pant.pdfThis latter point emphasizes the connection of FDA with big data. Big data are assumed

12 Pantelis Zenon Hadjipantelis and Hans-Georg Muller

where in analogy with Eq. 3, the ρk are the non-negative ordered eigenvaluesρ1 ≥ ρ2 ≥ . . .0 and ψk are the corresponding orthonormal eigenfunctions of V de-scribing its principal modes of variation. The Wi, j are assumed to be independentwith EWi, j= 0 and Var(Wi, j) = σ2

W . The functional data can then be decomposedinto two processes Y and V with Karhunen-Loeve representations :

Y (t) =µY (t)+∞

∑k=1

ξkφk(t) (Eq. 5) and (22)

V (t) =µV (t)+∞

∑k=1

ζkψk(t). (23)

Operationally, after we compute the FPCA of the original data Yi, j yielding fitsYi, j = yi(ti, j) for the underlying smooth processes, we use the logarithms of thesquared differences between Yi and Yi as our new sample to conduct a second FPCAstep. In the notation of Eq. 16, the total decomposition of the observed measure-ments then becomes:

Yi, j = Yi(t j)+ sign(Ri, j)exp[Vi(ti, j)+Wi, j]12 . (24)

6 Functional data analysis for temperature trends

In this section we illustrate the insights one can obtain by using functional dataanalysis techniques on historical temperature data. We begin by showing the find-ings obtained by a basic FPCA approach, followed by the insights gained from theinspection of the functional cross-covariance operator and finally from the subse-quent FVPA (Functional Variance Process Analysis) of our data. In related work,Horvath and Kokoszka present a detailed introductory application on the subject ofdetecting changes in the mean function [41] for temperature data.

Our illustration data are daily temperature time-series for two cities in NorthernCalifornia, Redwood City (3728′N, 12214′W) and Davis (3833′N, 12144′W- Station IDs: USC00047339 & USC00042294 for Redwood City and Davis re-spectively), downloaded from the Global Historical Climatology Network(GHCN)[63]. While the two cities are relatively close (∼ 150km) and of similar elevation (∼ 15m), Davis is landlocked while Redwood City is adjacent to the south San Fran-cisco Bay; as large bodies of water are known to act as natural “heat reservoirs”we expect that the two cities have different temperature patterns. For both locationswe were able to obtain daily temperature data for approximately a whole century(86 and 108 years of observations for Redwood City and Davis, respectively). Thisallows us to look at changes across time. We focus our attention to two variables ofinterest, the daily maximum temperature and the daily minimum temperature.

In addition to the difference between coastal and continental weather patterns,we are interested to see if there are long-term trends such as a warming effect. Bothstations are in urban or peri-urban locations. Therefore, an urban warming effect

Page 13: Functional Data Analysis for Big Data: A case study on ...anson.ucdavis.edu/~mueller/pant.pdfThis latter point emphasizes the connection of FDA with big data. Big data are assumed

Functional Data Analysis for Big Data: A case study on California temperature trends 13

Jan Apr Jul Oct-10

-5

0

5

10

15

20

Degre

es C

els

ius

Daily Minimum - 1933

Raw

Smoothed

Jan Apr Jul Oct-10

-5

0

5

10

15

20

Degre

es C

els

ius

Daily Minimum - 1934

Raw

Smoothed

Jan Apr Jul Oct-10

-5

0

5

10

15

20

Degre

es C

els

ius

Daily Minimum - 1964

Raw

Smoothed

Fig. 1 The raw and smoothed temperature curves for the daily minimum in Redwood City in 1933,1934 and 1964.

may be present. Our basic unit of analysis is the daily trajectory of the tempera-ture extrema records across the span of a whole year; we make no assumptions ofseasonality or stationarity.

The original data are obviously non-smooth (Fig. 1) in their raw form; neverthe-less it is reasonable to assume that there is an underlying smoothly varying processthat encapsulates the dynamics behind the temperature extrema recorded. The de-viations from a smooth function can be viewed as aberrations or as measurementerrors. We will study the decomposition of the temperature signals into a functionalprocess that reflects the long-term trends and a variance process that reflects theaberration patterns.

For the regular FDA part, the smoothing procedures can be either directly appliedto densely sampled data for each subject (here: actual temperature profiles) or thepooled data before obtaining the mean and covariance functions. Here we chooseto smooth the data directly for computational reasons. First, smoothing the raw dataallows us to massively parallelize the smoothing procedure. When dealing with bigdata it is paramount to avoid moving data around for efficient computation. Oneshould aim to move the computation close to the data rather than the other wayaround. If someone chooses to smooth the covariance then one would deal with one(potentially massive) object that would need to segmented, distributed across nodes,processed and then recombined. We avoid this by employing the fact that data mightbe already split and thus we ”bring the computation” to the data.

Second, all smoothing procedures experience edge effects which is a problemthat needs to be addressed. The remedy is often to avoid smoothing near the edges(eg. [32]), to truncate the smoothed curves (eg. [70]), to enforce knot locations forsplines (eg. [75]) or to rebalance the weights intelligently (eg. [35]). We minimizeedge effects during smoothing as well as parallelize our smoothing procedure bycombining adjacent years. In Fig. 2, we show the basic idea where one segmentsdata into two main data segments (Chunks A and B). After segmenting the data intochunks, the chunks can be processed independently. Importantly, a small part ofthe data will be duplicated (shown in a dark gray) in both chunks. This duplicationserves to avoid edge effects. When smoothing the data we smooth only half of theduplicated part present in each chunk (shown in gray and black dotted lines for

Page 14: Functional Data Analysis for Big Data: A case study on ...anson.ucdavis.edu/~mueller/pant.pdfThis latter point emphasizes the connection of FDA with big data. Big data are assumed

14 Pantelis Zenon Hadjipantelis and Hans-Georg Muller

chunks A and B respectively). Then we avoid duplicating calculations and distributethe computation among all available resources. As a final comment we note thatthe actual duplicate chunk is relatively small in comparison with the overall spacerequirements; it needs to be only of size 2bq, b being the bandwidth used and qbeing the number of segments used.

CHUNK ACHUNK B

ENTIRE DATASET

Fig. 2 Example where the original dataset to be smoothed is segmented into two chunks A and B.The two chunks overlap (dark gray box) so information is shared and smoothing edge effects arealleviated near the segmentation limits. The smoothing output grid-points (shown in widely anddensely dotted lines for chunks A and B respectively) do not overlap so one does not duplicate anycalculations.

Analyzing the data and looking at the respective means, it is evident that thetemperature curves of Davis exhibit more extreme weather patterns in the courseof a typical year (Fig. 3). During winter-time the daily extrema in Davis appearmarginally lower than in Redwood City (Fig. 3, both plots), signifying that Davisis colder than Redwood city. During summer Davis experiences significantly higherdaily maximum temperatures (Fig. 3, RHS). These patterns strongly suggest that theyearly variation of average temperatures is higher in Davis than in Redwood City.

To quantify the differences further, we also quantify the daily spread, which cor-responds to the daily maximum minus the daily minimum. As can be seen in Fig. 4which shows the means as well as the covariance of the daily spreads in each region,the temperature spread in Davis is higher than in Redwood City for approximately10 months (Fig. 4, LHS). In addition it is evident that the variance of the spreadacross the year is of higher amplitude in Davis compared to Redwood City. Davis inlate autumn (mid-October to late-November) appears to have very strong intra-daytemperature changes indicating that the temperature patterns in Davis exhibit morevariation that those in Redwood City, the latter being possibly modulated by thecoastal location of Redwood City.

An issue of interest is the covariance/correlation of the annual temperaturecurves between the two locations and the detection of the periods where synergyis strongest. The cross-covariance estimator (Eq. 6) is of interest here. Fig. 5 showsthe cross-covariance surfaces between the three temperature related measurementswe examined. First of all it is notable that the surfaces are not symmetric across theirdiagonal; this is expected.

The off-diagonal elements appear to be routinely of lower absolute magnitudehinting to little covariance between the temperatures in Davis and Redwood Cityduring different seasons. The daily minima cross-covariance seems significantly

Page 15: Functional Data Analysis for Big Data: A case study on ...anson.ucdavis.edu/~mueller/pant.pdfThis latter point emphasizes the connection of FDA with big data. Big data are assumed

Functional Data Analysis for Big Data: A case study on California temperature trends 15

Jan Feb Mar Apr May Jun Jul Aug Sep Oct Nov Dec0

5

10

15

20

25

30

35

40

Degre

es C

els

ius

Mean Daily Minimum

Davis

Redwood City

Jan Feb Mar Apr May Jun Jul Aug Sep Oct Nov Dec0

5

10

15

20

25

30

35

40

Degre

es C

els

ius

Mean Daily Maximum

Davis

Redwood City

Fig. 3 Smoothed estimates for the mean daily temperature extrema recorded in Davis from 1908to 2015 and in Redwood City from 1930 to 2015 using an Epanechnikov kernel and bandwidth of49 days.

flatter and does not exhibit significant troughs (or bumps) compared to the dailymaxima and daily spreads of temperatures. This suggests that the daily minima co-vary more homogeneously across the two regions as well as across time; a trendthat is in line with the lower spread in the mean daily minima curves (Fig. 3, LHS).Focusing on the ridges of the surfaces it is telling that the time periods where thecovariance of the two functional datasets is highest differ between the type of mea-surement examined. In particular, while the minima covary most during January,suggesting that cool or warm winters are cool or warm for both places, the max-ima do not covary at the same time. Spring daily maximum temperatures appearto covary the most while, based on the behavior of the daily spreads, daily spreadsappear covary the most in late autumn or early winter. This can also be seen byvisually inspecting the cross correlation matrices (Fig. 6).

Fig. 4 Smoothed estimates for the mean daily temperature spread recorded in Davis from 1908to 2015 and in Redwood City from 1930 to 2015 (LHS plot) using an Epanechnikov kernel andbandwidth of 49 days. The auto-covariance surface of the daily temperatures in Davis (central plot)and in Redwood City (RHS plot).

Page 16: Functional Data Analysis for Big Data: A case study on ...anson.ucdavis.edu/~mueller/pant.pdfThis latter point emphasizes the connection of FDA with big data. Big data are assumed

16 Pantelis Zenon Hadjipantelis and Hans-Georg Muller

Fig. 5 Smoothed estimates for the cross-covariance surfaces between Davis and in Redwood Cityfrom 1930 to 2015 for daily temperature minima (LHS), maxima (central plot) and spread (RHS).

Fig. 6 Smoothed estimates for the cross-correlation surfaces between Davis and in Redwood Cityfrom 1930 to 2015 for daily temperature minima (LHS), maxima (central plot) and spread (RHS).

A question of interest is whether the data reflect climate change. It is gener-ally accepted [39] that the world as a whole is getting warmer. We note that in ouranalysis we do not incorporate any soil moisture, solar radiance or ocean-wind in-formation which limits the interpretability of our data analysis as these factors aregenerally thought to be important cofounders of the Northern California temperaturedynamics.

Jan Feb Mar Apr May Jun Jul Aug Sep Oct Nov Dec

0

0.02

0.04

0.06

0.08

0.1

φ1 Minima

Davis 56%

Redwood City 55%

Jan Feb Mar Apr May Jun Jul Aug Sep Oct Nov Dec

0

0.02

0.04

0.06

0.08

0.1

φ1 Maxima

Davis 34%

Redwood City 41%

Jan Feb Mar Apr May Jun Jul Aug Sep Oct Nov Dec

0

0.02

0.04

0.06

0.08

0.1

φ1 Spread

Davis 48%

Redwood City 52%

Fig. 7 The first estimated eigenfunction for the daily temperature minima (LHS), maxima (centralplot) and spread (RHS) as a function of the day of the year. The Fraction-of-Variance-Explainedfor each component is shown in the legend.

Page 17: Functional Data Analysis for Big Data: A case study on ...anson.ucdavis.edu/~mueller/pant.pdfThis latter point emphasizes the connection of FDA with big data. Big data are assumed

Functional Data Analysis for Big Data: A case study on California temperature trends 17

We begin by examining the first estimated eigenfunction φ1 (Fig. 7). Larger ab-solute values are associated with higher variation in the direction of the first eigen-function, as the sign of the eigenfunction is arbitrary. The peaks of the various com-ponents are located at different times between different locations suggesting that theperiods of maximum variation do not fully coincide.

Examining the second principal mode of variation φ2 (Fig. 8) it appears that it en-capsulates additional strongly expressed shapes in each of the respective underlyingtemperature processes. Something that is noteworthy is that neither of the eigen-functions is periodic, thus reflecting the non-monotonicity of the original sampleseries (Fig. 1). We also note that one can assess the fraction of variance explainedby each component using its respective eigenvalue ratio (Eq. 4. Using φ1 and φ2 wecan obtain the FPC scores ξ1 and ξ2 by ξ j =

∫(y− µ)(t)φ j(t)dt, j = 1,2. A ques-

tion of interest is whether there are trends in ξ1 and ξ2 over calendar years. Overallclimate trends could be reflected in trends in some of the FPC scores. We thereforeregress the scores ξ j against calendar time t (measured t in years) for j = 1,2,applying either linear or non-parametric regression.

Fitting simple linear regressions, we obtain the fits in Fig. 9 and the t-valuesof Table 2. While the maxima estimates are relatively flat (Fig. 2, middle plot) theminima process has a clear upward pattern along time. This pattern suggests thatnights are becoming warmer. In the temperature spread process the estimated β1show a clear decreasing trend. This shows that the daily spread is getting smaller inamplitude because the daily minima get closer to the daily maxima.

For the non-parametric regression approach we used a local linear smoother withGaussian kernel to compute smooth estimates for the time-trends in this data; thebandwidth used was equal to 12 years. Fig. 10 shows that the trends are indeedroughly linear in the case of temperature minima and temperature spread and haveincreasing and decreasing tendancies respectively, across the 20th century. On thecontrary, temperature maxima are out of sync between the two locations.

In a similar manner we examined the estimated slopes for the second modeof variation. While visually some slopes (Fig. 11, LHS) appear significant at first

Jan Feb Mar Apr May Jun Jul Aug Sep Oct Nov Dec

-0.1

-0.05

0

0.05

0.1

φ2 Minima

Davis 16%

Redwood City 17%

Jan Feb Mar Apr May Jun Jul Aug Sep Oct Nov Dec

-0.1

-0.05

0

0.05

0.1

φ2 Maxima

Davis 28%

Redwood City 18%

Jan Feb Mar Apr May Jun Jul Aug Sep Oct Nov Dec

-0.1

-0.05

0

0.05

0.1

φ2 Spread

Davis 19%

Redwood City 16%

Fig. 8 The second estimated eigenfunction for the daily temperature minima (LHS), maxima (cen-tral plot) and spread (RHS) as a function of the day of the year. The Fraction-of-Variance-Explainedfor each component is shown in the legend.

Page 18: Functional Data Analysis for Big Data: A case study on ...anson.ucdavis.edu/~mueller/pant.pdfThis latter point emphasizes the connection of FDA with big data. Big data are assumed

18 Pantelis Zenon Hadjipantelis and Hans-Georg Muller

glance, after correcting for multiple comparisons their effects are statistically in-significant.

Based on these findings we conclude that there are suggestions of warming effectin our current dataset. As mentioned before, several extrema influencing phenom-ena (eg. urban microclimate, soil moisture, etc.) have not been actively accountedfor; a full climate change assessment should definitely need to account for suchconfounders; for example it is known that the increasing irrigation in Californiamitigates nightly cooling.

t-values (p-values) Minima Maxima SpreadDavis 10.480 (<1e-9) -0.090 (0.928) -7.311 (<1e-9)

Redwood City 5.471 (<1e-5) 0.395 (0.694) -4.414 (<1e-4)

Table 2 The t-values and p-values for the slopes of the fitted simple linear regressions, assumingnormality for the first principal component scores ξ1.

As a final step of this analysis we study patterns of the random noise in the data;for example whether the behavior of the noise in the data so that we can assess po-

1900 1920 1940 1960 1980 2000

-30

-20

-10

0

10

20

30

ξ1 - Temp. Minima

Davis

Redwood City

1900 1920 1940 1960 1980 2000

-30

-20

-10

0

10

20

30

ξ1 - Temp. Maxima

Davis

Redwood City

1900 1920 1940 1960 1980 2000

-30

-20

-10

0

10

20

30

ξ1 - Temp. Spread

Davis

Redwood City

Fig. 9 The estimates for the first principal scores (ξ1) for daily temperature minima (LHS), max-ima (central plot) and spread (RHS), plotted against calendar year, and overlaid with least squaresfit of simple linear regression.

1900 1920 1940 1960 1980 2000

-30

-20

-10

0

10

20

30

ξ1 - Temp. Minima

Davis

Redwood City

1900 1920 1940 1960 1980 2000

-30

-20

-10

0

10

20

30

ξ1 - Temp. Maxima

Davis

Redwood City

1900 1920 1940 1960 1980 2000

-30

-20

-10

0

10

20

30

ξ1 - Temp. Spread

Davis

Redwood City

Fig. 10 The estimates for the first principal scores (ξ1) for daily temperature minima (LHS),maxima (central plot) and spread (RHS), plotted against calendar year, and overlaid with non-parametric fit of local linear kernel regression.

Page 19: Functional Data Analysis for Big Data: A case study on ...anson.ucdavis.edu/~mueller/pant.pdfThis latter point emphasizes the connection of FDA with big data. Big data are assumed

Functional Data Analysis for Big Data: A case study on California temperature trends 19

t-values (p-values) Minima Maxima SpreadDavis -1.642 (0.104) -0.168 (0.867) 1.904 (0.060)

Redwood City 1.463 (0.148) -0.318 (0.751) 1.019 (0.312)

Table 3 The t-values and p-values for the slopes of the fitted simple linear regressions, assumingnormality for the second principal component scores ξ2.

tential issues of heteroskedasticity; to see whether the behavior of temperature pat-terns over shorter time spans has become more erratic in recent times. We representeach temperature series with two trajectories; one for the actual smooth temperatureextrema process and a second one for the realization of the smooth variance processthat reflects patterns in the deviations from smooth temperature curves. We examinethe same three processes as previously: temperature daily minimum, maximum andspread.

Looking first at the respective mean processes we notice that they appear verysimilar between the two different locations (Fig. 12). This suggests that the noisevariance patterns are not location-specific between the two locations. This appearsin line with findings about the correlation of temperature anomalies for neighboringstations [38]; as Hansen et al. note: “[correlation] typically remains above 50% todistances of about 1200 miles at most latitudes”[39].

1900 1920 1940 1960 1980 2000

-30

-20

-10

0

10

20

30

ξ2 - Temp. Minima

Davis

Redwood City

1900 1920 1940 1960 1980 2000

-30

-20

-10

0

10

20

30

ξ2 - Temp. Maxima

Davis

Redwood City

1900 1920 1940 1960 1980 2000

-30

-20

-10

0

10

20

30

ξ2 - Temp. Spread

Davis

Redwood City

Fig. 11 The estimates second principal scores (ξ2) for the daily temperature minima (LHS), max-ima (central plot) and spread (RHS).

Jan Feb Mar Apr May Jun Jul Aug Sep Oct Nov Dec0

0.5

1

1.5

2

2.5

De

gre

es C

els

ius

Mean Daily Min. Variance Process

Davis

Redwood City

Jan Feb Mar Apr May Jun Jul Aug Sep Oct Nov Dec0

0.5

1

1.5

2

2.5

De

gre

es C

els

ius

Mean Daily Max. Variance Process

Davis

Redwood City

Jan Feb Mar Apr May Jun Jul Aug Sep Oct Nov Dec0

0.5

1

1.5

2

2.5

De

gre

es C

els

ius

Mean Daily Spread Variance Process

Davis

Redwood City

Fig. 12 The estimated mean functions for the functional variance processes V (•) in Eq. 20 for thedaily temperature minima (LHS) and maxima (RHS). The smoothing bandwidth for the Epanech-nikov kernel used was 63 days.

Page 20: Functional Data Analysis for Big Data: A case study on ...anson.ucdavis.edu/~mueller/pant.pdfThis latter point emphasizes the connection of FDA with big data. Big data are assumed

20 Pantelis Zenon Hadjipantelis and Hans-Georg Muller

In Davis the overall mean variance is typically slightly higher than in RedwoodCity, again hinting towards the tampering effect of the nearby sea mass in RedwoodCity. The minima variation trend appears more variable during the winter. Similarlythe variance in the maxima seems to be highest in late spring with a small maximumin early autumn.

Examining the principal modes of variation (Fig. 13 & 14) the Fraction-of-Variance-Explained (FVE) from the first estimated eigenfunction is comparable tothe FVE of the second estimated eigenfunction in the case of daily minimum anddaily maximum variation. This strongly suggests that two or more strong indepen-dent sources of variance are in place. In addition, particularly in the variance pro-cesses of the daily temperature maximum and temperature spread, there are clearlydefined peaks of variation as reflected in the first eigenfunction. The peaks are par-tially aligned in the daily temperature spread.The second estimated eigenfunctionfor all three variances processes shows strong seasonal patterns.

In terms of time-trends, perhaps counter-intuitively the variance processes appearto have clear downward trends (Fig. 15, Table 4 for the linear regression and Eq. 16for the non-parametric smoothed trends). This means that local variation from thesmooth process it at all, declined over the years. Confounders of such trends can be

Jan Feb Mar Apr May Jun Jul Aug Sep Oct Nov Dec-0.06

-0.04

-0.02

0

0.02

0.04

0.06

0.08

0.1

0.12

φ1 Minima

Davis 33%

Redwood City 41%

Jan Feb Mar Apr May Jun Jul Aug Sep Oct Nov Dec-0.06

-0.04

-0.02

0

0.02

0.04

0.06

0.08

0.1

0.12

φ1 Maxima

Davis 34%

Redwood City 32%

Jan Feb Mar Apr May Jun Jul Aug Sep Oct Nov Dec-0.06

-0.04

-0.02

0

0.02

0.04

0.06

0.08

0.1

0.12

φ1 Spread

Davis 49%

Redwood City 38%

Fig. 13 The first estimated eigenfunction for the variance process of daily temperature minima(LHS) and maxima (RHS) as a function of the day of the year. The Fraction-of-Variance-Explainedfrom each component in its respective process is shown in the legend.

Jan Feb Mar Apr May Jun Jul Aug Sep Oct Nov Dec-0.1

-0.08

-0.06

-0.04

-0.02

0

0.02

0.04

0.06

0.08

0.1

φ2 Minima

Davis 25%

Redwood City 22%

Jan Feb Mar Apr May Jun Jul Aug Sep Oct Nov Dec-0.1

-0.08

-0.06

-0.04

-0.02

0

0.02

0.04

0.06

0.08

0.1

φ2 Maxima

Davis 23%

Redwood City 22%

Jan Feb Mar Apr May Jun Jul Aug Sep Oct Nov Dec-0.1

-0.08

-0.06

-0.04

-0.02

0

0.02

0.04

0.06

0.08

0.1

φ2 Spread

Davis 18%

Redwood City 21%

Fig. 14 The second estimated eigenfunction for the variance process of daily temperature minima(LHS), maxima (RHS) as a function of the day of the year. The Fraction-of-Variance-Explainedfrom each component in its respective process is shown in the legend.

Page 21: Functional Data Analysis for Big Data: A case study on ...anson.ucdavis.edu/~mueller/pant.pdfThis latter point emphasizes the connection of FDA with big data. Big data are assumed

Functional Data Analysis for Big Data: A case study on California temperature trends 21

urbanisation, increased CO2 in the air or increased irrigation, all or some of whichmight conspire to force temperatures closer to their overall smooth trend. Analyzingthe second estimated eigenfunction similarly revealed no graphically obvious orstatistically significant time trends (analysis not shown).

t-values (p-values) Minima Maxima SpreadDavis -4.216 (<1e-4) 0.552 (0.582) -9.263 (<1e-9)

Redwood City -6.991 (<1e-7) 0.918 (0.362) -3.888 (<1e-3)

Table 4 The t-values and p-values for the slopes of the fitted simple linear regressions, assumingnormality for the second principal component scores ζ1 of the functional variance process.

Overall the FVPA shows that indeed there is a clear presence of structured noisein the data. In this section we showed how functional data analysis can be used ona (potentially) massive dataset to gain insights about a system that exhibits complextime dynamics.

1900 1920 1940 1960 1980 2000-10

-8

-6

-4

-2

0

2

4

6

8

10

ξ1 - Temp. Minima

Davis

Redwood City

1900 1920 1940 1960 1980 2000-10

-8

-6

-4

-2

0

2

4

6

8

10

ξ1 - Temp. Maxima

Davis

Redwood City

1900 1920 1940 1960 1980 2000-10

-8

-6

-4

-2

0

2

4

6

8

10

ξ1 - Temp. Spread

Davis

Redwood City

Fig. 15 The estimates for the first principal scores (ζ1) for the variance process of daily temper-ature minima (LHS), maxima (central plot) and spread (RHS), plotted against calendar year, andoverlaid with least squares fit of simple linear regression.

1900 1920 1940 1960 1980 2000-10

-8

-6

-4

-2

0

2

4

6

8

10

ξ1 - Temp. Minima

Davis

Redwood City

1900 1920 1940 1960 1980 2000-10

-8

-6

-4

-2

0

2

4

6

8

10

ξ1 - Temp. Maxima

Davis

Redwood City

1900 1920 1940 1960 1980 2000-10

-8

-6

-4

-2

0

2

4

6

8

10

ξ1 - Temp. Spread

Davis

Redwood City

Fig. 16 The estimates for the first principal scores (ζ1) for the variance process of daily temper-ature minima (LHS), maxima (central plot) and spread (RHS), plotted against calendar year, andoverlaid with non-parametric fit of local linear kernel regression.

Page 22: Functional Data Analysis for Big Data: A case study on ...anson.ucdavis.edu/~mueller/pant.pdfThis latter point emphasizes the connection of FDA with big data. Big data are assumed

22 Pantelis Zenon Hadjipantelis and Hans-Georg Muller

7 Conclusions

While in this work we focused on temperature data, it should be noted that func-tional data analysis applications are routinely used in areas like genomics and medi-cal imaging where massive datasets are typically encountered; compare [11]. As thepresence of sensors into mobile as well as wearable devices becomes more common-place, even larger data collections of function-valued characteristics are poised tobecome part of standard analytics tasks. Meaningful and principled techniques fromfunctional data analysis will therefore play an increasing role. Research of speedingup common FDA computations is still needed. With fast algorithms in place thatwill include proper parallelization, FDA offers an important analysis framework forbig-data analysts.

8 Acknowledgements

Pantelis Hadjipantelis and Hans-Georg Muller would like to thank Paul Ullrich,Dept. of Land, Air and Water Resources, UC Davis, for his comments on an earlydraft of this work.

References

1. Anestis Antoniadis, Gerard Gregoire, and Ian W McKeague. Wavelet methods for curve esti-mation. Journal of the American Statistical Association, 89(428):1340–1353, 1994.

2. John AD Aston, Jeng Min Chiou, and Jonathan P. Evans. Linguistic pitch analysis usingfunctional principal component mixed effect models. Journal of the Royal Statistical Society:Series C (Applied Statistics), 59(2):297–317, 2010.

3. David Barber. Bayesian Reasoning and Machine Learing. Cambridge University Press, 2012.Chapt. 12, 16, 19 & 21.

4. Leo Breiman. Bagging predictors. Machine Learning, 24:123–140, 1996.5. Andreas Bruns. Fourier-, Hilbert- and wavelet-based signal analysis: are they really different

approaches? Journal of Neuroscience Methods, 137(2):321 – 332, 2004.6. Peter Buhlmann and Sara Van De Geer. Statistics for high-dimensional data: methods, theory

and applications. Springer Science & Business Media, 2011.7. Richard Cangelosi and Alain Goriely. Component retention in principal component analysis

with application to cDNA microarray data. Biology Direct, 2:2+, 2007.8. Peter E Castro, WH Lawton, and EA Sylvestre. Principal modes of variation for processes

with continuous sample curves. Technometrics, 28(4):329–337, 1986.9. Dong Chen and Hans-Georg Muller. Nonlinear manifold representations for functional data.

The Annals of Statistics, 40(1):1–29, 2012.10. Kehui Chen and Hans-Georg Muller. Modeling repeated functional observations. Journal of

the American Statistical Association, 107(500):1599–1609, 2012.11. Kehui Chen, Xiaoke Zhang, Alexander Petersen, and Hans-Georg Muller. Quantifying

infinite-dimensional data: Functional data analysis in action. Statistics in Biosciences, pages1–23, 2015.

Page 23: Functional Data Analysis for Big Data: A case study on ...anson.ucdavis.edu/~mueller/pant.pdfThis latter point emphasizes the connection of FDA with big data. Big data are assumed

Functional Data Analysis for Big Data: A case study on California temperature trends 23

12. Jeng Min Chiou, Hans Georg. Muller, and Jane Ling Wang. Functional quasi-likelihood re-gression models with smooth random effects. Journal of the Royal Statistical Society: SeriesB Statistical Methodology), 65(2):405–423, 2003.

13. Norma Coffey, Andrew J Harrison, Orna A Donoghue, and Kathy Hayes. Common functionalprincipal components analysis: A new approach to analyzing human movement data. Humanmovement science, 30(6):1144–1166, 2011.

14. John P Cunningham, Vikash Gilja, Stephen I Ryu, and Krishna V Shenoy. Methods for esti-mating neural firing rates, and their application to brain–machine interfaces. Neural Networks,22(9):1235–1246, 2009.

15. Anthony Christopher Davison. Statistical models. Cambridge University Press, 2003. Chapt.4.

16. Pedro Delicado. Functional k-sample problem when data are density functions. Computa-tional Statistics, 22(3):391–410, 2007.

17. Pedro Delicado, Ramon Giraldo, Carlos Comas, and Jorge Mateu. Statistics for spatial func-tional data: some recent contributions. Environmetrics, 21(3-4):224–239, 2010.

18. Ian L Dryden. Statistical analysis on high-dimensional spheres and shape spaces. Annals ofStatistics, pages 1643–1665, 2005.

19. Ian L Dryden, Alexey Koloydenko, and Diwei Zhou. Non-euclidean statistics for covariancematrices, with applications to diffusion tensor imaging. The Annals of Applied Statistics, pages1102–1123, 2009.

20. Bradley Efron and Robert J Tibshirani. An introduction to the bootstrap. CRC press, 1994.21. Randall L Eubank. Nonparametric regression and spline smoothing. CRC press, 1999.22. Jianqing Fan and Irene Gijbels. Local polynomial modelling and its applications: monographs

on statistics and applied probability, volume 66. CRC Press, 1996.23. Louis Ferre and Anne-Francoise Yao. Functional sliced inverse regression analysis. Statistics,

37(6):475–488, 2003.24. David Freedman and Persi Diaconis. On the histogram as a density estimator: L 2 theory.

Probability theory and related fields, 57(4):453–476, 1981.25. Inc. Gartner. Gartner says solving ’big data’ challenge involves more than just managing

volumes of data. http://www.gartner.com/newsroom/id/1731916. Accessed: 2015-09-01.26. Theo Gasser and Hans-Georg Muller. Estimating regression functions and their derivatives by

the kernel method. Scandinavian Journal of Statistics, pages 171–185, 1984.27. Ali Ghodsi. Dimensionality reduction a short tutorial. Technical report, Department of Statis-

tics and Actuarial Science, Univ. of Waterloo, Ontario, Canada, 2006.28. Irving John Good. Some applications of the singular decomposition of a matrix. Technomet-

rics, 11(4):823–831, 1969.29. Esther Grabe, Greg Kochanski, and John Coleman. Connecting intonation labels to mathe-

matical descriptions of fundamental frequency. Language and Speech, 50(3):281–310, 2007.30. Spencer Graves, Giles Hooker, and James Ramsay. Functional data analysis with R and

MATLAB. Springer, New York, 2009.31. Wensheng Guo. Functional mixed effects models. Biometrics, 58:121–128, 2002.32. Pantelis Z Hadjipantelis, John AD Aston, and Jonathan P Evans. Characterizing fundamental

frequency in Mandarin: A functional principal component approach utilizing mixed effectmodels. Journal of the Acoustical Society of America, 131(6):4651–64, 2012.

33. Pantelis Z Hadjipantelis, Muller Hans-Georg Aston, John AD, and Jonathan P Evans. Unify-ing amplitude and phase analysis: A compositional data approach to functional multivariatemixed-effects modeling of mandarin chinese. Journal of the American Statistical Association,110(510):545–559, 2015.

34. Nathan Halko, Per-Gunnar Martinsson, Yoel Shkolnisky, and Mark Tygert. An algorithm forthe principal component analysis of large data sets. SIAM Journal on Scientific computing,33(5):2580–2594, 2011.

35. Peter Hall and Hans-Georg Muller. Order-preserving nonparametric regression, with appli-cations to conditional distribution and quantile function estimation. Journal of the AmericanStatistical Association, 98:598–608, 2003.

Page 24: Functional Data Analysis for Big Data: A case study on ...anson.ucdavis.edu/~mueller/pant.pdfThis latter point emphasizes the connection of FDA with big data. Big data are assumed

24 Pantelis Zenon Hadjipantelis and Hans-Georg Muller

36. Peter Hall, Hans Georg. Muller, and Jane Ling Wang. Properties of principal componentmethods for functional and longitudinal data analysis. The Annals of Statistics, 34(3):1493–1517, 2006.

37. Peter Hall, Hans-Georg Muller, and Fang Yao. Modelling sparse generalized longitudinalobservations with latent gaussian processes. Journal of the Royal Statistical Society: Series B(Statistical Methodology), 70(4):703–723, 2008.

38. James Hansen and Sergej Lebedeff. Global trends of measured surface air temperature. Jour-nal of Geophysical Research: Atmospheres (1984–2012), 92(D11):13345–13372, 1987.

39. James Hansen, Reto Ruedy, Mki Sato, and Ken Lo. Global surface temperature change. Re-views of Geophysics, 48(4), 2010.

40. Jaroslaw Harezlak, Michael C Wu, Mike Wang, Armin Schwartzman, David C Christiani, andXihong Lin. Biomarker discovery for arsenic exposure using functional data. Analysis andfeature learning of mass spectrometry proteomic data. The Journal of Proteome Research,7(1):217–224, 2008.

41. Lajos Horvath and Piotr Kokoszka. Inference for Functional Data with Applications, volume200. Springer Science & Business Media, 2012.

42. David C Hoyle. Automatic PCA dimension selection for high dimensional data and smallsample sizes. Journal of Machine Learning Research, 9(12):2733–2759, 2008.

43. Janine Illian, Erica Benson, John Crawford, and Harry Staines. Principal component analy-sis for spatial point processesassessing the appropriateness of the approach in an ecologicalcontext. In Case studies in spatial point process modeling, pages 135–150. Springer, 2006.

44. Alan J Izenman. Modern Multivariate Statistical Techniques: Regression, Classification andManifold Learning. Springer Verlag, New York, 2008. Chapt.6.

45. Gareth M James and Trevor J Hastie. Functional linear discriminant analysis for irregularlysampled curves. Journal of the Royal Statistical Society. Series B, Statistical Methodology,pages 533–550, 2001.

46. Gareth M James, Trevor J Hastie, and Catherine A Sugar. Principal component models forsparse functional data. Biometrika, 87(3):587–602, 2000.

47. Ci-Ren Jiang, John AD Aston, and Jane-Ling Wang. Smoothing dynamic positron emissiontomography time courses using functional principal components. NeuroImage, 47(1):184–193, 2009.

48. Ci-Ren Jiang, Wei Yu, Jane-Ling Wang, et al. Inverse regression for longitudinal data. TheAnnals of Statistics, 42(2):563–591, 2014.

49. Ian Jolliffe. Principal component analysis. Wiley Online Library, 2005. Chapt. 3.50. Kim Kenobi, Ian L Dryden, and Huiling Le. Shape curves and geodesic modelling.

Biometrika, 97(3):567–584, 2010.51. Jurgen Kleffe. Principal components of random variables with values in a separable Hilbert

space. Statistics: A Journal of Theoretical and Applied Statistics, 4:391–406, 1973.52. Christopher G. Knight, Rees Kassen, Holger Hebestreit, and Paul B. Rainey. Global analy-

sis of predicted proteomes: functional adaptation of physical properties. Proceedings of theNational Academy of Sciences of the United States of America, 101(22):8390–8395, 2004.

53. Ron Kohavi and Mehran Sahami. Error-based and entropy-based discretization of continuousfeatures. In KDD, pages 114–119, 1996.

54. Sebastian Kurtek, Eric Klassen, Zhaohua Ding, and Anuj Srivastava. A novel riemannianframework for shape analysis of 3d objects. In IEEE Conference on Computer Vision andPattern Recognition, CVPR 2010, pages 1625–1632. IEEE, 2010.

55. Daniel D Lee and H Sebastian Seung. Algorithms for non-negative matrix factorization. InAdvances in Neural Information Processing Systems, pages 556–562, 2001.

56. Daniel J Levitin, Regina L Nuzzo, Bradley W Vines, and James O Ramsay. Introduction tofunctional data analysis. Canadian Psychology/Psychologie canadienne, 48(3):135, 2007.

57. Ker-Chau Li. Sliced inverse regression for dimension reduction. Journal of the AmericanStatistical Association, 86(414):316–327, 1991.

58. Bitao Liu and Hans-Georg Muller. Estimating derivatives for samples of sparsely observedfunctions, with application to online auction dynamics. Journal of the American StatisticalAssociation, 104(486):704–717, 2009.

Page 25: Functional Data Analysis for Big Data: A case study on ...anson.ucdavis.edu/~mueller/pant.pdfThis latter point emphasizes the connection of FDA with big data. Big data are assumed

Functional Data Analysis for Big Data: A case study on California temperature trends 25

59. Xueli Liu and Hans-Georg Muller. Functional convex averaging and synchronization for time-warped random curves. Journal of the American Statistical Association, 99(467):687–699,2004.

60. Yanyuan Ma and Liping Zhu. A review on dimension reduction. International StatisticalReview, 81(1):134–150, 2013.

61. Michael W Mahoney. Randomized algorithms for matrices and data. Foundations and Trendsin Machine Learning, 3(2):123–224, 2011.

62. Stephane G Mallat. A Theory for Multiresolution Signal Decomposition: The Wavelet Repre-sentation. Pattern Analysis and Machine Intelligence, IEEE Transactions on, 11(7):674–693,1989.

63. Matthew J Menne, Imke Durre, Russell S Vose, Byron E Gleason, and Tamara G Houston. Anoverview of the global historical climatology network-daily database. Journal of Atmosphericand Oceanic Technology, 29(7):897–910, 2012.

64. James Mercer. Functions of positive and negative type, and their connection with the theoryof integral equations. Philosophical Transactions of the Royal Society of London. Series A,Containing Papers of a Mathematical or Physical Character, 209:415–446, 1909.

65. Thomas P Minka. Automatic choice of dimensionality for PCA. Advances in Neural Infor-mation Processing Systems, 15:598–604, 2001.

66. Jeffrey S. Morris and Raymond J. Carroll. Wavelet-based functional mixed models. Journalof the Royal Statistical Society, Series B, 68:179–199, 2006.

67. Hans-Georg Muller, Ulrich Stadtmuller, and Fang Yao. Functional variance processes. Journalof the American Statistical Association, 101(475):1007–1018, 2006.

68. Hans-Georg Muller, Shuang Wu, Alexandros D Diamantidis, Nikos T Papadopoulos, andJames R Carey. Reproduction is adapted to survival characteristics across geographically iso-lated medfly populations. Proceedings of the Royal Society of London B: Biological Sciences,page rspb20091461, 2009.

69. Christos H Papadimitriou. Computational complexity. John Wiley and Sons Ltd., 2003. Chapt.11.

70. Alexander Petersen and Hans-Georg Muller. Functional data analysis for density functions bytransformation to a hilbert space. The Annals of Statistics, 44(1):183–218, 2016.

71. Sonia Petrone, Michele Guindani, and Alan E Gelfand. Hybrid dirichlet mixture models forfunctional data. Journal of the Royal Statistical Society: Series B (Statistical Methodology),71(4):755–782, 2009.

72. Davide Pigoli and Laura M Sangalli. Wavelets in functional data analysis: Estimation ofmultidimensional curves and their derivatives. Computational Statistics & Data Analysis,56(6):1482–1498, 2012.

73. James O Ramsay. Multilevel modeling of longitudinal and functional data. Modeling intrain-dividual variability with repeated measures data: Methods and applications, pages 171–201,2002.

74. James O Ramsay, Kevin G Munhall, Vincent L Gracco, and David J Ostry. Functional dataanalyses of lip motion. The Journal of the Acoustical Society of America, 99(6):3718–3727,1996.

75. James O Ramsay and Bernard W Silverman. Functional Data Analysis. Springer Verlag, NewYork, 2005. Chapt. 3, 4 & 7.

76. C Radhakrishna Rao. The theory of least squares when the parameters are stochastic and itsapplication to the analysis of growth curves. Biometrika, 52(3/4):447–458, 1965.

77. Carl Edward Rasmussen. Gaussian processes in machine learning. In Advanced lectures onmachine learning, pages 63–71. Springer, 2004.

78. Sam T Roweis and Lawrence K Saul. Nonlinear dimensionality reduction by locally linearembedding. Science, 290(5500):2323–2326, 2000.

79. Laura M Sangalli, James O Ramsay, and Tim O Ramsay. Spatial spline regression models.Journal of the Royal Statistical Society. Series B (Methodological), 75(4):1–23, 2013.

80. Bernhard Scholkopf, Alexander Smola, and Klaus-Robert Muller. Nonlinear component anal-ysis as a kernel eigenvalue problem. Neural computation, 10(5):1299–1319, 1998.

Page 26: Functional Data Analysis for Big Data: A case study on ...anson.ucdavis.edu/~mueller/pant.pdfThis latter point emphasizes the connection of FDA with big data. Big data are assumed

26 Pantelis Zenon Hadjipantelis and Hans-Georg Muller

81. David W Scott. On optimal and data-based histograms. Biometrika, 66(3):605–610, 1979.82. Bernard W Silverman. Some aspects of the spline smoothing approach to non-parametric

regression curve fitting. Journal of the Royal Statistical Society. Series B (Methodological),pages 1–52, 1985.

83. Herbert A Sturges. The choice of a class interval. Journal of the American Statistical Associ-ation, 21(153):65–66, 1926.

84. Joshua B Tenenbaum, Vin De Silva, and John C Langford. A global geometric framework fornonlinear dimensionality reduction. science, 290(5500):2319–2323, 2000.

85. Susanne Theis. Deriving probabilistic short-range forecasts from a deterministic high-resolution model. PhD thesis, University of Bonn - Universitat Bonn, 2005. PhD Thesis.

86. Robert Tibshirani. Regression shrinkage and selection via the lasso. Journal of the RoyalStatistical Society. Series B (Methodological), 58(1):pp. 267–288, 1996.

87. Ledyard R Tucker. Determination of parameters of a functional relationship by factor analysis.Psychometrika, 23:19–23, 1958.

88. Mariano J. Valderrama. An overview to modelling functional data. Computational Statistics,22(3):331–334, 2007.

89. Shanshan Wang, Wolfgang Jank, and Galit Shmueli. Explaining and forecasting online auctionprices and their dynamics using functional data analysis. Journal of Business & EconomicStatistics, 26(2), 2008.

90. Kilian Q Weinberger, Fei Sha, and Lawrence K Saul. Learning a kernel matrix for nonlin-ear dimensionality reduction. In Proceedings of the twenty-first international conference onMachine learning, page 106. ACM, 2004.

91. Alle Meije Wink and Jos B T M Roerdink. Denoising functional MR images: A Comparisonof Wavelet Denoising and Gaussian Smoothing. Medical Imaging, IEEE Transactions on,23(3):374–387, 2004.

92. Rafi Witten and Emmanuel Candes. Randomized algorithms for low-rank matrix factoriza-tions: sharp performance bounds. Algorithmica, 72(1):264–281, 2015.

93. Keith J Worsley, CH Liao, J Aston, V Petre, GH Duncan, F Morales, and AC Evans. A generalstatistical analysis for fmri data. Neuroimage, 15(1):1–15, 2002.

94. Wenjing Yang, Hans-Georg Muller, and Ulrich Stadtmuller. Functional singular compo-nent analysis. Journal of the Royal Statistical Society: Series B (Statistical Methodology),73(3):303–324, 2011.

95. Fang Yao, Hans Georg Muller, and Jane-Ling Wang. Functional data analysis for sparselongitudinal data. Journal of the American Statistical Association, 100(470):577–590, 2005.

96. Xianyang Zhang, Xiaofeng Shao, Katharine Hayhoe, and Donald J Wuebbles. Testing thestructural stability of temporally dependent functional observations and application to climateprojections. Electronic Journal of Statistics, 5:1765–1796, 2011.