26
Data fusion for remote sensing applications Anne H. S. Solberg March 17, 2006 Abstract With a growing number of satellite sensors the coverage of the earth in space, time and the electromagnetic spectrum is increasing fast. To be able to utilize all this information, a number of approaches for data fusion have been presented. The “multi” concept for remote sensing applications refers to multisource, multiscale, multipolarization, multifrequency, and multitemporal imagery. We present and discuss methods for multisource image analysis and provide a tutorial on the subject on data fusion for remote sensing. The main focus is on methods for multisource, multiscale and multitemporal image classi- fication. Guidelines to be used in choosing the best architecture and approach for data fusion for a given application are provided. 1 Introduction Earth observation is currently developing more rapidly than ever before. During the last decade and the near future the number of satellites has been growing steadily, and the coverage of the Earth in space, time and the electromagnetic spectrum is increasing correspondingly fast. The accuracy in classifying a scene can be increased by using images from sev- eral sensors operating at different wavelengths of the electromagnetic spectrum. The interaction between the electromagnetic radiation and the earth’s surface is character- ized by certain properties at different frequencies of electromagnetic energy. Sensors with different wavelengths provide complementary information about the surface. In addition to image data, prior information about the scene might be available in the form of map data from Geographic Information Systems (GIS). The merging of mul- tisource data can create a more consistent interpretation of the scene compared to an interpretation based on data from a single sensor. This development opens up for a potential significant change in the approach of analysis of earth observation data. Traditionally, analysis of such data bas been by means of analysis of a single satellite image. The emerging exceptionally good coverage in space, time and the spectrum opens for analysis of time series of data, combining different sensor types, combining imagery of different scales, and better * Anne Solberg is with Department of Informatics, University of Oslo, 0314 Oslo, Norway, and with the Norwegian Computing Center, 0314 Oslo, Norway (e-mail: anne@ifi.uio.no) 1

Data fusion for remote sensing applications - … · 4.1 A general introduction to multisensor data fusion for remote sensing applications Fusion can be performed at either the signal,

  • Upload
    buiphuc

  • View
    224

  • Download
    0

Embed Size (px)

Citation preview

Page 1: Data fusion for remote sensing applications - … · 4.1 A general introduction to multisensor data fusion for remote sensing applications Fusion can be performed at either the signal,

Data fusion for remote sensing applications

Anne H. S. Solberg∗

March 17, 2006

Abstract

With a growing number of satellite sensors the coverage of the earth inspace, time and the electromagnetic spectrum is increasing fast. To be ableto utilize all this information, a number of approaches for data fusion havebeen presented. The “multi” concept for remote sensing applications refers tomultisource, multiscale, multipolarization, multifrequency, and multitemporalimagery. We present and discuss methods for multisource image analysis andprovide a tutorial on the subject on data fusion for remote sensing. The mainfocus is on methods for multisource, multiscale and multitemporal image classi-fication. Guidelines to be used in choosing the best architecture and approachfor data fusion for a given application are provided.

1 Introduction

Earth observation is currently developing more rapidly than ever before. During thelast decade and the near future the number of satellites has been growing steadily,and the coverage of the Earth in space, time and the electromagnetic spectrum isincreasing correspondingly fast.

The accuracy in classifying a scene can be increased by using images from sev-eral sensors operating at different wavelengths of the electromagnetic spectrum. Theinteraction between the electromagnetic radiation and the earth’s surface is character-ized by certain properties at different frequencies of electromagnetic energy. Sensorswith different wavelengths provide complementary information about the surface. Inaddition to image data, prior information about the scene might be available in theform of map data from Geographic Information Systems (GIS). The merging of mul-tisource data can create a more consistent interpretation of the scene compared toan interpretation based on data from a single sensor.

This development opens up for a potential significant change in the approachof analysis of earth observation data. Traditionally, analysis of such data bas beenby means of analysis of a single satellite image. The emerging exceptionally goodcoverage in space, time and the spectrum opens for analysis of time series of data,combining different sensor types, combining imagery of different scales, and better

∗Anne Solberg is with Department of Informatics, University of Oslo, 0314 Oslo, Norway, andwith the Norwegian Computing Center, 0314 Oslo, Norway (e-mail: [email protected])

1

Page 2: Data fusion for remote sensing applications - … · 4.1 A general introduction to multisensor data fusion for remote sensing applications Fusion can be performed at either the signal,

integration with ancillary data and models. Thus, data fusion to combine data fromseveral sources is becoming increasingly more important in many remote sensingapplications.

This paper provides a tutorial on data fusion for remote sensing application. Themain focus is on methods for multisource image classification, but separate sectionson multisensor image registration, multiscale classification, and multitemporal imageclassification are also included. The remainder of this chapter is organized in thefollowing manner: In section 2 the “multi” concept in remote sensing is presented.Multisensor data registration is treated in section 3. Classification strategies for mul-tisensor applications are discussed in section 4. Multitemporal image classifiation isdiscussed in section 5, while multiscale approaches are discussed in section 6. Con-cluding remarks are given in section 7.

2 The “multi” concept in remote sensing

The variety of different sensors already available or being planned creates a numberof possibilities for data fusion to provide better capabilities for scene interpretation.This is referred to as the “multi” concept in remote sensing. The “multi” conceptincludes: multitemporal, multispectral/multifrequency, multipolarization, multiscale,and multisensor image analysis. In addition to the concepts discussed here, imagingusing multiple incidence angles can also provide additional information [21, 28].

2.1 The multispectral/multifrequency aspect

The measured backscatter values for an area varies with the wavelength band. A land-use category will give different image signal depending on the frequency used, andby using different frequencies, a spectral signature which characterizes the land-usecategory can be found. A description of the scattering mechanisms for optical sensorscan be found in [3], while [75] contains a thorough discussion of backscattering mech-anisms in the microwave region. Multispectral optical sensors have demonstrated thiseffect for a substantial number of applications for several decades, they are now fol-lowed by high-spatial-resolution multispectral sensors like e.g. Ikonos and Quickbird,and by hyperspectral sensors from satellite platforms (e.g. Hyperion).

2.2 The multitemporal aspect

The term multitemporal refers to the repeated imaging of an area over a time period.By analyzing an area through time, it is possible to develop interpretation techniquesbased on an object’s temporal variations, and to discriminate different pattern classesaccordingly. Multitemporal imagery allows the study of how the backscatter of differ-ent area varies with time, weather conditions and seasons. It also allows monitoringof processes changing over time.

The principal advantage of multitemporal analysis is the increased amount of in-formation for the study area. The information provided for a single image is, forcertain applications, not sufficient to properly distinguish between the desired pat-tern classes. This limitation can sometimes be resolved by examining the pattern of

2

Page 3: Data fusion for remote sensing applications - … · 4.1 A general introduction to multisensor data fusion for remote sensing applications Fusion can be performed at either the signal,

temporal changes in the spectral signature of an object. This is particularly impor-tant for vegetation applications. Multitemporal image analysis is discussed in moredetail in section 5.

2.3 The multipolarization aspect

The multipolarization aspect is related to microwave image data. The polarizationof an electromagnetic wave refers to the orientation of the electric field during prop-agation. A review of the theory and features of polarization is given in [74, 82].

2.4 The multisensor aspect

With an increasing number of operational and experimental satellites, informationabout a phenomenon can be captured using different types of sensors.

Fusion of images from different sensors requires some additional preprocessingand poses certain difficulties which are not solved in traditional image classifiers.Each sensor has its own characteristics and the image captured usually containsvarious artifacts which should be corrected or removed. The images also need to begeometrically corrected and co-registered. Since the multisensor images often are notacquired on the same data, the multitemporal nature of the data must often also beaccounted for.

Figure 1 shows a simple visualization of two Synthetic Aperture Radar imagesfrom an oil spill in the Baltic sea, imaged by the ENVISAT ASAR sensor and theRadarsat SAR sensor. The images were taken a few hours apart. During this time,the oil slick has drifted somewhat, and it has become more irregular in shape.

2.5 Other sources of spatial data

The preceding sections have addressed spatial data in the form of digital imagesobtained from remote sensing satellites. For most regions, additional informationis available in the form of various kinds of maps, e.g. topography, ground cover,elevation, etc. Frequently, maps contain spatial information not obtainable from asingle remotely sensed image. Such maps represent a valuable information resourcein addition to the satellite images. To integrate map information with a remotelysensed image, the map must be available in digital form, e.g. in a GIS system.

3 Multisensor data registration

A prerequisite for data fusion is that the data are co-registered, and geometricallyand radiometrically corrected. Data co-registration can be simple if the data are geo-referenced. In that case, the co-registration consists merely of resampling the imagesto a common map projection. However, an image matching step is often necessary toobtain subpixel accuracy in matching. Complicating factors for multisensor data arethe different appearance of the same object imaged by different sensors, and non-rigidchanges in object position between multitemporal images.

3

Page 4: Data fusion for remote sensing applications - … · 4.1 A general introduction to multisensor data fusion for remote sensing applications Fusion can be performed at either the signal,

Figure 1: Example of multisensor visualization of an oil spill in the Baltic Sea createdby combining an ENVISAT ASAR image with a Radarsat SAR image take a few hourslater.

The image resampling can be done at various stages of the image interpretationprocess. Resampling an image affects the spatial statistics of neighboring pixel, whichis of importance for many radar image feature extraction methods that might usespeckle statistics or texture. When fusing a radar image with other data sources, asolution might be to transform the other data sources to the geometry of the radarimage. When fusing a multitemporal radar image, an alternative might be to useimages from the same image mode of the sensor, e.g. only ascending scenes with agiven incidence angle range. If this cannot be done and the spatial information fromthe original geometry is important, the data can be fused and resampling done afterclassification by the sensor-specific classifiers.

A image matching step can be necessary to achieve sub-pixel accuracy in the co-registration even if the data are georeferenced. A survey of image registration methodsis given in Zitova and Flusser [84]. A full image registration process generally consistsof four steps:

1. Feature extraction. This is the step where regions, edges, contours etc. that canbe used to represent tie-points in the set of images to be matched are extracted.This is a crucial step, as the registration accuracy can be no better than whatis achieved for the tie points.

Feature extraction can be grouped into area-based methods [18, 19], feature-based methods [24,49,59], and hybrid approaches [84]. In area-based methods,

4

Page 5: Data fusion for remote sensing applications - … · 4.1 A general introduction to multisensor data fusion for remote sensing applications Fusion can be performed at either the signal,

the grey levels of the images are used directly for matching, often by statisti-cal comparison of pixel values in small windows, and they are best suited forimages from the same or highly similar sensors. Feature-based methods will beapplication dependent, as the type of features to use as tie points needs to betailored to the application. Features can be extracted either from the spatialdomain (edges, lines, regions, intersections etc.) or from the frequency domain(e.g. wavelet features). Spatial features can perform well for matching datafrom heterogeneous sensors, e.g. optical and radar images. Hybrid approachesuse both area-based and feature-based techniques, e.g. by combining both acorrelation-based matching with an edge-based approach, and they are usefulfor matching data from heterogeneous sensors.

2. Feature matching. In this step, the correspondence between the tie-points orfeatures in the sensed image and the reference image is found. Area-basedmethods for feature extraction use either correlation, Fourier-transform meth-ods or optical flow [44]. Feature-based methods use the equivalence betweencorrelation in the spatial domain and multiplication in the Fourier-domain toperform matching in the Fourier domain [24, 49]. Correlation-based methodsare best suited for data from similar sensors. The optical flow approach involvesestimation of the relative motion between two images and is a broad approach.It is commonly used in video analysis, but only a few studies have used it forremote sensing applications [29, 30].

3. Transformation selection concerns the choice of mapping function and estima-tion of its parameters based on the established feature correspondence. Theaffine transform model is commonly used for remote sensing applications, wherethe images normally are pre-processed for geometrical correction, a step thatjustifies the use of affine transforms.

4. Image resampling. In this step, the image is transformed by means of themapping function. Image values in no-integer coordinates are computed bythe appropriate interpolation technique. Normally, either a nearest neighboror a bilinear interpolation is used. Nearest neighbor interpolation is applicablewhen no new pixel values should be introduced. Bilinear interpolation is oftena good trade-off between accuracy and computational complexity compared toe.g cubic or higher order interpolation.

4 Multisensor image classification

The literature on data fusion in the computer vision and machine intelligence do-mains is substantial. For an extensive review of data fusion, we recommend thebook by Abidi and Gonzales [1]. Multi-sensor architectures, sensor management, anddesigning sensor setup is also thoroughly discussed in [81].

5

Page 6: Data fusion for remote sensing applications - … · 4.1 A general introduction to multisensor data fusion for remote sensing applications Fusion can be performed at either the signal,

4.1 A general introduction to multisensor data fusion forremote sensing applications

Fusion can be performed at either the signal, pixel, feature or decision level of rep-resentation (see Figure 2). In signal-based fusion, signals from different sensors arecombined to create a new signal with a better signal-to-noise ratio than the originalsignals [58]. Techniques for signal-level data fusion typically involve classic detectionand estimation methods [35]. If the data are non-commensurate, they must be fusedat a higher level.

Pixel-based fusion consists of merging information from different images on apixel-by-pixel basis to improve the performance of image processing tasks such assegmentation [46]. Feature-based fusion consists of merging features extracted fromdifferent signals or images [5]. In feature-level fusion, features are extracted frommultiple sensor observations, then combined into a concatenated feature vector andclassified using a standard classifier. Symbol-level or decision-level fusion consistsof merging information at a higher level of abstraction. Based on the data fromeach single sensor, a preliminary classification is performed. Fusion then consists ofcombining the outputs from the preliminary classifications.

The main approaches to data fusion in the remote sensing literature are statisticalmethods [10,68,70,80], Dempster-Shafer theory [27,37,43], and neural networks [10,61]. We will discuss each of these approaches in the following sections. The best leveland methodology for a given remote sensing application depends on several factors:the complexity of the classification problem, the available data set, and the goal ofthe analysis.

4.2 Decision-level data fusion for remote sensing applications

In the general multisensor fusion case, we have a set of images X1...XP from P sensors.The class labels of the scene are denoted C. The Bayesian approach is to assign eachpixel to the class that maximizes the posterior probabilities P (C|X1, .....XP )

P (C|X1, ...., XP ) =P (X1, ..., XP |C)P (C)

P (X1, ..., XP ), (1)

where P (C) is the prior model for the class labels.For decision-level fusion, the following conditional independence assumption is

used.

P (X1, ..., XP | C) ≡ P (X1 | C) · · ·P (XP | C).

This assumption means that the measurements from the different sensors areconsidered to be conditionally independent.

6

Page 7: Data fusion for remote sensing applications - … · 4.1 A general introduction to multisensor data fusion for remote sensing applications Fusion can be performed at either the signal,

Figure 2: An illustration of data fusion on different levels.

7

Page 8: Data fusion for remote sensing applications - … · 4.1 A general introduction to multisensor data fusion for remote sensing applications Fusion can be performed at either the signal,

Figure 3: Schematic view of weighting the outputs from sensor-specific classifiers indecision-level fusion.

4.3 Combination schemes for combining classifier outputs

In the data fusion literature (see, e.g., [9]), various alternative methods have beenproposed for combining the outputs from the sensor-specific classifiers by weightingthe influence of each sensor. This is termed consensus theory. The weighting schemescan be linear, logarithmic or of a more general form (see Figure 3).

The simplest choice, the linear opinion pool (LOP), is given by

LOP (X1, ...., XP ) =

P∑

p=1

P (Xp|C)λp. (2)

The logarithmic opinion pool (LOGP) is given by

LOGP (X1, ...., XP ) =

P∏

p=1

P (Xp|C)λp, (3)

which is equivalent to the Bayesian combination if the weights λp are equal. Thisweighting scheme contradicts the statistical formulation in which the sensor’s uncer-tainty is supposed to be modeled by the variance of the probability density function.

The weights are supposed to represent the sensor’s reliability. The weights can beselected by heuristic methods based on e.g. their goodness [6] by weighting a sensor’sinfluence by a factor proportional to its overall classification accuracy on the trainingdata set. An alternative approach for a linear combination pool is to use a geneticalgorithm [73].

An approach using a neural net to optimize the weights is presented in [9]. Yetanother possibility is to choose the weights in such a way that they not only weighthe individual data sources but also the classes within the data sources [53].

Benediktsson et al. [6, 9] use a multilayer perceptron (MLP) neural network tocombine the class-conditional probability densities P (Xp|C). This allows a moreflexible, non-linear combination scheme. They compare the classification accuracyusing MLPs to linear and logarithmic opinion pools, and find that the neural netcombination performs best.

8

Page 9: Data fusion for remote sensing applications - … · 4.1 A general introduction to multisensor data fusion for remote sensing applications Fusion can be performed at either the signal,

Benediktsson and Sveinsson [7] provide a comparison of different weighting schemesfor a linear and logarithmic opinion pool, genetic algorithm with and without prun-ing, parallel consensus neural nets, and conjugate gradient back propagation (CGBP)nets on a single multisource data set. The best results were achieved by using a CGBPnet to optimize the weights in a logarithmic opinion pool.

A study that contradicts the weighting of different sources is found in [65]. In thisstudy, three different data sets (optical and radar) were merged using the logarithmicopinion pool, and the weights were varied between 0 and 1. Best results for all threedata sets were found by using equal weights.

4.4 Statistical multisource classification

Statistical methods for fusion of remotely sensed data can be divided into four cat-egories: the augmented vector approach, stratification, probabilistic relaxation, andextended statistical fusion. In the augmented vector approach, data from differentsources are concatenated as if they were measurements from a single sensor. This isthe most common approach for many application-oriented applications of multisourceclassification, because no special software is needed. This is an example of pixel-levelfusion.

Such a classifier is difficult to use when the data cannot be modelled with acommon probability density function, or when the data set includes ancillary data,e.g. from a GIS system. The fused data vector is then classified using ordinary single-source classifiers [45]. Stratification has been used to incorporate ancillary GIS datain the classification process. The GIS data are stratified into categories and then aspectral model for each of these categories is used [31].

Richards et al. [57] extended the methods used for spatially contextual classifi-cation based on probabilistic relaxation to incorporate ancillary data. The methodsbased on extended statistical fusion [10, 43] were derived by extending the conceptsused for classification of single-sensor data. Each data source is considered indepen-dently, and the classification results are fused using weighted linear combinations.

By using a statistical classifier one often assumes that the data has a multivariateGaussian distribution. Recent developments in statistical classifiers based on regres-sion theory include choices of non-linear classifiers [32]- [60]. For a comparison ofneural nets and regression-based non-linear classifiers, see [66].

4.5 Neural nets for multisource classification

Many multisensor studies have used neural nets because no specific assumptions aboutthe underlying probability densities are needed (see e.g. [33,63]). A drawback of neu-ral nets in this respect is that they act like a black box in that the user cannot controlhow the different data sources are used. It is also difficult to explicitly use a spatialmodel for neighboring pixels (but one can extend the input vector from measurementsfrom a single pixel to measurements from neighboring pixels). Guan et al. [34] uti-lized contextual information by using a network of neural networks with which theybuilt a quadratic regularizer. Another drawback is that specifying a neural networkarchitecture involves specifying a large number of parameters. A classification exper-iment should take care in choosing them and apply different configurations, making

9

Page 10: Data fusion for remote sensing applications - … · 4.1 A general introduction to multisensor data fusion for remote sensing applications Fusion can be performed at either the signal,

the complete training process very time consuming (see [52,63]). Hybrid approachescombining statistical methods and neural networks for data fusion have also been pro-posed [9]. Benediktsson et al. [9] apply a statistical model to each individual source,and use neural nets to reach a consensus decision. Most applications involving aneural net use a multilayer perceptron or radial basis function network, but otherneural network architectures can be used (see, e.g., [8, 17,78]).

Neural nets for data fusion can be applied both at the pixel-, feature-, and deci-sion level. For pixel- and feature-level fusion a single neural net is used to classify thejoint feature vector or pixel measurement vector. For decision-level fusion, a networkcombination like the one outlined in Figure 4 is often used [61]. A multilayer percep-tron neural net is first used to classify the images from each source separately. Then,the outputs from the sensor-specific nets are fused and weighted in a fusion network.

Figure 4: Network architecture for decision-level fusion using neural networks.

4.6 A closer look at Dempster-Shafer evidence theory for

data fusion

Dempster-Shafer theory of evidence provides a representation of multisource datausing two central concepts: plausibility and belief. Mathematical evidence theorywas first introduced by Dempster in the 1960’s, and later extended by Shafer [62].

A good introduction to Dempster-Shafer evidence theory for remote sensing datafusion is given in [37].

Plausibility (Pls) and belief (Bel) are derived from a mass function m, which isdefined on the [0, 1] interval. The belief and plausibility functions for an element Aare defines as:

Bel(A) =∑

B⊆A

m(B) (4)

10

Page 11: Data fusion for remote sensing applications - … · 4.1 A general introduction to multisensor data fusion for remote sensing applications Fusion can be performed at either the signal,

P ls(A) =∑

B∩A 6=∅

m(B). (5)

They are sometimes referred to as lower and upper probability functions. Thebelief value of hypothesis A can be interpreted as the minimum uncertainty valueabout A, and its plausibility as the maximum uncertainty [37].

Evidence from p different sources is combined by combining the mass functionsm1...mp by

{

m(∅) = 0

If K 6= 1, m(A) =P

B1∩···Bp=A

Q

1≤i≤p mi(Bi)

1−K

where K =∑

B1∩···∩Bp=∅

1≤i≤p mi(Bi) is interpreted as a measure of conflictbetween the different sources.

The decision rule used to combine the evidence from each sensor varies fromdifferent applications, either maximum of plausibility or maximum of belief (withvariations). The performance of Dempster-Shafer theory for data fusion does howeverdepend on the methods used to compute the mass functions. Lee et al. [46] assign non-zero mass function values only to the single classes, while [37] proposes two strategiesfor assigning mass function values to sets of classes according to the membership fora pixel for these classes.

The concepts of evidence theory belongs to a different school than Bayesian mul-tisensor models. Researchers coming from one school often have a tendency to dislikemodelling used in the alternative theory. Not many neutral comparisons of these twoapproaches exist. The main advantage of this approach is its robustness in the wayinformation from several heterogeneous sources is combined. A disadvantage is theunderlying basic assumption that the evidence from different sources is independent.According to [38], Bayesian theory assumes that imprecision about uncertainty in themeasurements is assumed to be null and uncertainty about an event is only measuredby the probability. The author disagrees with this by pointing out that in Bayesianmodelling, uncertainty about the measurements can be modelled in the priors. Priorsof this kind are not always used, however. Priors in a Bayesian model can also beused to model spatial context and temporal class development. It might be arguedthat Dempster-Shafer theory can be more appropriate for a high number of hetero-geneous sources. However, most papers on data fusion for remote sensing considertwo or maximum three different sources.

4.7 Contextual methods for data fusion

Remote sensing data have an inherent spatial nature. To account for this, contextualinformation can be incorporated in the interpretation process. Basically, the effect ofcontext in an image labelling problem is that when a pixel is considered in isolation, itmay provide incomplete information about the desired characteristics. By consideringthe pixel in context with other measurements, more complete information might bederived.

Only a limited set of studies have involved spatially contextual multisource clas-sification. Richards et. al [57] extended the methods used for spatial contextualclassification based on probabilistic relaxation to incorporate ancillary data. Binaghi

11

Page 12: Data fusion for remote sensing applications - … · 4.1 A general introduction to multisensor data fusion for remote sensing applications Fusion can be performed at either the signal,

et al. [13] presented a knowledge-based framework for contextual classification basedon fuzzy set theory. Wan and Fraser [78] used multiple self-organizing maps for con-textual classification. Le Hegarat-Mascle et al. [37] combined the use of a Markovrandom field model with Dempster-Shafer theory. Smits and Dellepiane [64] used amultichannel image segmentation method based on Markov random fields with adap-tive neighborhoods. Markov random fields have also been used for data fusion inother application domains [20, 79].

4.8 Using Markov random fields to incorporate ancillary data

Solberg et al. [67, 69] used a Markov random field model to include map data intothe fusion. In this framework, the task is to estimate the class labels of the scene Cgiven the image data X and the map data M (from a previous survey):

P (C|X,M) = P (X|C,M)P (C)

with respect to C.The spatial context between neighboring pixels in the scene is modelled in P (C)

using the common Ising model. By using the equivalence between Markov randomfields and Gibbs distribution

P (·) =1

Zexp−U(·),

where U is called the energy function and Z is a constant, the task of maximizingP (C|X,M) is equivalent to minimizing the sum

U =P

i=1

Udata(i) + Uspatial,map.

Uspatial is the common Ising model:

Uspatial = βs

k∈N

I(ci, ck)

andUmap = βm

k∈M

t(ci|mk).

mk is the class assigned to the pixel in the map, and t(ci|mk) is the probability ofa class transition from class mk to class ci. This kind of model can also be used formultitemporal classification [69].

4.9 A summary of data fusion architectures

Table 1 gives a schematic view on different fusion architectures applied to remotesensing data.

12

Page 13: Data fusion for remote sensing applications - … · 4.1 A general introduction to multisensor data fusion for remote sensing applications Fusion can be performed at either the signal,

Pixel level fusion

Advantages Simple. No special classifier software needed.Correlation between sources utilized.Well suited for change detection.

Limitations Assumes that the data can be modelled usinga common probability density function.Source reliability cannot be modelled.

Feature level fusion

Advantages Simple. No special classifier software needed.Sensor-specific features gives advantage over pixel-based fusion.Well suited for change detection.

Limitations Assumes that the data can be modelled usinga common probability density function.Source reliability cannot be modelled.

Decision level fusion

Advantages Suited for data with different probability densities.Source-specific reliabilities can be modelled.Prior information about the source combination can be modelled.

Limitations Special software often needed.

Table 1: A summary of data fusion architectures

5 Multitemporal image classification

For most applications where multisource data are involved, it is not likely that allthe images are acquired at the same time. When the temporal aspect is involved, theclassification methodology must handle changes in pattern classes between the imageacquisitions, and possibly also use different classes.

To find the best classification strategy for a multitemporal data set, it is useful toconsider what the goal of the analysis is and how complex the multitemporal imagedata to be used might be. Multitemporal image classification can be applied fordifferent purposes:

• Monitor/identify specific changes. If the goal is to monitor changes, multitem-poral data are required, either in the form of a combination of existing mapsand new satellite imagery, or a set of satellite images. For identifying changes,different fusion levels can be considered. A high number of methods for changedetection exists, ranging from pixel-level to decision-level fusion. Examples ofpixel-level change detection are classical unsupervised approaches like imagemath, image regression, and principal component analysis of a multitemporalvector of spectral measurements or derived feature vectors like normalized vege-tation indexes. In this paper, we will not go into detail on these well-establishedunsupervised methods. Decision-level change detection includes post classifica-tion comparisons, direct multidate classification and more sophisticated classi-fiers.

13

Page 14: Data fusion for remote sensing applications - … · 4.1 A general introduction to multisensor data fusion for remote sensing applications Fusion can be performed at either the signal,

Statistical multisensor classifiers

Advantages Good control over the process.Prior knowledge can be included if the model is adaptedto the application.Inclusion of ancillary data simple using a Markov Random field approach.

Limitations Assumes a particular probability density function.

Dempster-Shafer multisensor classifiers

Advantages Useful for representation of heterogeneous sources.Inclusion of ancillary data simple.Well suited to model a high number of sources.

Limitations Performance depends on selected mass functions.Not many comparisons with other approaches.

Neural net multisensor classifiers

Advantages No assumption about probability densities needed.Sensor-specific weights can easily be estimated.Suited for heterogeneous sources.

Limitations The user has little control over the fusion process and howdifferent sources are used.Involves a large number of parameters and a risk of overfitting.

Hybrid multisensor classifiers

Advantages Can combine the best of statistical and neural netor Dempster-Shafer approaches.

Limitations More complex to use.

Table 2: A summary of decision-level fusion strategies

• Improved quality in discriminating between a set of classes. Sometimes, partsof an area might be covered by clouds, and a multitemporal image set is neededto map all areas. For microwave images, the signature depends on temperatureand soil moisture content, and several images might be necessary to obtain goodcoverage of all regions in an area as two classes can have different mechanismsaffecting their signature. For this kind of application, a data fusion modelthat takes source reliability weighting into account should be considered. Anexample concerning vegetation classification in a series of SAR images is shownin Figure 5.

• Discriminate between classes based on their temporal signature development.By analyzing an area through time and studying how the spectral signaturechanges, it can be possible to discriminate between classes that are not separableon a single image. Consider e.g. vegetation mapping. Based on a single image,we might be able to discriminate between deciduous and conifer trees, but notbetween different kinds of conifer or deciduous. By studying how the spectralsignature varies during the growth season, we might also be able to discriminatebetween different vegetation species.

14

Page 15: Data fusion for remote sensing applications - … · 4.1 A general introduction to multisensor data fusion for remote sensing applications Fusion can be performed at either the signal,

Figure 5: Multitemporal image from 13 different dates during August-December 1991for agricultural sites in Norway. The ability to identify ploughing activity in a SARimage depends on the soil moisture content at the given date.

It is also relevant to consider the available data set. How many images can beexpected to be included in the analysis? Most studies use bi-temporal datasets, whichare easy to obtain. Obtaining longer time series of images can sometimes be difficultdue to sensor repeat cycles and weather limitations. In Northern Europe, cloudcoverage is a serious limitation for many applications of temporal trajectory analysis.Obtaining long time series tend to be easier for low/medium resolution images fromsatellites with frequent passes.A principal decision in multi-temporal image analysis is if the images are to be com-bined on the pixel-level or the decision-level. Pixel-level fusion consists of combiningthe multi-temporal images into a joint data set and performing the classificationbased on all data at the same time. In decision-level fusion, a classification is firstperformed for each time, and then the individual decisions are combined to reacha consensus decision. If no changes in the spectral signatures of the objects to bestudied have occurred between the image acquisitions, this is very similar to classifiercombination [6].

5.1 Multitemporal classifiers

In the following, we describe the main approaches for multitemporal classification.The methods utilize temporal correlation in different ways. Temporal feature correla-tion means that the correlation between the pixel measurements or feature vectors atdifferent times are modelled. Temporal class correlation means that the correlationbetween the class labels of a given pixel at different times is modelled.

15

Page 16: Data fusion for remote sensing applications - … · 4.1 A general introduction to multisensor data fusion for remote sensing applications Fusion can be performed at either the signal,

Direct multidate classification. In direct compound or stacked vector classifi-cation, the multitemporal data set is merged at the pixel-level into one vector ofmeasurements, followed by classification using a traditional classifier. This is a sim-ple approach that utilizes temporal feature correlation. However, the approach mightnot be suited when some of the images are of lower quality due to noise. An exampleof this classification strategy is to use multiple self-organizing maps (MSOM) [77] asa classifier for compound bi-temporal images.

Cascade classifiers. Swain [72] presented the initial work on using cascade classi-fiers. In a cascade-classifier approach the temporal class correlation between multi-temporal images is utilized in a recursive manner. To find a class label for a pixel attime t2, the conditional probability for observing class ω given the images x1 and x2

is modelled as:P (ω|x1, x2).

Classification was performed using a Maximum Likelihood classifier. In severalpapers by Bruzzone and co-authors [14, 15] the use of cascade classifiers has beenextended to unsupervised classification using multiple classifiers (combining bothMaximum Likelihood classifiers and radial basis function neural nets).

Markov chain and Markov random field classifiers. Solberg et al. [69] de-scribe a method for classification for multisource and multitemporal images wherethe temporal changes of classes are modelled using Markov chains with transitionprobabilities. This approach utilizes temporal class correlation. In the Markov ran-dom field model presented in [70], class transitions are modelled in terms of Markovchains of possible class changes and specific energy functions are used to combinetemporal information with multisource measurements, and ancillary data. Bruzzoneand Prieto [16] use a similar framework for unsupervised multitemporal classification.

Approaches based on characterizing the temporal signature. Several pa-pers have studied changes in vegetation parameters, for a review see [22]. In [2, 50]the temporal signatures of classes are modelled using Fourier series (using temporalfeature correlation). Not many approaches have integrated phenological models forthe expected development of vegetation parameters during the growth season. Au-rdal et al. [4] model the phenological evolution of mountain vegetation using HiddenMarkov Models. The different vegetation classes can be in one of a predefined setof states related to their phenological development, and classifying a pixel consistsof selecting the class that has the highest probability of producing a given series ofobservations. The performance of this model is compared to a compound maximumlikelihood approach and found to give comparable results for a single scene, but morerobust when testing and training on different images.

Other decision-level approaches to multitemporal classification. Jeon [41]developed a spatio-temporal classifier utilizing both the temporal and the spatial con-text of the image data. Khazenie and Crawford [42] proposed a method for contextualclassification using both spatial and temporal correlation of data. In this approach,

16

Page 17: Data fusion for remote sensing applications - … · 4.1 A general introduction to multisensor data fusion for remote sensing applications Fusion can be performed at either the signal,

the feature vectors are modelled as resulting from a class-dependent process plus acontaminating noise process, and the noise is correlated in both space and time. Mid-delkoop et al. [48] presented a knowledge-based classifier which used land-cover datafrom preceding years. An approach to decision level change detection using evidencetheory is given in [38].

A summary of approaches for multitemporal image classifiers is given in Table 3.

Direct multidate classifier

Advantages Simple. Temporal feature correlation between image measurements utilized.

Limitations Is restricted to pixel-level fusion.Not suited for data sets containing noisy images.

Cascade classifiers

Advantages Temporal correlation of class labels considered.Information abouth special class transitions can be modelled.

Limitations Special software needed.

Markov chain/MRF classifiers

Advantages Spatial and temporal correlation of class labels considered.Information abouth special class transitions can be modelled.

Limitations Special software needed.

Temporal signature trajectory approaches

Advantages Can discriminate between classes not separable at a single point in time.Can be used either at feature-level or decision-level.Decision-level approaches allow flexible modelling.

Limitations Feature-level approaches can be sensitive to noise.A time series of images needed (can be difficult to get morethan bi-temporal).

Table 3: A discussion of multitemporal classifiers

6 Multiscale image classification

Most of the approaches to multisensor image classification do not treat the multiscaleaspect of the input data. The most common approach is to resample all the imagesto be fused to a common pixel resolution.

In other domains of science, much work on combining data sources at differentresolutions exists, e.g. in epidemiology [12], in estimation of hydraulic conductivity forcharacterizing groundwater flow [25], in estimation of environmental components [39].These approaches are mainly for situations where the aim is to estimate an underlyingcontinuous variable.

The remote sensing literature contains many examples of multiscale/multisensordata visualization. Many multispectral sensors, like e.g. SPOT XS or Ikonos, providea combination of multispectral band and a panchromatic band of a higher resolution.Several methods for visualizing such multiscale data sets have been proposed, and

17

Page 18: Data fusion for remote sensing applications - … · 4.1 A general introduction to multisensor data fusion for remote sensing applications Fusion can be performed at either the signal,

they are often based on overlaying a multispectral image on the panchromatic imageusing different colors. We will not go into a detailed description of such techniques,but refer the reader to surveys like [51, 56, 76]. Van der Meer [26] studied the effectof multisensor image fusion in terms of information content for visual interpreta-tion, and concluded that image fusion aiming at improving the visual content andinterpretability was more successful for homogeneous data than for heteorogeneousdata.

For classification problems, Puyou-Lascassies [55] and Zhukov et al. [83] consid-ered unmixing of low-resolution data by using class label information obtained fromclassification of high-resolution data. The unmixing is performed through severalsequential steps, but no formal model for the complete data set is derived. Price [54]proposed unmixing by relating the correlation between low-resolution data and high-resolution data resampled to low-resolution, to correlation between high-resolutiondata and low-resolution data resampled to high resolution. The possibility of mixedpixels was not taken into account.

In [23], separate classifications were performed based on data from each resolution.The resulting resolution-dependent probabilities were averaged over the resolutions.

Multi-resolution tree models are sometimes used for multi-scale analysis (seee.g. [47]). Such models yield a multi-scale representation through a quad tree, inwhich each pixel at a given resolution is decomposed into four child pixels at higherresolution, which are correlated. This gives a model where the correlation betweenneighbor pixels depend on the pixel locations in an arbitrary (i.e. not problem re-lated) way.

The multi-scale model presented in [71] is based on the concept of a referenceresolution and is developed in a Bayesian framework [11]. The reference resolutioncorresponds to the highest resolution present in the data set. For each pixel ofthe input image at the reference resolution it is assumed that there is an underlyingdiscrete class. The observed pixel values are modeled conditionally on the classes. Theproperties of the class label image are described through an a priori model. Markovrandom fields have been selected for this purpose. Data at coarser resolutions aremodeled as mixed pixels, i.e., the observations are allowed to include contributionsfrom several distinct classes. In this way it is e.g. possible to exploit spectrallyricher images at lower resolution to obtain more accurate classification results at thereference level, without smoothing the results as much as if we simply oversampledthe low resolution data to the reference resolution prior to the analysis.

Methods that use a model for the relationship between the multiscale data mightoffer advantages compared to simple resampling both in terms of increased classifica-tion accuracy and by being able to describe relationships between variables measuredat different scales. This can provide tools to predict high-resolution properties fromcoarser resolution properties. Of particular concern in the establishment of statisticalrelationships is quantification of what is lost in precision at various resolutions, andthe associated uncertainty.

The potential of using multiscale classifiers will also depend on the level of detailneeded for the application, and might be related to the typical size of the structuresone wants to identify in the images. Even simple resampling of the coarsest resolutionto the finest resolution, followed by classification using a multisensor classifier can help

18

Page 19: Data fusion for remote sensing applications - … · 4.1 A general introduction to multisensor data fusion for remote sensing applications Fusion can be performed at either the signal,

improve the classification result. The gain obtained by using a classifier that explicitlymodels the data at different scales depends on the set of classes used, but also theregions used to train and test the classifier. For scenes with a high level of detail, e.g.urban scenes, the performance gain might be large. However, it depends also on howthe classifier performance is evaluated. If the regions used for testing the classifierare well inside homogeneous regions and not close to other classes, the difference inperformance in terms of overall classification accuracy might not be large, but visualinspection of the level of detail in the classified images can reveal the higher level ofdetail.

A summary of multiscale classification approaches is given in Table 4.

Resampling combined with single-scale classifier

Advantages Simple. Works well enough for homogeneous regions.

Limitations Can fail in identifying small or detailed structures.

Classifier with explicit multiscale model

Advantages Can give increased performance for small or detailed structures.

Limitations More complex software needed.Not necessary for homogeneous regions

Table 4: A discussion of multiscale classifiers

7 Concluding remarks

A number of different approaches for data fusion for remote sensing applications havebeen presented in the literature. A prerequisite for data fusion is that the data areco-registered and geometrically and radiometrically corrected.

In general, these is no consensus on which multisource or multitemporal classifi-cation approach that works best. Different studies and comparisons report differentresults. There is still a need for a better understanding on which methods are mostsuited to different applications types, and also broader comparison studies. The bestlevel and methodology for a given remote sensing application depends on several fac-tors: the complexity of the classification problem, the available data set, the numberof sensors involved, and the goal of the analysis.

Some guidelines for selecting the methodlogy and architecture for a given fusiontask are given below.

Fusion level. Decision-level fusion gives best control and allows weighting the in-fluence of each sensor. Pixel-level fusion can be suited for simple analysis, e.g. fastunsupervised change detection.

Selecting a multisensor classifier. If decision level fusion is selected, three mainapproaches for fusion should be considered: the statistical approach, neural networks,or evidence theory. A hybrid approach can also be used to combine these approaches.

19

Page 20: Data fusion for remote sensing applications - … · 4.1 A general introduction to multisensor data fusion for remote sensing applications Fusion can be performed at either the signal,

If the sources are believed to provide data of different quality, weighting schemes forconsensus combination of the sensor-specific classifiers should be considered.

Selecting a multitemporal classifier. To find the best classification strategy fora multitemporal data set, the complexity of the class separation problem must beconsideren in light of the available data set. If the classes are difficult to separate,it might be necessary to use methods for characterizing the temporal trajectory ofsignatures. For pixel-level classification of multitemporal imagery, the direct multi-date classification approach can be used. If specific knowledge about certain types ofchanges needs to be modelled, Markov chain/Markov Random Field approaches orcascade classifiers should be used.

Approaches for multiscale data. Multiscale images can either be resampled to acommon resolution, or a classifier with implicit modelling of the relationship betweenthe different scales can be used. For classification problems involving small or detailedstructures (e.g. urban areas) or heteorogeneous the latter is recommended.

Acknowledgements

The author would like to thank Line Eikvil for valuable input, in particular regardingmultisensor image registration.

References

[1] M. A. Abidi and R. C. Gonzalez. Data Fusion in Robotics and Machine Intelli-gence. Academic Press, Inc., 1992.

[2] L. Andres, W. A. Salas, and D. Skole. Fourier analysis of multi-temporal AVHRRdata applied to a land cover classification. Int. J. Remote Sensing, 15:1115–1121,1994.

[3] G. Asrar. Theory and Applications of Optical Remote Sensing. Wiley, New York,1989.

[4] Lars Aurdal, Ragnar Bang Huseby, Line Eikvil, Rune Solberg, DagrunVikhamar, and Anne Solberg. Use of hiddel markov models and phenology formultitemporal satellite image classification: applications to mountain vegetationclassification. In MULTITEMP 2005, pages 220–224, May 2005.

[5] N. Ayache and O. Faugeras. Building, registrating, and fusing noisy visual maps.Int. J. Robot. Res., 7:45–64, 1988.

[6] J. A. Benediktsson and I. Kanellopoulos. Classification of multisource and hy-perspectral data based on decision fusion. IEEE Tr. Geosc. Remote Sensing,37:1367–1377, 1999.

20

Page 21: Data fusion for remote sensing applications - … · 4.1 A general introduction to multisensor data fusion for remote sensing applications Fusion can be performed at either the signal,

[7] J. A. Benediktsson and J. Sveinsson. Multisource remote sensing data classifi-cation based on consensus and pruning. IEEE Trans. Geosc. Remote Sensing,41:932–936, 2003.

[8] J. A. Benediktsson, J. R. Sveinsson, and O. K. Ersoy. Optimized combinationof neural networks. In IEEE International Symposium on Circuits and Systems(ISCAS’96), pages 535–538, Atlanta, Georgia, May 1996.

[9] J. A. Benediktsson, J. R. Sveinsson, and P. H. Swain. Hybrid consensys theoreticclassification. IEEE Tr. Geosc. Remote Sensing, 35:833–843, 1997.

[10] J. A. Benediktsson and P. H. Swain. A method of statistical multisource classi-fication with a mechanism to weight the influence of the data sources. In IEEESymp. Geosc. Rem. Sens. (IGARSS), pages 517–520, Vancouver, Canada, July1989.

[11] J. Besag. Towards Bayesian image analysis. Journal of Applied Statistics,16(3):395–407, 1989.

[12] N. G. Besag, K. Ickstadt, and R. L. Wolpert. Spatial poisson regression for healthand exposure data measured at disparate resolutions. Journal of the AmericanStatistical Association, 452:1076–1088, 2000.

[13] E. Binaghi, P. Madella, M. G. Montesano, and A. Rampini. Fuzzy contextualclassification of multisource remote sensing images. IEEE Trans. Geosc. RemoteSensing, 35:326–340, 1997.

[14] L. Bruzzone and R. Cossu. A multiple-cascade-classifier system for a robust andpartially unsupervised updating of land-cover maps. IEEE Trans. Geosc. Rem.Sens., 40:1984–1996, 2002.

[15] L. Bruzzone and D. F. Prieto. Unsupervised retraining of a maximum-likelihoodclassifier for the analysis of multitemporal remote-sensing images. IEEE Trans.Geosc. Rem. Sens., 39:456–460, 2001.

[16] L. Bruzzone and D. F. Prieto. An adaptive semiparametric and context-basedapproach to unsupervised change detection in multitemporal remote-sensing im-ages. IEEE Trans. Image Proc., 11:452–466, 2002.

[17] G. A. Carpenter, M. N. Gjaja, S. Gopal, and C. E. Woodcock. ART neuralnetworks for remote sensing: Vegetation classification from Landsat TM andterrain data. In IEEE Symp. Geosc. Rem. Sens. (IGARSS), pages 529–531,Lincoln, Nebraska, May 1996.

[18] P. Chalermwat and T. El-Chazawi. Multi-resolution image registration usinggenetics. In Proc. ICIP, pages 452–456, 1999.

[19] H. M. Chen, M. K. Arora, and P.K. Varshney. Mutial information-based imageregistration for remote sensing data. International Journal of Remote Sensing,24:3701–3706, 2003.

21

Page 22: Data fusion for remote sensing applications - … · 4.1 A general introduction to multisensor data fusion for remote sensing applications Fusion can be performed at either the signal,

[20] P. B. Chou and C. M. Brown. Multimodal reconstruction and segmentationwith Markov Random Fields and HCF optimization. In Proceedings of the 1988DARPA Image Understanding Workshop, pages 214–221, 1988.

[21] J. Cimino, A. Brandani, D. Casey, J. Rabassa, and S. D. Wall. Multiple incidenceangle SIR-B experiment over Argentina: Mapping of forest units. IEEE Trans.Geosc. Remote Sensing, 24:498–509, 1986.

[22] P. Coppin, K. Jonkheere, B. Nackaerts, and B. Muys. Digital change detectionmethods in ecosystem monitoring: a review. International Journal of RemoteSensing, 25:1565–1596, 2004.

[23] M. M. Crawford, S. Kumar, M. R. Ricard, J. C. Gibeaut, and A. Neuenshwan-der. Fusion of airborne polarimetric and interferometric SAR for classification ofcoastal enviromnents. IEEE Trans. Geoscience and Remote Sensing, 37(3):1306–1315, 1999.

[24] X. Dai and S. Khorram. A feature-based image registration algorithm usingimproved chain-code representation combined with invariant moments. IEEETr. Geosc. Rem. Sens., 37:17–38, 1999.

[25] M. M. Daniel and A. S. Willsky. A multiresolution methodology for signal-levelfusion and data assimilation with applications to remote sensing. Proceedings ofthe IEEE, 85(1):164–180, 1997.

[26] F. Van der Meer. What does multisensor image fusion add in terms of informationcontent for visual interpretation? Int. J. Remote Sensing, 18:445–452, 1997.

[27] J. Desachy, L. Roux, and E-H. Zahzah. Numeric and symbolic data fusion: Asoft computing approach to remote sensing image analysis. Pattern RecognitionLetters, 17:1361–1378, 1996.

[28] C. Elachi, J. Cimino, and M. Settle. Overview of the shuttle imaging radar-Bpreliminary scientific results. Science, 232:1511–1516, 1986.

[29] D. Fedorov, L. M. G. Fonseca, C. Kennedy, and B. S. Manjunath. Automaticregistration and mosaicking system for remotely sensed imagery. In Proc. 9thInternational Symposium on Remote Sensing, pages 22–27, Crete, Greece, 2002.

[30] L. Fonseca, G. Hewer, C. Kenney, and B. Manjunath. Registration and fusionof multispectral images using a new control point assessment method derivedfrom optical flow ideas. In Proc. Algorithms for Multispectral and HyperspectralImagery V, pages 104–111. SPIE, Orlondo, USA, 1999.

[31] S. E. Franklin. Ancillary data input to satellite remote sensing of complex terrainphenomena. Computers and Geosciences, 15:799–808, 1989.

[32] J. Friedman. Multivariate adaptive regression splines (with discussion). Annalsof Statistics, 19:1–141, 1991.

22

Page 23: Data fusion for remote sensing applications - … · 4.1 A general introduction to multisensor data fusion for remote sensing applications Fusion can be performed at either the signal,

[33] P. Gong, R. Pu, and J. Chen. Mapping ecological land systems and classificationuncertainties from digital elevation and forest-cover data using neural networks.Photogrammetric Engineering & Remote Sensing, 62:1249–1260, 1996.

[34] L. Guan, J. A. Anderson, and J. P. Sutton. A network of networks processingmodel for image regularization. IEEE Trans. Neural Networks, 8:169–174, 1997.

[35] D. L. Hall and J. Llinas. An introduction to multisensor data fusion. Proceedingsthe IEEE, 85(1):6–23, 1997.

[36] T. Hastie, R. Tibshirani, and A. Buja. Flexible discriminant analysis by optimalscoring. J. American Statistical Association, 89:1255–1270, 1994.

[37] S. L. Hegarat-Mascle, I. Bloch, and D. Vidal-Madjar. Application of Dempster-Shafer evidence theory to unsupervised classification in multisource remote sens-ing. IEEE Trans. Geosc. Remote Sensing, 35:1018–1031, 1997.

[38] S. Le Hegarat-Mascle and R. Seltz. Automatic change detection by evidentialfusion of change indices. Remote Sensing of Environment, 91:390–404, 2004.

[39] D. Hirst, G. Storvik, and A. R. Syversveen. A hierarchical modelling approachto combining environmental data at different scales. Journal of Royal StatisticalSociety, Series C, 52:377–390, 2003.

[40] J.-N. Hwang, D. Li, M. Maechelr, D. Martin, and J. Schimert. Projection pursuitlearning networks for regression. Engineering Applications Artificial Intelligence,5:193–204, 1992.

[41] B. Jeon and D. A. Landgrebe. Classification with spatio-temporal interpixelclass dependency contexts. IEEE Trans. Geosc. Rem. Sens., 30:663–672, 1992.

[42] N. Khazenie and M. M. Crawford. Spatio-temporal autocorrelated model forcontextual classification. IEEE Trans. Geosc. Remote Sensing, 28:529–539, 1990.

[43] H. Kim and P. H. Swain. A method for classification of multisource data usinginterval-valued probabilities and its application to HIRIS data. In Proc. Work-shop on Multisource Data Integration in Remote Sensing, pages 75–82. NASAConference Publication 3099, Maryland, June 1990.

[44] Z.-D. Lan, R. Mohr, and P. Remagnino. Robust matching by partial correlation.In British Machine Vision Conference, pages 651–660, 1996.

[45] D. G. Leckie. Synergism of synthetic aperture radar and visible/infrared data forforest type discrimination. Photogrammetric Engineering and Remote Sensing,56:1237–1246, 1990.

[46] T. Lee, J. A. Richards, and P. H. Swain. Probabilistic and evidential approachesfor multisource data analysis. IEEE Trans. Geosc. Remote Sensing, 25:283–293,1987.

23

Page 24: Data fusion for remote sensing applications - … · 4.1 A general introduction to multisensor data fusion for remote sensing applications Fusion can be performed at either the signal,

[47] M. R. Luettgen, W. Clem Karl, and A. S. Willsky. Efficient multiscale regular-ization with applications to the computation of optical flow. IEEE Trans. ImageProcessing, 3(1):41–63, 1994.

[48] J. Middelkoop and L. L. F. Janssen. Implementation of temporal relationships inknowledge based classification of satellite images. Photogrammetric Engineering& Remote Sensing, 57:937–945, 1991.

[49] D. M. Mount, N. S. Netanyahu, and L. Le Moigne. Efficient algorithms for robustfeature matching. Pattern Recognition, 32:17–38, 1999.

[50] L. Olsson and L. Eklundh. Fourier series for analysis of temporal sequences ofsatellite sensor imagery. Int. Journal Remote Sensing, 15:3735–3741, 1994.

[51] G. Pajares and J. M. de la Cruz. A wavelet-based image fusion tutorial. PatternRecognition, 37:1855–1871, 2004.

[52] J. D. Paola and R. A. Schowengerdt. The effect of neural-network structure on amultispectral land-use/ land-cover classification. Photogrammetric Engineering& Remote Sensing, 63:535–544, 1997.

[53] M. Petrakos, J. A. Benediktsson, and I. Kannelopoulos. The effect of classifieragreement on the accuracy of the combined classifier in decision level fusion.IEEE Trans. Geosc. Remote Sensing, 39:2539–2546, 2001.

[54] J. C. Price. Combining multispectral data of differing spatial resolution. IEEETrans. Geoscience and Remote Sensing, 37(3):1199–1203, 1999.

[55] P. Puyou-Lascassies, A. Podaire, and M. Gay. Extracting crop radiometric re-sponses from simulated low and high spatial resolution satellite data using alinear mixing model. int. J. Remote Sensing, 15(18):3767–3784, 1994.

[56] T. Ranchin, B. Aiazzi, L. Alparone, S. Baronti, and L. Wald. Image fusion - theARSIS concept and some successful implementations. ISPRS J. Photogramm.Remote Sensing, 58:4–18, 2003.

[57] J. A. Richards, D. A. Landgrebe, and P. H. Swain. A means for utilizing ancil-lary information in multispectral classification. Remote Sensing of Environment,12:463–477, 1982.

[58] J. M. Richardson and K. A. Marsh. Fusion of multisensor data. Int. J. Robot.Res, 7:78–96, 1988.

[59] E. Rignot, R. Kwok, J.C. Curlander, J. Homer, and I. Longstaff. Automatedmultisensor registration: requirements and teqniques. Photogrammetric Engi-neering and Remote Sensing, 57:1029–1038, 1991.

[60] B. D. Ripley. Flexible non-linear approaches to classification. In V. Cherkassky,J. H. Friedman, and H. Wechsler, editors, From Statistics to Neural Networks.Theory and Pattern Recognition Applications, pages 105–126. NATO ASI SeriesF: Computer and Systems Sciences, Springer Verlag, 1994.

24

Page 25: Data fusion for remote sensing applications - … · 4.1 A general introduction to multisensor data fusion for remote sensing applications Fusion can be performed at either the signal,

[61] S. B. Serpico and F. Roli. Classification of multisensor remote-sensing imagesby structured neural networks. IEEE Tr. Geosc. Remote Sensing, 33:562–578,1995.

[62] G. Shafer. A Mathematical Theory of Evidence. Princeton University Press,1976.

[63] A. K. Skidmore, B. J. Turner, W. Brinkhof, and E. Knowles. Performance ofa neural network: Mapping forests using GIS and remotely sensed data. Pho-togrammetric Engineering & Remote Sensing, 63:501–514, 1997.

[64] P. C. Smits and S. G. Dellepiane. Synthetic aperture radar image segmentationby a detail preserving Markov random field approach. IEEE Trans. Geosc. Rem.Sens., 35:844–857, 1997.

[65] A. Solberg, G. Storvik, and R. Fjørtoft. A comparison of criteria for decisionfusion and parameter estimation in statistical multisensor image classification.In IEEE Symp. Geosc. Rem. Sens. (IGARSS’02), page None, July 2002.

[66] A. H. Solberg. Flexible nonlinear contextual classification. Pattern RecognitionLetters, 25:1501–1508, 2004.

[67] A. H. Schistad Solberg. Contextual data fusion applied to forest map revision.IEEE Trans. Geosc. Rem. Sens., 37:1234–1243, 1999.

[68] A. H. Schistad Solberg, A. K. Jain, and T. Taxt. Multisource classification ofremotely sensed data: Fusion of Landsat TM and SAR images. IEEE Trans.Geosc. Rem. Sens., 32:768–778, 1994.

[69] A. H. Schistad Solberg, T. Taxt, and Anil K. Jain. A Markov random fieldmodel for classification of multisource satellite imagery. IEEE Trans. Geosc.Rem. Sens., 34:100–113, 1996.

[70] A. Schistad Solberg. Texture fusion and classification based on flexible discrimi-nant analysis. In International Conference on Pattern Recognition (ICPR), pages596–600, Vienna, Austria, August 1996.

[71] G. Storvik, R. Fjørtoft, and A. Solberg. A bayesian approach to classificationin multiscale remote sensing data. IEEE Trans. Geosc. Rem. Sens., 43:539–547,2005.

[72] P. H. Swain. Bayesian classification in a time-varying environment. IEEE Trans.Sys. Man and Cyber., 8:879–883, 1978.

[73] B. C. K. Tso and P. M. Mather. Classification of multisource remote sensingimagery using a genetic algorithm and Markov random fields. IEEE Tr. onGeosc. Remote Sensing, 37:1255–1260, 1999.

[74] F. T. Ulaby and C. Elachi. Radar Polarimetry for Geoscience Applications.Artec House Inc., 1990.

25

Page 26: Data fusion for remote sensing applications - … · 4.1 A general introduction to multisensor data fusion for remote sensing applications Fusion can be performed at either the signal,

[75] F. T. Ulaby, R. K. Moore, and A. K. Fung. Microwave Remote Sensing, Activeand Passive, Vols. I-III. Artech House Inc., 1981, 1982, 1986.

[76] L. Wald. Data fusion: Definitions and achitectures - Fusion of images of differentspatial resolutions. Ecole des Mines Press, 2002.

[77] W. Wan and D. Fraser. Multisource data fusion with multiple self-organizingmaps. IEEE Tr. on Geosc. Remote Sensing, 37:1344–1349, 1999.

[78] W. Wan and D. Fraser. A self-organizing map model for spatial and temporalcontextual classification. In IEEE Symp. Geosc. Rem. Sens. (IGARSS), pages1867–1869, Pasadena, California, August 1994.

[79] W.A. Wright. A Markov random field approach to data fusion and colour seg-mentation. Image Vision Comp., 7:144–150, 1989.

[80] S. Wu. Analysis of data acquired by shuttle imaging radar SIR-A and LandsatThematic Mapper over Baldwin county, Alabama. In Proc. Machine Processingof Remotely Sensed Data Symposium, pages 173–182, West Lafayette, Indiana,June 1985.

[81] N. Xiong and P. Svensson. Multi-sensor management for information fusion:issues and approaches. Information Fusion, 3:163–180, 2002.

[82] H. A. Zebker and J. J. Van Zyl. Imaging radar polarimetry: A review. Proceed-ings of the IEEE, 79:1583–1606, 1991.

[83] B. Zhukov, D. Oertel, F. Lanzl, and G. Reinhackel. Unmixing-based multisensormultiresolution image fusion. IEEE Trans. Geoscience and Remote Sensing,37(3):1212–1226, 1999.

[84] B. Zitova and J. Flusser. Image registration methods: a survey. Image andVision Computing, 21:977–1000, 2003.

26