37
Independent Component Analaysis 1

Independent Component Analysis

Embed Size (px)

Citation preview

Page 1: Independent Component Analysis

1

Independent Component Analaysis

Page 2: Independent Component Analysis

2

Significant recent advances in the field of statistical signal processing should be brought to the attention of the biomedical engineering community.

Algorithms have been proposed to separate multiple signal sources based solely on their statistical independence, instead of the usual spectral differences.

These algorithms promise to:

lead to more accurate source modeling,

more effective artifact rejection algorithms.

Page 3: Independent Component Analysis

3

Motivation

Problem: Decomposing a mixed signal into independent sources

Ex.

Given: Mixed Signal

Our Objective is to gain:

Source1 News

Source2 Song ICA (Independent Component Analysis) is a quite powerful technique to

separate independent sources

Page 4: Independent Component Analysis

4BSS and ICA

Cocktail party or Blind Source Separation (BSS) problem Ill posed problem, unless assumptions are made!

Most common assumption is that source signals are statistically independent. This means knowing value of one of them gives no information about the other.

Methods based on this assumption are called Independent Component Analysis methods statistical techniques for decomposing a complex data set into

independent parts.

It can be shown that under some reasonable conditions, if the ICA assumption holds, then the source signals can be recovered up to permutation and scaling.

Page 5: Independent Component Analysis

5

Statistical independence

In probability theory, to say that two events are independent means that the occurrence of one event makes it neither more nor less probable that the other occurs.

Examples of such independent random variables are the value of a dice thrown and of a coin tossed, or speech signal and background noise originating from

a ventilation system at a certain time instant.

This means if we have two signals A and B one signal doesn’t give any information with regards of other signal this is the definition of statistical independent with aspect to ICA

Definitions

)()()()()(

)()|( ,

, yPXPyxPyPxP

yxPxyP yxyxy

x

yxy

Page 6: Independent Component Analysis

6Example

cocktail-party problem

The microphones give us two recorded time signals. We denote them with x=(x1(t), x2(t)). x1 and x2 are the amplitudes and t is the time index. We denote the independent signals by s=(s1(t),s2(t)); A - mixing matrix (2x2)

x1(t) = a11s1 +a12s2

x2(t) = a21s1 +a22s2a11,a12,a21, and a22 are some parameters that depend on the distances of the microphones from the speakers. It would be very good if we could estimate the two original speech signals s1(t) and s2(t), using only the recorded signals x1(t) and x2(t). We need to estimate the aij., but it is enough to assume that s1(t) and s2(t), at each time instant t, are statistically independent. The main task is to transform the data (x); s=Ax to independent components, measured by function: F(s1,s2)

Page 7: Independent Component Analysis

7

Recordings in real environments

Separation of Music & SpeechExperiment-Setup:- office room (5m x 4m)- two distant talking mics- 16kHz sampling rate

40cm

60cm

Page 8: Independent Component Analysis

8The ICA model

s1 s2

s3 s4

x1 x2 x3 x4

a11

a12a13

a14

xi(t) = ai1*s1(t) + ai2*s2(t) + ai3*s3(t) + ai4*s4(t)

Here, i=1:4.

In vector-matrix notation, and dropping index t, this is

x = A * s

Page 9: Independent Component Analysis

9

Restrictions on ICA

To make sure that the basic ICA model just given can be estimated, we have to make certain assumptions and restrictions.

1. The independent components are assumed statistically independent

2. The independent components must have nongaussian distributions

3. For simplicity, we assume that the unknown mixing matrix is square.

In other words, the number of independent components is equal to the number of observed mixtures.

Page 10: Independent Component Analysis

10

Central Limit Theorem

Central limit theorem states that linear combination of given random variables is more Gaussian as compared to the original variables themselves.

Example:

Repeated throwing of nonbiased dices, with the result given by the sum of the dice faces.

• For one dice the distribution is uniform;

• For two dices the sum is piecewise linearly distributed;

• For a set of k dices, the distribution is given by a

piecewise k-order polynomial distribution;

• With the increase of k, the distribution of the sum tends towards a more Gaussian one.

See Matlab script !!

Page 11: Independent Component Analysis

11

One Dice

Page 12: Independent Component Analysis

12

2 dices

Page 13: Independent Component Analysis

13

5 dices

Page 14: Independent Component Analysis

14

9 Dices

Page 15: Independent Component Analysis

15

10 Dices

Page 16: Independent Component Analysis

16Why does ICA require signals to be non-Gaussian ?

Central limit theorem states that linear combination of given random variables is more Gaussian as compared to the original variables themselves.

So if we have two independent Gaussian sources after mixing them they become more Gaussian as central limit theorem state, which make it impossible to recover the original sources

Page 17: Independent Component Analysis

17

Mathematical framework

Very general-purpose method of signal processing and data analysis. Definition of ICA

Statistical latent variables model n linear mixtures of n independent components

N linear mixtures of N independent components

, and , are considered random variables, not proper time signals. The values of the signals are considered samples (instantiations) of the random variables, not functions of time. The mean value is taken zero, without loss of generality

Page 18: Independent Component Analysis

18

Page 19: Independent Component Analysis

19

The linear mixing equation (the ICA model):

Denoting by aj – the j th column of matrix A; the model becomes

The ICA model is a generative model, i.e., it describes how the observed data are generated by mixing the components si .

The independent components are latent variables, i.e., not directly observable.

The mixing matrix A is also unknown.

We observe only the random vector x, and we must estimate both A and s. This must be done under as general assumptions as possible.

Page 20: Independent Component Analysis

20

ICA is a special case of blind source separation (BSS)

Blind means that we know very little, if anything, on the mixing matrix, and make little assumptions on the source signals.

Basic ICA assumption: The source components are

statistically independent they have (unknown) distributions as

Non-gaussian as possible optimize a certain contrast function

The problem is finding W - the unmixing matrix that gives

- the best estimate of the independent source vector:

• If the unknown mixing matrix A is square and nonsingular, then

• Else, the best unmixing matrix, that separates sources as independent as possible, is given by the generalized inverse Penrose-Moore matrix:

Page 21: Independent Component Analysis

21

Ambiguities of ICA

Permutation Ambiguity:

Assume that P is a nxn permutation matrix.

Given only the s , we cannot distinguish between W and PW

The permutation of the original sources is ambiguous

Page 22: Independent Component Analysis

22

Ambiguities of ICA

Scaling Ambiguity:

We cannot recover the “correct” scaling of the sources

Solution: Data whitening (sphering)

Scaling a speaker's speech signal by some positive factor affects only the volume of that speaker's speech.

Also, sign changes do not matter: and sound identical when played on a speaker.

Page 23: Independent Component Analysis

23

Whitening (sphering)

Suppose is a random column vector with covariance matrix M and mean zero.

Whitening X means simply multiplying it by

Steps: Find the covariance matrix of X, Cov(X) =

Find the inverse of the square root of Cov(X) and call it invcov

Find the mean of X then subtract it from X

Finally multiply invcov by X to get the whitened version of X

Whitening solves half of the ICA problem !!

See matlab script !!

Page 24: Independent Component Analysis

24

ICA Algorithm

To find the unmixing matrix W it is a maximization problem

We want to find the unmixing matrix W such that the non-Guassinaty is maximized

W is the parameter that we want to estimate

Given a training set the log likelihood is

Page 25: Independent Component Analysis

25

ICA Algorithm

Maximize using gradient ascent:

By taking the derivatives of using:

Page 26: Independent Component Analysis

26

ICA Algorithm

Page 27: Independent Component Analysis

27

Usages of ICA

Separation of Artifacts in MEG (magneto-encephalography) data

Reducing Noise in Natural Images

Telecommunications (CDMA [Code-Division Multiple Access] mobile communications)

Page 28: Independent Component Analysis

28Application domains of ICA

Blind source separation Image denoising Medical signal processing – fMRI, ECG, EEG Modelling of the hippocampus and visual

cortex Feature extraction, face recognition Compression, redundancy reduction Watermarking Clustering Time series analysis (stock market, microarray

data)

Page 29: Independent Component Analysis

29

Image denoising

Wiener filtering

ICA filtering

Noisy image

Original image

Page 30: Independent Component Analysis

30

Automatic Image Segmentation

Page 31: Independent Component Analysis

31

Barcode Classification Results

Classifying 4 data sets: linear, postal, matrix, junk

Page 32: Independent Component Analysis

32

Image De-noising

Page 33: Independent Component Analysis

33

Filling in missing data

Page 34: Independent Component Analysis

34

ICA applied to Brainwaves

An EEG recording consists of activity arising from many brain and extra-brain processes

Page 35: Independent Component Analysis

35

ICA applied to Brainwaves

Page 37: Independent Component Analysis

37

Thank You