• Keine Ergebnisse gefunden

Independent Component Analysis ICA The Cocktail-Party Problem

N/A
N/A
Protected

Academic year: 2022

Aktie "Independent Component Analysis ICA The Cocktail-Party Problem"

Copied!
11
0
0

Wird geladen.... (Jetzt Volltext ansehen)

Volltext

(1)

The Cocktail-Party Problem

Faith Riggold 1964 “Cocktail-Party”

ICA

2

Independent Component Analysis

(2)

ICA: the task

Given

n

Signals

X

that are a linear mixture of unknown source signals

S

, can we estimate the source signals? The “BSS: Blind Source Separation” or “Cocktail-Party”

problem:

Estimated Separated Sources

W Y

X = A·S Y=W·X

unknown to be estimated

Unknown Source Signals

S A

Observed Signals

“linear mixture of sources”

X

ICA: the problem

4

Given: X = { xi( t ) | 1 ≤ i ≤ n, 1 ≤ t ≤ T }, a n x T data matrix Problem: How to decompose the matrix into

X = AS

with

an unknown mixing matrix

A

and unknown source signal Si

X = A

x

···· S1●····

···· S2●····

···· Sm●····

So far only linearity is assumed. -> Many solutions & ambiguities !!!

Ambiguities:

1. The variance (scaling ) of Si cannot be determined, either a scalar multiplication of Aor of S

Ambiguities:

1. The variance (scaling ) of Si cannot be determined, either a scalar multiplication of Aor of S

2. The order of the sources is arbitrary.

-> we should normalize the sources.

(3)

ICA: the data model

X = A

x

···· S1●····

···· S2●····

···· Sm●····

Question:

What could be an assumption on the sources

S

that helps to decompose the

X

into

A

and

S

? Assumption:

1) All sources signals Si , the rows of

S,

are statistically independent.

2) Since we can not estimate the magnitude of Si , we fix it to E[SiSiT] =1 ==> E[

SS

T] =

I

6

Definition:

Random variables (vectors)

y

iare statistically independent if

P

(y1, y2,….,yn) =

P

(yi)

ICA: statistical independence

==> for any function

g

i

E

[

g

1(y1)

g

2(y2)···

g

n(yn)] =

∏ E

[

g

i(yi) ]

(4)

ICA: what about PCA

=> E[

s

] = 0 , E[

s ²

] = 1

uncorrelated versus independence

=>

s,t

are uncorrelated

=> E [ f( s ) g( t )] = 2E [ s4] = 2

 

3 4

5 and E [ f( s )] E[ g( t )] = 2

E[] 2E[ ]

=>

s,t

are stat. dependent -√3

½

√3

+√3

p(s) : uniform distribution

define t = 2s

²

=> E[ s·t ] = 2E[ ] =0

define f(x) = x

²

and g(x) = x

also E[ s ]E[ t ] =0 ?? =>

s ,t

statistically independent??

8

If more than one source signal is Gaussian we can not separate the sources with ICA.

Let us assume:

all source signals are Gaussian, uncorrelated and of unit variance.

Then an orthogonal mixing matrix would generate signals Xi with a

completely symmetrical Gaussian joint density function.

ICA: gaussian signals are of no use

A completely symmetrical joint density function contains no information on the structure of the mixing matrix

A

.

(5)

ICA: the data model continued

Assumptions:

1) non-Gaussian source signals Si (except possibly one).

2) All sources signals Si , the rows of

S

are statistically independent.

3) Since we can not estimate the magnitude of Si , we fix it to E[SiSiT] =1 ==> E[

SS

T] =

I

ICA: approach

10

We search for a matrix W (ideally W = A-1) so that the rows yi of

Y = WX

1. are maximal statistically independent, 2. are maximal non Gaussian,

3. and of variance E[yiyiT] =1.

(6)

ICA: PCA as preprocessing

PCA

• computes the axis of maximum variance

• these axis are uncorrelated ( but only for Gaussian data statistically independent).

Now we can normalize the axis by their variance => Z = V·X=V·A·S with E[ziziT] =1.

PCA normalize by variance

“whitening the data”

for non Gaussian data 

ICA: the procedure

12

Source Mixture

Whitened Signals

Z = V·X=V·A·S

X = A·S Y=W·X

S

X X X

(7)

ICA: the problem reformulated twice

2.) Independence approach :

1. Measure the independence between the signals.

2. Find the signals that maximize this independence.

1.)

Non-Gaussian approach :

By the central limit theorem, the PDF of a sum of n independent random variables tends to a Gaussian random variable.

1. Find a measure ofnon-Gaussianity.

2. Find W such that the outputs PDF are as different as possible from the Gaussian function.

ICA: measure of non-Gaussian

14

There exist several approaches to measure if a pdf is Gaussian or not!

However, we do not know the full pdf!

 it is more reasonable to use more global measures of the distribution such as mean, variance,…

Remember, moments and cumulants (semi-invariants) are easy to compute!

1

[ ] ( ), 1, 2, ...

N

i i

i

I

m E x x P x i

ith moment:

(8)

ICA: measure of non-Gaussian

1 1

2 2

2 2 1

4

( ) [ ]

( ) [ ] -

...

( ) [ ] [ ] [ ]

[ ] [ ]

i i

i j i j

i j k l i j k l i j k l

i k j l

x E x m

x x E x x m m

x x x x E x x x x E x x E x x E x x E x x

[ ] [ ]

[ ] [ ]

i j k l

i l j k

E x x E x x E x x E x x

The cumulants of distribution are:

The Kurtosis is then defined:

4 2 2

4 2 2

( i) [ i ] - 3 ( [ i ])

ku rt xE x E x

  

ICA: measure of non-Gaussian

16

Both cumulants and kurtosis are good to measure the deviation of a distribution from being Gaussian:

For ICA the Kurtosis is commonly applied:

For finding the independent components:

optimize

W

so that is maximum: ( )

n

j j

ku rt Y

(9)

ICA: 2nd approach: independence

Remember:

Kullbach-Leibler divergence measures a distance between two pdf’s (! not symmetric):

     

, ln b 

a b a

a

p x

L p p p x d x

p x

 

Definition of stat. independent

p

(y1, y2,….,yn) =

p

(yi)

Now we measure L ( p(y1, y2,….,yn) , p(yi) )

   

( )

( ), ( ) ln

n

i i

i

p y

L p Y p y p Y d Y

p Y

 

  

ICA: measure of independence

18

Now we measure L

(

p(y1, y2,….,yn) , p(yi) )

   

( )

( ), ( ) ln

n

i i

i

p y

L p Y p y p Y d Y

p Y

 

  

     

ln ln ( )

n

i i

p Y p Y d Y p Y p y d Y



   

ln ( )

n

i i

H Y p Y p y d Y

 

   

n i i

H Y H y

(10)

ICA: application to images

Mixtures

ICA: application to images

20

PCA

ICA

(11)

ICA: application to images

ICA

Originals

Referenzen

ÄHNLICHE DOKUMENTE

We give an example of a pure group that does not have the independence property, whose Fitting subgroup is neither nilpotent nor definable and whose soluble radical is neither

Another area of scholarship, much of it new in the last thirty years, has been the exploration of the impact of the western Enlightenment among educated Greeks from

We believe that the research community needs to bring the powerful concepts of declarative specification to current data analysis systems, in order to achieve the broad big

In der ICA-Auswertung zeigte sich, dass es für die Gedächtnisfunktion spezifische Netzwerke im Cortex gibt, die durch eine abgestimmte Stimulierung aktiviert oder

According to Mudde’s definition of populism that was explained earlier in the study, populism is an ideology that opposes two groups – ‘the pure people’ vs the ‘corrupt

Cell based biochip data as well as spectroscopic data are used to demonstrate that, if the interaction contains information, the prediction performance can be improved by including

These responses are strongly reduced for stimuli played in the middle of the sequence as shown by depicting an average non-target response observed at the 45th position of

Figure 7.16: Network Based on the 3,000 Most Frequent Unigrams in the Main Corpus (No Transformation, Burrows Delta, 6 Nearest Neighbours, Genres)... By now, it has become clear