• Keine Ergebnisse gefunden

Pattern Recognition

N/A
N/A
Protected

Academic year: 2022

Aktie "Pattern Recognition"

Copied!
13
0
0

Wird geladen.... (Jetzt Volltext ansehen)

Volltext

(1)

Pattern Recognition

Clustering, Self-Organizing Maps

(2)

Clustering

The task: partition a set of objects into “meaningful” subsets (clusters). The objects in a subset should be “similar”.

Notations:

Set of Clusters Set of indices Feature vectors

Partitioning

for ,

(3)

Clustering

Let and each cluster has a “representative”

The task reads:

Alternative variant is to consider the clustering as

a mapping that assigns a cluster number to each

(4)

K-Means Algorithm

Initialize centers randomly, Repeat until convergence:

1. Classify:

2. Update centers:

• The task is NP

• converges to a local optimum (depends on the initialization)

(5)

Sequential K-Means

Repeat infinitely:

1. Chose randomly a feature vector from the training data 2. Classify it:

3. Update the -th center:

with a decreasing step

• converges to the same, as the parallel version

• is a special case of Robbins-Monro Algorithm

(6)

Some variants

Other distances, e.g. instead of

In the K-Means algorithm the classification step remains the same, the update step – the geometric median of

(a bit complicated as the average ).

Another problem: features may be not additive ( does not exist) Solution: K-Medioid Algorithm ( is a feature vector from the

training set)

(7)

A generalization

Observe (for the Euclidean distance):

In what follows:

with a Distance Matrix that can be defined in very different ways.

Example: Objects are nodes of a weighted graph, is the length of the shortest path from to .

Distances for “other” objects (non-vectors):

• Edit (Levenshtein) distance between two symbolic sequences

• For graphs – distances based on graph isomorphism etc.

(8)

An application – color reduction

Objects are pixels, features are RGB-values. Decompose an image into parts that correspond to “characteristic” colors.

(8 colors)

(9)

The task is to “approximate” a dataset by a neural network of a certain topology.

An example – stereo in “flatland”.

The input space is 3- (or more) dimensional, the set of points is however isomorphic to a 2D-space (up to noises).

Cohonen Networks, Self-Organizing Maps

(10)

Self-Organizing Maps

SOM-s (usually) consist of RBF-neurons , each one represents (covers) a “part” of the input space (specified by the centers ).

The network topology is given by means of a distance . Example – neurons are nodes of a weighted graph, distances are shortest paths. For the “flatland” example the graph is a 2D-grid with unit weight for all edges.

(11)

Self-Organizing Maps, sequential algorithm

1. Chose randomly a feature vector from the training data (white) 2. Compute the “winner”-neuron (dark-yellow)

3. Compute the neighborhood of in the network (yellow)

4. Update the weights of all neurons from

(12)

is monotonously decreasing with respect to (time) and Without 3) – the sequential K-Means.

Parallel variants:

Go through all feature vectors, sum up the gradients, apply.

Example for :

The network fits into the data distribution (unfolds).

Self-Organizing Maps, algorithms

(13)

Conclusion

Before:

1. Neuron – linear classifier

2. Feed-Forward Networks – complex classifiers 3. Hopfield Networks – structured output

4. Cohonen Networks – model fitting

Basic tasks, modeling, network architectures, algorithms etc.

The next block (3-4 classes)

– probability theory, inference and learning …

Referenzen

ÄHNLICHE DOKUMENTE

The MON-Tree showed good scalability when increasing the number of objects and time units in the index structure, and the query window and time interval in querying.. In our tests,

Whenever a particular level of analysis of Nature is populated with objects whose internal structure engenders specic action capable of changing or cre- ating other objects,

Example: Objects are nodes of a weighted graph, is the length of the shortest path from to4. Distances for “other”

Racionalismo y empirismo claman sin embargo por la objetividad, es decir, porque el sujeto se mantenga impávido ante los objetos, pues los objetos están allí para ser ingeridos por

The unknown shape is described with the help of the level set function; the advantage of the shape function is that no mesh movement or re-meshing is necessary but an

Hence, if the projection is chosen at random, the shadow vertex method is a randomized simplex algorithm with polynomial expected running time for any flow linear program.. In

The resulting shape description allows to separately analyze the deformation behavior of shapes and their structural relations, while the different levels of the shape hierarchy

Nonetheless, an increasingly networked and standardized German and European repository infrastructure is being built in projects and initiatives such as, for example, DRIVER