• Keine Ergebnisse gefunden

Machine Learning

N/A
N/A
Protected

Academic year: 2022

Aktie "Machine Learning"

Copied!
15
0
0

Wird geladen.... (Jetzt Volltext ansehen)

Volltext

(1)

Machine Learning

Clustering, Self-Organizing Maps

(2)

Clustering

The task: partition a set of objects into “meaningful” subsets (clusters). The objects in a subset should be “similar”.

Notations:

Set of Clusters Set of indices Feature vectors

Partitioning

for ,

(3)

Clustering

Let and each cluster has a “representative”

The task reads:

Alternative variant is to consider the clustering as

a mapping that assigns a cluster number to each

(4)

K-Means Algorithm

Initialize centers randomly, Repeat until convergence:

1. Classify:

2. Update centers:

• The task is NP

• converges to a local optimum (depends on the initialization)

(5)

Sequential K-Means

Repeat infinitely:

1. Chose randomly a feature vector from the training data 2. Classify it:

3. Update the -th center:

with a decreasing step

• converges to the same, as the parallel version

• is a special case of Robbins-Monro Algorithm

(6)

Some variants

Other distances, e.g. instead of

In the K-Means algorithm the classification step remains the same, the update step – the geometric median of

(a bit complicated as the average ).

Another problem: features may be not additive ( does not exist) Solution: K-Medioid Algorithm ( is a feature vector from the training set)

(7)

A generalization

Observe (for the Euclidean distance):

In what follows:

with a Distance Matrix that can be defined in very different ways.

Example: Objects are nodes of a weighted graph, is the length of the shortest path from to .

Distances for “other” objects (non-vectors):

• Edit (Levenshtein) distance between two symbolic sequences

• For graphs – distances based on graph isomorphism etc.

(8)

An application – color reduction

Objects are pixels, features are RGB-values.

Partition the RGB-space into “characteristic” colors.

(8 colors)

(9)

Another application – superpixel segmentation

Object are pixels. Features are RGBXY-values.

→ Those pixels belong to the same cluster that are close to each other both spatially and in in the RGB-space

SLIC Superpixels: http://ivrg.epfl.ch/research/superpixels

(10)

Cohonen Networks, Self-Organizing Maps

The task is to “approximate” a dataset by a neural network of a certain topology.

An example – stereo in “flatland”.

The input space is 3- (or more) dimensional, the set of points is

(11)

Self-Organizing Maps

SOM-s (usually) consist of RBF-neurons , each one represents (covers) a “part” of the input space (specified by the centers ).

The network topology is given by means of a distance . Example – neurons are nodes of a weighted graph, distances are shortest paths. For the “flatland” example the graph is a 2D-grid with unit weight for all edges.

(12)

Self-Organizing Maps, sequential algorithm

1. Chose randomly a feature vector from the training data (white) 2. Compute the “winner”-neuron (dark-yellow)

3. Compute the neighborhood of in the network (yellow)

4. Update the weights of all neurons from

(13)

Self-Organizing Maps, algorithms

is monotonously decreasing with respect to (time) and Without 3) – the sequential K-Means.

Parallel variants:

Go through all feature vectors, sum up the gradients, apply.

Example for :

The network fits into the data distribution (unfolds).

(14)

K-Means ↔ Expectation Maximization

EM – compute posteriors for the Expectation step

K-Means – classify object

(15)

Conclusion

Before:

1. Probability theory – models, inference, learning 2. → Discriminative learning → Classifiers

Neural networks:

1. Feed-Forward Networks – complex classifiers 2. Hopfield Networks – structured output

3. Cohonen Networks – clustering (unsupervised), model fitting Next topics – further classifiers:

1. Support Vector Machines, Kernels 2. Empirical Risk minimization

3. Principal Component Analysis

4. Combining classifiers – Decision Trees, AdaBoost

Referenzen

ÄHNLICHE DOKUMENTE

Racionalismo y empirismo claman sin embargo por la objetividad, es decir, porque el sujeto se mantenga impávido ante los objetos, pues los objetos están allí para ser ingeridos por

The unknown shape is described with the help of the level set function; the advantage of the shape function is that no mesh movement or re-meshing is necessary but an

Example: Objects are nodes of a weighted graph, is the length of the shortest path from to.. Distances for “other”

 To test the effects of the digital learning objects created in this study on an Estonian basic school EFL class with the hypothesis being that the learners who

Whenever a particular level of analysis of Nature is populated with objects whose internal structure engenders specic action capable of changing or cre- ating other objects,

Nonetheless, an increasingly networked and standardized German and European repository infrastructure is being built in projects and initiatives such as, for example, DRIVER

Af- fects “begin in the interstitial space between agents” (Kraftl, 2013, p. 50, original emphasis). In this article, I take spaces in-between to be liminal ones and advance

If this is the case one says that the field K is real (or formally real), otherwise nonreal. We assume that the reader is familiar with the basic theory of quadratic forms