• Keine Ergebnisse gefunden

Pattern Recognition

N/A
N/A
Protected

Academic year: 2022

Aktie "Pattern Recognition"

Copied!
16
0
0

Wird geladen.... (Jetzt Volltext ansehen)

Volltext

(1)

Pattern Recognition

2

(2)

Perceptrons

by M.L. Minsky and S.A. Papert (1969)

Books:

4

Pattern Recognition, fourth Edition (Hardcover) by Sergios Theodoridis, Konstantinos Koutroumbas Publisher: Academic Press; 4th edition ( 2006, 2008) Language: English

ISBN-10: 1597492728

4th Edition 3 rd Edition 2nd Edition

(3)

Books:

Pattern Recognition and Machine Learning by Christopher Bishop

Publisher: Springer; 1 edition (August 17, 2006) ISBN: 0387310738

Pattern Classification, second Edition (Hardcover) by Richard O. Duda, Peter E. Hart and David G.

Stork

Publisher: Wiley Interscience 2 edition (2001) Language: English

ISBN: 0-471-05669-3

Introduction to Pattern Recognition

6

Today:

• Machine Perception

• An Example

• Pattern Recognition Systems

• The Design Cycle

• Learning

• Conclusion

(4)

Pattern Recognition

Build a machine that can recognize patterns.

Machine Perception :

– Optical Character Recognition (OCR), – Speech recognition,

– Email Spam Detection,

– Skin Detection based on pixel color, – Texture classification,

– …..

Pattern Recognition

8

Base technology for:

– Image analysis,

– Speech understanding, – Document analysis, – Bioinformatics,

– Time series prediction.

(5)

An Example: Sea bass / Salmon

“Sorting incoming fish on a conveyor according to species using optical sensing.”

Sea bass

Species

Salmon

10

• Length

• Lightness

• Width

• Number and shape of fins

• Position of the mouth, etc…

This is the set of all suggested features to explore for further use in our classification task!

Sea bass / Salmon

Problem Analysis

Set up a camera and take some sample images to extract features:

(6)

1. Preprocessing

Use a segmentation operation to isolate fish from one another and from the background.

2. Feature extraction

Information from a single fish is sent to a feature extractor whose purpose is to reduce the data by measuring certain features.

(Mega Pixel -> few numbers)

3. The features are passed to a classifier.

Sea bass / Salmon

Sea bass / Salmon

12

(7)

Sea bass / Salmon

Example of feature: length of the fish

Training error: 90 / 316 = 28%

Decision: If length < l* then salmon else sea bass l*

14

Training error: 90 / 316 = 28%

The length is a poor feature alone!

Select the lightness as a possible feature.

Sea bass / Salmon

(8)

Sea bass / Salmon

Example of feature: lightness of the fish

Training error: 16 / 316 = 5%

Decision: If lightn. < x*, then salmon else sea bass

16

• Threshold decision boundary and cost relationship.

– Move our decision boundary toward smaller values of lightness in order to minimize the cost (reduce the number of sea bass that are classified as salmon!).

Sea bass / Salmon

(9)

Now we use 2 features instead of 1:

Adopt the lightness and add the width of the fish.

Fish x

T = [x1, x2] Lightness Width

Sea bass / Salmon

Sea bass / Salmon

18

Training error: 8 / 316 = 2,5%

Linear decision function:

(10)

• We might add other features that are not correlated with the ones we already have. A precaution should be taken not to reduce the performance by adding

“noisy features”.

Sea bass / Salmon

• Ideally, the best decision boundary should be the one which provides an optimal performance such as in the following figure:

Sea bass / Salmon

20

(11)

However, our satisfaction is premature because the central aim of designing a

classifier is to correctly classify novel input.

Issue of generalization!

Sea bass / Salmon

Sea bass / Salmon

22

Training error: 9 / 316 = 2,5%

Quadratic decision function:

(12)

Pattern Recognition Systems:

• Sensing

– Use of a transducer (camera or microphone).

– PR system depends on the bandwidth, the resolution, sensitivity distortion of the transducer.

• Segmentation and grouping

– Patterns should be well separated and should not overlap.

Pattern Recognition Systems

24

input sensing segmentation feature extraction

Segmentation:

– Isolate relevant data from the sensor output stream

Feature extraction:

– Discriminative

– Invariant to translation, rotation and scale….

Classification: Use a feature

(13)

The Design Cycle:

start collect data choose features

choose model train classifier evaluate classifier

end

error

<T

>T

26

• Data Collection:

• Feature Choice: Depends on the

characteristics of the problem domain.

The Design Cycle

– What type of sensor?

– How do we know when we have collected an adequately large and representative set of examples for training and testing the system?

– simple to extract,

– invariant to irrelevant transformation, – insensitive to noise and

(14)

• Model Choice:

• Training:

The Design Cycle

– Depends on the model chosen.

– Use data to determine the parameters of a classifier.

– There are many different procedures for training classifiers and choosing models.

– e.g. should we use a linear or a quadratic decision function?

– Can we estimate the probability distribution function that models the features?

28

• Evaluation:

– Measure the error rate on the validation set of examples that is different from the training set.

– This tests the generalization performance.

– If not good enough, go back to either of the design step.

The Design Cycle

(15)

Computational Complexity:

– More complex classifier are more computationally expensive.

– What is the optimal trade-off between computational ease and performance?

– (How does an algorithm scale as a function of the number of features, patterns or

categories?)

The Design Cycle

Learning

30

• Supervised learning

• Unsupervised learning

– The system forms clusters or “natural groupings”

of the input patterns.

– Difficult: still the focus of intense research.

– A teacher provides a category label or cost for each pattern in the training set.

(16)

Conclusion

• The number, complexity and magnitude of the sub- problems of Pattern Recognition appear often to be overwhelming.

• Many of these sub-problems can indeed be solved.

• Many fascinating unsolved problems still remain.

Referenzen

ÄHNLICHE DOKUMENTE

We expected that the quality of a group strategy (i.e., its theoretical accuracy) likely depends on the average quality of individual strategies (here: measured by the recognition

Recall that all the coefficients are significant at the 1 percent level, so the β i estimates provide strong evidence that all the monthly releases contain incremental information

assess in real-life situations. The Harry Potter series seems to be particularly lenient for this purpose. Part of the popularity of the series is explained by the fact that

Furthermore, the costs in the present research on trials with six target events measured around 150 msec or so, consid- erably smaller than the costs observed in Smith (2003),

Organizarea contabilităţii de gestiune şi calculaţiei costurilor pe baze ştiinţifice în întreprinderile din industria fierului şi oţelului asigură premisele pentru

Government expenditure, private consumption expenditure, tax revenue, government debt, disposable income 4, government budget deficit and wealth 5 are the variables used in this

*Mixed PDK resins consist of circular PDK resin produced from X% recovery of waste PDK resin and the makeup produced from virgin materials for a 20,000 MT manufacturing plant..

The optimization of transport infrastructure as such is not a clearly defined process. Besides the optimization criteria - this report demonstrates both economic and energy-