• Keine Ergebnisse gefunden

Raw data

N/A
N/A
Protected

Academic year: 2021

Aktie "Raw data"

Copied!
64
0
0

Wird geladen.... (Jetzt Volltext ansehen)

Volltext

(1)

Multimedia Databases Multimedia Databases

Wolf-Tilo Balke Silviu Homoceanu

Institut für Informationssysteme

Technische Universität Braunschweig

http://www.ifis.cs.tu-bs.de

(2)

• Lecture

– 22.10.2009 – 04.02.2010

– 15:00-17:15 (3 lecture hours with a break) – Exercises, detours, and home work

discussion integrated into lecture

0. Organizational Issues

discussion integrated into lecture

• 4 Credits

• Exams

– Oral exam

– 50% of exercise points needed

to be eligible for the exam

(3)

• Recommended literature

– Schmitt: Ähnlichkeitssuche in Multimedia-Datenbanken,

Oldenbourg, 2005

0. Organizational Issues

Oldenbourg, 2005

– Steinmetz: Multimedia-Technologie:

Grundlagen, Komponenten und

Systeme, Springer, 1999

(4)

– Castelli/Bergman: Image Databases, Wiley, 2002

– Khoshafian/Baker: Multimedia and

0. Organizational Issues

– Khoshafian/Baker: Multimedia and Imaging Databases, Morgan

Kaufmann, 1996

– Sometimes: original papers (on our Web page)

(5)

• Course Web page

– http://www.ifis.cs.tu-bs.de/teaching/ws-0910/mmdb – Contains slides, excercises,

0. Organizational Issues

Contains slides, excercises, related papers and a video of the lecture

– Any questions? Just drop

us an email…

(6)

1. Introduction

1.1 What are multimedia databases?

1.2 Multimedia database applications 1.3 Evaluation of retrieval techniques

1. Introduction

1.3 Evaluation of retrieval techniques

(7)

• What are multimedia databases (MMDB)?

– Databases + multimedia = MMDB

1.1 Multimedia Databases

• Key words: databases and multimedia

• We already know databases, so what is

multimedia?

(8)

• Multimedia

– The concept of multimedia expresses the integration of different digital media types

– The integration is usually performed in a document

1.1 Basic Definitions

– Basic media types are text, image, vector graphics,

audio and video

(9)

• Text

– Text data, Spreadsheets, E-Mail, …

• Image

– Photos (Bitmaps), Vector graphics, CAD, …

• Audio

1.1 Data Types

• Audio

– Speech- and music records, annotations, wave files, MIDI, MP3, …

• Video

– Dynamical image record, frame-sequences,

MPEG, AVI, …

(10)

• Earliest definition of information retrieval:

“Documents are logically interdependent digitally encoded texts“

• Extension to multimedia documents allows the

1.1 Documents

• Extension to multimedia documents allows the

additional integration of other media types as

images, audio or video

(11)

• Document types

– Media objects are documents which are of only one type (not necessarily text)

– Multimedia objects are general documents which allow an arbitrary combination of

1.1 Documents

which allow an arbitrary combination of different types

• Multimedia data is transferred through the

use of a medium

(12)

• Medium

– A medium is a carrier of information in a communication connection

– It is independent of the transported information

1.1 Basic Definitions

– The used medium can also be

changed during information

transfer

(13)

• Book

1.1 Medium Example

– Communication between author and reader

– Independent from content – Independent from content – Hierarchically built on text

and images

– Reading out loud represents

medium change to sound/audio

(14)

• Based on receiver type

– Visual/optical medium – Acoustic mediums

– Haptical medium – through tactile senses

1.1 Medium Classification

Haptical medium – through tactile senses

– Olfactory medium – through smell – Gustatory medium – through taste

• Based on time

– Dynamic

– Static

(15)

• We now have seen…

– What multimedia is

– And how it is transported (through some medium)

1.1 Multimedia Databases

• But… why do we need databases?

– Most important operations of databases are data

storage and data retrieval

(16)

• Persistent storage of multimedia data, e.g.:

– Text documents

– Vector graphics, CAD – Images, audio, video

1.1 Multimedia Databases

Images, audio, video

• Content-based retrieval

– Efficient content based search

– Standardization of meta-data (e. g., MPEG-7, MPEG-21)

(17)

• Stand-alone vs. database storage model?

– Special retrieval functionality as well as corresponding optimization can be provided in both cases…

– But in the second case we also get the general advantages of databases

1.1 Multimedia Databases

advantages of databases

Declarative query language

Orthogonal combination of the query functionality

Query optimization, Index structures

Transaction management, recovery

...

(18)

1.1 Historical Overview

Retrieval procedures for text documents (Information Retrieval)

Relational Databases and SQL

1960

1970

Presence of multimedia objects intensifies

SQL-92 introduces BLOBs First Multimedia-Databases

1980

1990

2000

(19)

• Relational Databases use the data type BLOB (binary large object)

– Uninterpreted data

– Retrieval through metadata like e.g., file name, size, author, …

1.1 Commercial Systems

author, …

• Object-relational extensions feature enhanced retrieval functionality

– Semantic search

– IBM DB2 Extenders, Oracle Cartridges, …

– Integration in DB through UDFs, UDTs, Stored

Procedures, …

(20)

• Requirements for multimedia databases (Christodoulakis, 1985)

– Classical database functionality – Maintenance of unformatted data

1.1 Requirements

Maintenance – Consideration of

special storage and

presentation devices

(21)

• To comply with these requirements the following aspects need to be considered

– Software architecture – new or extension of existing databases?

– Content addressing – identification of the objects

1.1 Requirements

– Content addressing – identification of the objects through content-based features

– Performance – improvements using indexes,

optimization, etc.

(22)

– User interface – how should the user interact with the system? Separate structure from content!

– Information extraction – (automatic) generation of content-based features

– Storage devices – very large storage capacity,

1.1 Requirements

– Storage devices – very large storage capacity, redundancy control and compression

– Information retrieval – integration of some

extended search functionality

(23)

• Retrieval means the choice between data objects which can be based on…

– a SELECT condition (exact match) – or a defined similarity connection

(best match)

1.1 Retrieval

(best match)

• Retrieval may cover the delivery

of the results to the user, too

(24)

• Closer look at the search functionality

– „Semantic“ search functionality

– Orthogonal integration of classical and extended functionality

1.1 Retrieval

– Search does not directly access the media objects – Extraction, normalization and indexing of content-

based features

– Meaningful similarity/distance measures

(25)

• “Retrieve all images showing a sunset !”

1.1 Content-based Retrieval

• What exactly do these images have in common?

(26)

• Usually 2 main steps

– Example: image databases

1.1 Schematic View

Digitization Image

collection Image analysis Image analysis

and feature Image

Digitization

extraction and feature

extraction

Image database

Digitization Image

query Image analysis

extraction Image analysis

and feature extraction

Similarity search

Search result Querying the database

Creating the database

(27)

1.1 Detailed View

Query Result

3. Query

preparation 5. Result preparation

Query plan & feature values Result data

4. Similarity computation & query processing

2. Extraction of features

1. Insert into the database MM-Database

Feature values Raw & relational data

Raw data

(28)

1.1 More Detailed View

Query Result

Result preparation Medium transformation

Format transformation

Result data Query preparation

Normalization Segmentation Feature extraction

Optimization

Query plan Feature values

Relevance feedback

MM Database MM-Database BLOBs/CLOBs

Similarity computation Query processing

Feature values Feature index Feature extraction Feature recognition Feature preparation

Relational DB

Structure data

Relational DB

Metadata Profile Structure data Pre-processing

Decomposition Normalization Segmentation

(29)

• Lots of multimedia content on the Web

– Social networking e.g., Facebook, MySpace, Hi5, Orkut, etc.

– Photo sharing e.g., Flickr, Photobucket, Imeem, Picasa, etc.

1.2 Applications

etc.

– Video sharing e.g., YouTube, Megavideo, Metacafe,

blip.tv, Liveleak, etc.

(30)

• Cameras are everywhere

– In London “there are at least 500,000 cameras in the city, and one study showed that in a single

day a person could expect to be filmed 300 times”

1.2 Applications

day a person could expect to be filmed 300 times”

(31)

• Picasa face recognition

1.2 Applications

(32)

• Picasa, face recognition example

1.2 Applications

(33)

• Picasa, learning phase

1.2 Applications

(34)

• Picasa example

1.2 Applications

(35)

• Consider a police investigation of a large-scale drug operation

• Possible generated data:

– Video data captured by surveillance cameras – Audio data captured

1.2 Sample Scenario

– Audio data captured

– Image data consisting of still photographs taken by investigators

– Structured relational data containing background

information

– Geographic information system data

(36)

• Possible queries

– Image query by example:

police officer has a photograph and wants to find the identity of the person in the picture

1.2 Sample Scenario

of the person in the picture

Query: “retrieve all images from the image library in which the person appearing in the (currently displayed)

photograph appears”

– Image query by keywords: police officer wants to examine pictures of “Tony Soprano”

Query: “retrieve all images from the image library in which

‘Tony Soprano’ appears"

(37)

– Video Query:

Query: “Find all video segments in which Jerry appears”

By examining the answer of the above query, the police officer hopes to find other people

1.2 Sample Scenario

the police officer hopes to find other people who have previously interacted with the victim

(38)

– Heterogeneous Multimedia Query:

Find all individuals who have been photographed with “Tony Soprano” and who have been convicted of attempted

murder in New Jersey and who have recently had electronic fund transfers made into their bank accounts from ABC

Corp.

1.2 Sample Scenario

Corp.

(39)

• Basic difference

Static

High number of search queries (read access), few modifications of the data state

Dynamic

Often modifications of the data state

Active

1.2 Characteristics

Active

Database functionality lead to application operations

Passive

Database reacts only at requests from outside

Standard search

Queries are answered through the use of metadata e.g., Google-image search

Retrieval functionality

Content based search on the multimedia repository e.g., Picasa face recognition

(40)

• Passive static retrieval

– Art historical use case

1.2 Example

(41)

– Possible hit in a multimedia database

1.2 Example

(42)

• Active dynamic retrieval

– Wetter warning through evaluation of satellite photos

1.2 Example

Extraction

Typhoon-Warning for the Philippines

Extraction

(43)

• Standard search

– Queries are answered through the use of metadata e.g., Google-image search

1.2 Example

(44)

• Retrieval functionality

– Content based e.g., Picasa face recognition

1.2 Example

(45)

• Basic evaluation of retrieval techniques

– Efficiency of the system

Efficient utilization of system resources

Scalable also over big collections

– Effectivity of the retrieval process

1.3 Towards Evaluation

– Effectivity of the retrieval process

High quality of the result

Meaningful usage of the system

– Weighting is application specific

(46)

• Characteristic values to measure efficiency are e.g.:

– Memory usage – CPU-time

1.3 Evaluating Efficiency

– Number of I/O-Operations – Response time

• Depends on the (Hardware-) environment

• Goal: the system should be efficient enough!

(47)

• Measuring effectivity is more difficult and always depending on the query

• Goal: define some query-dependent evaluation measures!

1.3 Evaluating Effectivity

measures!

– Objective quality metrics

– Result evaluation based on the query

– Independent from used querying interface and retrieval procedure

– Leads to comparability of different systems/algorithms

(48)

• Effectivity can be measured regarding some explicit query

– Main focus on evaluating the behavior of the system with respect to a query

– Relevance of the result set

1.3 Evaluating Effectivity

– Relevance of the result set

• But effectivity also needs to consider implicit information needs

– Main focus on evaluating the usefulness, usability and

user friendliness of the system

(49)

• To evaluate a retrieval system over some query, each document will be classified binary as

relevant or irrelevant with respect to the query

1.3 Relevance

– This classification is performed by “experts”

– The response of the system to the query will be compared to this manual classification

Compare the obtained response with the “ideal” result

(50)

• Subjective measure estimating to what degree the information need of the user is satisfied

– Difficult to measure (empirical studies) – Questionable instrument for comparing

procedures/systems

1.3 Usefulness (Pertinence)

procedures/systems

• Attention: useful documents can be irrelevant when considering the query (serendipity)

• In this lecture:

explicit query evaluation measures only!

(51)

1.3 Involved Sets

searched for

collection

found searched for

(= relevant)

found

(= query result)

(52)

• Irrelevant documents, classified as relevant by the system

– False alarms, false drops, …

1.3 False Positives

• Needlessly increase the result set

• Usually inevitable (ambiguity)

• Can be easily eliminated by the user

(53)

• Relevant documents classified by the system as irrelevant

– False dismissals

1.3 False Negatives

• Dangerous, since they can not be detected easily by the user

– Does the collection contain “better” documents?

– False positives are usually not as bad as false negatives

(54)

• Correct positives (correct alarms)

– All documents correctly classified by the system as relevant

• Correct negatives (correct dismissals)

1.3 Remaining Sets

– All documents correctly classified by the system as irrelevant

• All sets are disjunctive and their reunion is the

entire document collection

(55)

1.3 Overview

irrelevant relevant

System- evaluation User-

evaluation

cd irrelevant fa

fd

relevant ca

(56)

• {Relevant results} = fd + ca

• {Retrieved results} = ca + fa

1.3 Interpretation

collection

searched for found

ca

cd

fa

fd

(57)

• Precision measures the ratio of correctly returned documents relative to all returned documents

– P = ca / (ca + fa)

1.3 Precision

• Value between [0, 1]

(1 representing the best value)

• High number of false alarms mean

worse results

(58)

• Recall measures the ratio of correctly returned documents relative to all relevant documents

– R = ca / (ca + fd)

• Value between [0, 1]

1.3 Recall

• Value between [0, 1]

(1 representing the best value)

• High number of false drops mean

worse results

(59)

• Both measures only make sense, if considered at the same time

– E.g., get perfect recall by simply returning all

documents, but then the precision is extremely low…

• Can be balanced by tuning the system

1.3 Precision-Recall Analysis

• Can be balanced by tuning the system

– E.g., smaller result sets lead to better precision rates at the cost of recall

• Usually the average precision-recall of more

queries is considered (macro evaluation)

(60)

• Alarms (returned elements) can be easily divided in ca and fa

– Precision is easy to calculate

• Dismissals (not returned elements) are not so trivial to divide in cd und fd, because the entire

1.3 Actual Evaluation

trivial to divide in cd und fd, because the entire collection has to be classified

– Recall is difficult to calculate

• Standardized Benchmarks

– Provided connections and queries

– Annotated result sets

(61)

• Text REtrieval Conference

• De-Facto-Standard since 1992

• Establish average precision for 11 fixed recall points (0; 0,1; 0,2; …; 1) according to defined

1.3 TREC

points (0; 0,1; 0,2; …; 1) according to defined procedures (trec_eval)

• Different tracks, extended also for video data,

´Web retrieval and Question Answering

• Other initiatives: e.g., CLEF (cross-language

retrieval) or INEX (XML-Documents)

(62)

1.3 Example

4 cd

0,2 P

0,25 R 6

2 8

Q

1

fd ca

fa Query

8

0,5 0,8

0,525 0,8 Average

2 8

2 Q

2

1

(63)

• Precision-Recall-Graphs

1.3 Representation

Average precision of the system 3 at a recall-level of 0,2

System 1 System 2 System 3

Which system is the best?

What is more

important: recall or precision?

(64)

• Retrieval of images by color

• Introduction to color spaces

• Color histograms

• Matching

Next lecture

• Matching

Referenzen

ÄHNLICHE DOKUMENTE

Dieser Beitrag zeigt, wie mit der Analysesoftware Atlas.ti aus digitalisiertem Rohmaterial in Form von Audiodaten ein Klickbares Inhaltsverzeichnis (C-TOC) erstellt wird,

The creation of a Clickable Table of Contents (C-TOC) is one example of how to use this programme in order to structure raw data. To do this the interview—or other kind of

The unit extension can be easily dealt with adding declarations of the form !UNIT element 1 element 2 unit , where element 1 refers to the element defining the media type, element

Trotzdem kann auch diese Darstellung in der Praxis sehr schnell unhandlich werden. Diese Struktur eignet sich nicht um Anfragen zu beantworten, die Segmente manipulieren, wie dies

<Dxy der Kreuzkorrelation zwischen den Signalen des vorderen und des hinteren Brennstoff/Luft-Gemisch- Sensors, die beide vom Filtersystem verarbeitet wur- den; und

Article optimised or newly created by the editorial team with the text/image standard of the magazine. The focus should be on the added value of the content for the reader.

По результатам прохождения профессиональной переподготовки специалисты получают диплом государственного образца, удостоверяющий их право (квалификацию)

tablish a firm relationship with Herakles, but the set of sculpture on public display also aimed at depicting a still traditional (as most metopes show), but in some scenes