• Keine Ergebnisse gefunden

Task-Driven Sparse Coding for Classification of Motion Data

N/A
N/A
Protected

Academic year: 2022

Aktie "Task-Driven Sparse Coding for Classification of Motion Data"

Copied!
17
0
0

Wird geladen.... (Jetzt Volltext ansehen)

Volltext

(1)

Task-Driven Sparse Coding for Classification of Motion Data

Babak Hosseini, Barbara Hammer

24. July. 2017

Bielefeld University

Cognitive Interaction Technology - Centre of Excellence

1

(2)

Sparse coding

Sparse coding:

– Providing a sparse representation of data – Y: Matrix of data vectors (columns)

D: a matrix of basic primitives (Dictionary)X: Matrix of the Coefficient Vector

– Sparsity constraints/objectives

• 1-norm, cardinality, etc

(3)

Classification tasks

Sparse coding Classification:

– Classifier on top of sparse representation (X)

• SVM

• Linear/nonlinear classifier

• nonlinear feature mapping

(4)

Classification tasks

Sparse coding Classification:

– Classifier on top of sparse representation (X)

• SVM

• Linear/nonlinear classifier

• nonlinear feature mapping

!

(5)

Classification tasks

Sparse coding Classification:

– Classifier on top of sparse representation (X)

• SVM

• Linear/nonlinear classifier

– Augmenting the classifier to the main optimization:

f: Classifier function/objective

(6)

Classification tasks

Task Driven dictionary learning

– Sparse Coding

Classifier

– Bi-level optimization – Coupled

(7)

Classification tasks

Task Driven dictionary learning

If 𝑋is closed-form solution based on D

Augmenting {X,D} relationship into the 2ndoptimization

(8)

Classification tasks

Task Driven dictionary learning

If 𝑋is closed-form solution based on D

Augmenting {X,D} relationship into the 2ndoptimization Optimizing D based on the classification task

(9)

Task driven framework

Task-driven sparse coding:

Each 𝑥𝑖 use a different 𝐷𝐼𝑖

𝐼𝑖: selected columns of D to reconstruct 𝑦𝑖

Each 𝑥𝑖 would result in a different 𝑔𝑖(𝐷𝐼𝑖 , 𝑊)

𝑔 𝐷, 𝑊 = 𝑔𝑖(𝐷𝐼𝑖 , 𝑊)

!! Not a single structure

Solution: Stochastic GD methods

Batch Optimization

(10)

Task driven framework

Alternating Optimization:

• Solving wrt. X* , D*, W* in a sequence in a loop

(11)

Our task-driven framework

Task Driven Kernel Sparse Coding:

– Kernel space

Φ D = Φ Y ∗ 𝐴

𝐴 ∈ ℝ𝑁×𝑘 in the input space

– None-Negative framework

• {X,A} are positive  interpretability

– Linear classifier

• H: training labels

(12)

Derivations:

nnKOMP solution :

𝐼: selected columns from 𝐴

Calculating :

Task Driven K-Sparse Coding

(13)

Task Driven K-Sparse Coding

Task Driven Algorithm:

Loop till convergence:

Finding X*: None-negative kernel OMP

Finding A*: stochastic projected gradient descent Finding W*: linear programing (ridge regression)

(14)

Application

Motion data classification:

– Multi-dimension Time-series

– Kernel matrix:

• Pair-wise similarity between the motions {𝑦𝑖, 𝑦𝑗}

• Using DTW distance

i i i+2

time

(15)

Experiments

Results:

15

(16)

Conclusion and Future works

Summary

• Task-driven framework orients sparse coding towards the classification objective.

• The none-negative sparse representation improves classification performance.

• Non-negative kernel framework provides an interpretable model while classifying the data.

Future works:

Feature based classifier via an additional parameter – Online version of the problem

– Enhancing the optimization strategy  Speed, Robustness

16

(17)

Thank you very much!

17

Referenzen

ÄHNLICHE DOKUMENTE

dency patterns can find opposites that belong to different PoS categories is useful for understanding the benefits of syntactic information for relation extraction, as well as

Diese Durchschnitt er- setzen wir nun durch ein Zufallsexperiment: Wir ziehen einen einzelnen Datenpunkt z i und betrachten nur die partielle Ableitung, die sich f¨ ur diesen

4 we confirm that the results are similar to the convex case; i.e., SVRG reduces the variance and smoothly converges faster than the best-tuned SGD with learning rate scheduling,

In particular, Alon, Matias and Szegedy [AMS99] give an Ω(n) lower bound on the space complexity of any algorithm for estimating the frequency of the largest item given an

In particular, second order stochastic gradient and averaged stochastic gradient are asymptotically efficient after a single pass on the training set.. Keywords: Stochastic

For the classical (Signed) Max-Cut problem, and its wide generalization to strongly λ-extendable properties, parameterized above the classical Poljak-Turzík bound, we improved the

Because Rules 14–17 is exactly the subset of Rules 1–7 applicable to signed graphs without positive edges, and because the kernelization Rules 8–9 for Signed Max- Cut AEE and

► Asbestbedingte Erkrankungen sind für mehr als 60% aller Todesfälle infolge einer Berufserkrankung verantwortlich.. STOP dem Krebs