• Keine Ergebnisse gefunden

Selecting discriminative features with discriminative multiple canonical correlationanalysis for multi-feature information fusion

N/A
N/A
Protected

Academic year: 2022

Aktie "Selecting discriminative features with discriminative multiple canonical correlationanalysis for multi-feature information fusion"

Copied!
8
0
0

Wird geladen.... (Jetzt Volltext ansehen)

Volltext

(1)

Selecting Discriminative Features with Discriminative Multiple Canonical Correlation Analysis for Multi-Feature

Information Fusion

Lei Gao1, Lin Qi1, Ling Guan1,2

1 School of Information Engineering 2 Electrical and Computer Engineering Zhengzhou University Ryerson University 450001, Zhengzhou, China M5B 2K3, Toronto, Canada

email- iegaolei@gmail.com; ielqi@zzu.edu.cn lguan@ee.ryerson.ca Abstract: In this paper, it presents a novel approach for selecting discriminative features in multimodal information fusion based discriminative multiple canonical correlation analysis (DMCCA), which is the generalized form of canonical correlation analysis (CCA), multiple canonical correlation analysis (MCCA) and discriminative canonical correlation analysis (DCCA). The proposed approach identifies the discriminative features from the multi-feature in Fractional Fourier Transform (FRFT) domain, which are capable of simultaneously maximizing the within-class correlation and minimizing the between-class correlation, leading to better utilization of the multi-feature information and producing more effective pattern recognition results. The effectiveness of the introduced solution is demonstrated through extensive experimentation on a visual based emotion recognition problem.

1 Introduction

The effective utilization and integration of multiple information contents presented in different media sources are becoming an increasingly important research topic in many applications with the proliferation of multimedia and the advances in sensing technology.

Since single information based pattern analysis and recognition systems only afford low level of performance due to the drastic variation and noisy nature of the acquired signals, it leads to insufficient and inaccurate pattern representation of the perception of interest.

However, multiple data contains more information about the semantics presented in the media. The combination of multiple data may potentially provide a more complete and discriminatory description of the intrinsic characteristics of the pattern, and produce improved system performance compared with single information only [LYRYAM-2010].

Therefore, information fusion becomes an emerging and challenging research area in multimedia processing. The major difficulties lie in the identification of the inherent relationship between different information, and the design of a fusion strategy that can effectively utilize the complementary information presented in different channels. A wide variety of methods have been proposed in the literature to address the difficulties.

In general, there are three levels of information fusion: feature/data level, score level and decision level [AA-2003]. Compared with other methods, the advantage of the feature level fusion is as follows. As different feature vectors extracted from the same pattern tend to reflect different characteristics of the pattern, optimally combining these features not only keeps the effective discriminant information of multi-feature, but also eliminates the redundant information to certain degree, which is especially important to classification and recognition of large scale database in high dimensional feature space.

It is also the focus of this paper.

Generally, there exist two traditional classes of feature fusion methods which are serial feature fusion and parallel feature fusion [JJDJ-2003]. Recently, there has been extensive interest in the analysis of correlation based approaches for multi-feature information

(2)

fusion such as canonical correlation analysis (CCA) [XK-1999], kernel canonical correlation analysis (KCCA) [CP-2001], discriminative CCA (DCCA) [TSZJ-2007] and multi-set canonical correlation analysis (MCCA) [A-2002], which have been applied to audiovisual based talking-face biometric verification [HG-2007], medical imaging analysis [NTYV-2010], handwriting recognition [QSYPD-2005], audio-visual synchronization [MYEA-2007], joint blind source separation [TWV-2009], blind single- input and multiple-output (SIMO) channels equalization [JID-2005]. However, as CCA, KCCA or DCCA method only could deal with the mutual relationships between two random vectors and it lacks discriminative character for MCCA. In order to address the mentioned problems, the approach of DMCCA is introduced in [LLEL-2012], which extracts more discriminative characteristics of multi-feature information. Nevertheless, one important yet not well studied problem in DMCCA is that there is not any reliable approach to select the discriminative features from multi-feature information to achieve better recognition results.

In this paper, we conduct a novel approach for selecting discriminative features in multimodal information fusion based discriminative multiple canonical correlation analysis (DMCCA). The proposed approach identifies the more discriminative features in FRFT domain on a visual based emotion recognition problem. The effectiveness of the introduced solution is demonstrated through extensive experimentation. The remainder of this paper is organized as follows: the conception of FRFT and 2D-FRFT is briefly introduced in Section 2. In Section 3, the analysis and derivation of the proposed approach is presented. The emotion recognition system and experimental results are given in Section 4. Conclusions are drawn in Section 5.

2 FRFT and 2D-FRFT

FRFT is a generalized form of the FT, which can be interpreted as a rotation of the signal in the time-frequency plane [L-1994]. It contains simultaneity the time-frequency information of the signal, and is considered as a new tool for time-frequency analysis, especially in the area of image representation [HD-1993, A-1993]. In the area of facial expression recognition, some researches have shown its superiority with respect to other feature extraction tools [LLEXL-2010]. As different order features of 2D-FRFT contain different time-frequency information [CMH-2000], and the previous studies only focus on the single order FRFT features. Thus, it is reasonable to fuse the different features to improve the recognition results.

In this section, we introduce the definition and properties of the FRFT and the two- dimensional form 2D-FRFT.

Given a signal , its FRFT is defined as:

X uα

( ) {

F x tα

( ) } ( )

u x t K t u dt

( ) ( )

α ,

== ⎡⎣ ⎤⎦

−∞ (1)

2 2

1 cot exp cot csc ,

2 2

( ), 2

( ), (2 1)

( , )

j jt u tu n

t u n

t u n

K t uα

α α α α π

π

δ α π

δ α π

+

⎪ − =

⎪ + = ±

⎪⎩

=

(2)

where = (p* )/2 is the rotation angle in FRFT domain and p is the transform order.

The FRFT can be extended to its two-dimensional form. For a two-dimensional signal

( , )

x s t

, its 2D-FRFT is defined as:

(3)

fusion such as canonical correlation analysis (CCA) [XK-1999], kernel canonical correlation analysis (KCCA) [CP-2001], discriminative CCA (DCCA) [TSZJ-2007] and multi-set canonical correlation analysis (MCCA) [A-2002], which have been applied to audiovisual based talking-face biometric verification [HG-2007], medical imaging analysis [NTYV-2010], handwriting recognition [QSYPD-2005], audio-visual synchronization [MYEA-2007], joint blind source separation [TWV-2009], blind single- input and multiple-output (SIMO) channels equalization [JID-2005]. However, as CCA, KCCA or DCCA method only could deal with the mutual relationships between two random vectors and it lacks discriminative character for MCCA. In order to address the mentioned problems, the approach of DMCCA is introduced in [LLEL-2012], which extracts more discriminative characteristics of multi-feature information. Nevertheless, one important yet not well studied problem in DMCCA is that there is not any reliable approach to select the discriminative features from multi-feature information to achieve better recognition results.

In this paper, we conduct a novel approach for selecting discriminative features in multimodal information fusion based discriminative multiple canonical correlation analysis (DMCCA). The proposed approach identifies the more discriminative features in FRFT domain on a visual based emotion recognition problem. The effectiveness of the introduced solution is demonstrated through extensive experimentation. The remainder of this paper is organized as follows: the conception of FRFT and 2D-FRFT is briefly introduced in Section 2. In Section 3, the analysis and derivation of the proposed approach is presented. The emotion recognition system and experimental results are given in Section 4. Conclusions are drawn in Section 5.

FRFT is a generalized form of the FT, which can be interpreted as a rotation of the signal in the time-frequency plane [L-1994]. It contains simultaneity the time-frequency information of the signal, and is considered as a new tool for time-frequency analysis, especially in the area of image representation [HD-1993, A-1993]. In the area of facial expression recognition, some researches have shown its superiority with respect to other feature extraction tools [LLEXL-2010]. As different order features of 2D-FRFT contain different time-frequency information [CMH-2000], and the previous studies only focus on the single order FRFT features. Thus, it is reasonable to fuse the different features to improve the recognition results.

In this section, we introduce the definition and properties of the FRFT and the two- dimensional form 2D-FRFT.

Given a signal , its FRFT is defined as:

X uα

( ) {

F x tα

( ) } ( )

u x t K t u dt

( ) ( )

α ,

== ⎡⎣ ⎤⎦

−∞ (1)

2 2

1 cot exp cot csc ,

2 2

( ), 2

( ), (2 1)

( , )

j jt u tu n

t u n

t u n

K t uα

α α α α π

π

δ α π

δ α π

+

⎪ − =

⎪ + = ±

⎪⎩

=

(2)

where = (p* )/2 is the rotation angle in FRFT domain and p is the transform order.

The FRFT can be extended to its two-dimensional form. For a two-dimensional signal

( , )

x s t

, its 2D-FRFT is defined as:

X u v Fα β, ( , )= βs v

{

Fαt u

[

x t s( , )

] }

(3)

where = (p* )/2, = (q* )/2 are the rotation angles, and p, q are the transform orders in the 2D-FRFT. It satisfies the property of additivity in 2D-FRFT [SH-1998]:

Xα ϕ β θ+,+( , )u v F= θv v'

{

Fϕu u' Xα β,( , )u v' '

}

(4)

In the fields of digital image processing, the two dimensional discrete FRFT can be implemented by row-column computation as shown in [SH-1998].

3 The Approach of Selecting Discriminative Feature for DMCCA

Let X X1, 2,�XQ be Q sets zero-mean random samples as:

X x xi=∈[i1(1), i2(1)xin1(1),�x xi1( )c, i2( )cxinc( )c] Rp Ni× (5)

where i is the number sequence of the random samples,

x

ij( )m denotes the jth sample in the mth class, respectively, and

n

il is the number of samples in the lth class of

X

i set.

1 c l il

n N

=

= (6) where c is the total number of classes.

The aim of DMCCA is to seek the projection vectors ω=[ ,ω ω1 2�ωN]for feature extraction so that the within-class correlation is maximized and the between-class correlation is minimized. With the definition of DMCCA in [LLEL-2012], it can be written as:

1 ( )

1 C D D

N − ω ρ ω=

− (7) where

1 1 1

1

T T

N

T T

N N N

XX XAX

C

X AX X X

⎡⎛ ⎞⎤

⎢⎜ ⎟⎥

=⎢⎜ ⎟⎥

⎢⎜⎝ ⎟⎠⎥

⎣ ⎦

� � �

1 1 0

0

T

N NT

X X D

X X

⎡⎛ ⎞⎤

⎢⎜ ⎟⎥

=⎢⎜ ⎟⎥

⎢⎜⎝ ⎟⎠⎥

⎣ ⎦

� � �

1 1 0

0

i i

il il n nic ic n n

n n n n

I

A I R

I ×

×

×

×

⎡⎛ ⎞⎤

⎢⎜ ⎟⎥

⎢ ⎥

=∈⎜ ⎟

⎢⎜⎜ ⎟⎟⎥

⎢⎝ ⎠⎥

⎣ ⎦

� �

� with I=[1,1, 1]� TRn and ω=[ω ω1T, 2T�ωNT T] .

Therefore, the solution obtained is the eigenvector associated to the eigenvalue of equation (7). That is:

1 ( )*( )

1inv D C D

N − ω ρω=

− (8)

where inv means inverse transform of matrix. However, unless the covariance matrices D have full rank, the block matrix in Eq. (8) will become singular. An approach [TMH- 2003] to dealing with singular covariance matrices and to controlling complexity is to add a multiple of the identity matrix

λ Ι

,

λ > 0

to D.

(4)

Thus, the general form of equation (8) is written:

1 ( )*( )

1inv D C D

N + − ω ρω=

− (9)

where

when is insingular matrix I when is singular matrix

D D

D+ D λ D

= ⎨⎧⎩ +

From equation (7), the

ρ

is the criterion to seek the projection vectors for feature extraction. That is to say, the value of

ρ

is the key parameter to the effect of selecting discriminative features, and the larger

ρ

corresponds to the more discriminative features, which effectively maximize the within-class correlation and minimize the between-class correlation. While the smaller

ρ

corresponds to the less discriminative features. Thus, it is reasonable to evaluate the final information fusion results by judgment J( i) :

1

( )i N i i

J ρ ρ

=

=

(10) where

ρ

i is the ith eigenvalue of equation (9).

Furthermore, since

1

( 1 ( )*( ))

1

N i i

trance inv D C D

N + ρ

=

− =

(11)

where trance( ) represents the trace of the matrix. Thus, we can calculate J( i) through trance of the matrix instead of the eigenvalue decomposition when the number of features is large, which could significantly speed up the calculation process.

4 Emotion Recognition System and Experimental Results Analysis

4.1Emotion recognition system

In this paper, we demonstrate the effectiveness of the proposed method on a visual based emotion recognition problem with FRFT features. Figure 1 depicts a general block diagram of the proposed emotion recognition system.

Figure 1: The proposed emotion recognition system

For facial expression representation, a key image which corresponds to the frame with largest speech amplitude is first identified from the audiovisual. Then we detect the face region from the image frame using a color based method [YL-2008]. Then the face

(5)

Thus, the general form of equation (8) is written:

1 ( )*( )

1inv D C D

N + − ω ρω=

− (9)

where

when is insingular matrix I when is singular matrix

D D

D+ D λ D

= ⎨⎧⎩ +

From equation (7), the

ρ

is the criterion to seek the projection vectors for feature extraction. That is to say, the value of

ρ

is the key parameter to the effect of selecting discriminative features, and the larger

ρ

corresponds to the more discriminative features, which effectively maximize the within-class correlation and minimize the between-class correlation. While the smaller

ρ

corresponds to the less discriminative features. Thus, it is reasonable to evaluate the final information fusion results by judgmentJ( i) :

1

( )i N i i

J ρ ρ

=

=

(10) where

ρ

i is the ith eigenvalue of equation (9).

Furthermore, since

1

( 1 ( )*( ))

1

N i i

trance inv D C D

N + ρ

=

− =

(11)

where trance( ) represents the trace of the matrix. Thus, we can calculate J( i) through trance of the matrix instead of the eigenvalue decomposition when the number of features is large, which could significantly speed up the calculation process.

In this paper, we demonstrate the effectiveness of the proposed method on a visual based emotion recognition problem with FRFT features. Figure 1 depicts a general block diagram of the proposed emotion recognition system.

Figure 1: The proposed emotion recognition system

For facial expression representation, a key image which corresponds to the frame with largest speech amplitude is first identified from the audiovisual. Then we detect the face region from the image frame using a color based method [YL-2008]. Then the face

region is normalized to an image of size of 112 × 96. As the large dimensionality of the coefficients, we downsample each subband to a size of 32 × 32 to enhance the recognition process. The extracted FRFT features in each of the time segments are then analyzed and selected using the proposed analysis approach. Subsequently, the newly generated features, which represent the multi-set information among different patterns, are concatenated into projected vectors for classification through the method of DMCCA.

Then, the nearest-neighbor classifier is used for emotion recognition with visual features.

4.2Experimental results analysis

To evaluate the performance of the proposed method, we conduct experiments on the RML audiovisual emotion database [YL-2008], which consists of video samples from eight human subjects, speaking six different languages (English, Mandarin, Urdu, Punjabi, Persian, and Italian). In the experiment, a total of 288 samples are selected from the RML audiovisual database, each belonging to one of the six universal emotional states-anger(AN), disgust(DI), fear(FE), sadness(SA), surprise(SU) and happiness(HA).

Among the samples, 192 samples are chosen for training set and 96 are chosen for evaluation. As a benchmark, the performances of using single FRFT features in different transform orders (from 0.1 to 1.0) are first evaluated as Figure 2. From Figure 2, it is observed that recognition results are dissimilar in different transform orders due to different time-frequency information, which could be applied in the field of information fusion with expecting to reach higher recognition accuracy. In the experiment, we select the five single features (p=0.3, 0.7, 0.8, 0.9, 1.0) with better recognition accuracy from Figure 2 as the fused multi-feature. Based on equation (14) (15), the values of J( i)fused with different 2D-FRFT features in DMCCA method are shown as Table 1, which mean the discrimination of the fused features.

0.1 0.3 0.5 0.7 0.9 1

66 67 68 69 70 71 72 73 74 75

Recognition Results(%)

Transform Orders

Figure 2: Results of emotion recognition with single FRFT feature Table 1. Results of ( )J ρi with different number of 2D-FRFT features

FRFT Features 0.8, 0.9 0.8,0.9,1.0

( )i

J ρ -0.24 0.12

FRFT Features 0.7,0.8,0.9,1.0 0.3,0.7,0.8,0.9,1.0 ( )i

J ρ 8.93 0.04

From Table 1, it can be observed that the value of J( i) with four 2D-FRFT features(0.7,0.8,0.9,1.0) obtains the maximum 8.93, while the result with two 2D-FRFT features(0.8,0.9) arrives the minimum -0.24. Based on our analysis, as J( i) means the discrimination of the fused features, the recognition result fused with four 2D-FRFT features(0.7,0.8,0.9,1.0) should outperform others. In order to demonstrate the effectiveness of the proposed method on 2D-FRFT features domain and our

(6)

mathematical analysis, the experiments on visual emotion recognition using the method of DMCCA performed with the different number of 2D-FRFT features are shown as follows. In the experiments, it is worthwhile to point out that, as there are 32*32 dimensions in the visual recognition problem, we adopt the first 50 projected dimensions by the approach of DMCCA. The overall recognition rates are shown in Figure 3.

Furthermore, we also perform the highest recognition accuracy under different number of features, which can be shown as Table 2. Besides, in order to demonstrate the efficiency of the proposed method effectively, the relation between J( i) and highest recognition results is shown as Figure 4. It can be seen that the experimental results comply with our mathematical analysis presented in the previous section.

0 5 10 15 20 25 30 35 40 45 50

30 40 50 60 70 80 90 100

p=0.8,0.9 p=0.8,0.9,1.0 p=0.7,0.8,0.9,1.0 p=0.3,0.7,0.8,0.9,1.0

Projected Dimensions Recognition

Results (%)

Figure.3 Emotion recognition experimental results of different number FRFT features Table 2 Highest emotion recognition results of different FRFT features

FRFT Features 0.8, 0.9 0.8,0.9,1.0

Highest results(%) 83.33 85.42

FRFT Features 0.7,0.8,0.9,1.0 0.3,0.7,0.8,0.9,1.0

Highest results(%) 90.67 84.38

2 3 4 5

-1 0 1 2 3 4 5 6 7 8 9

2 3 4 5

83 84 85 86 87 88 89 90 91

Values of J(p)

The number of FRFT features

Highest results (%)

The number of FRFT features

Figure 4 Relation between highest recognition results and values of J( )ρi

Furthermore, in order to demonstrate the effectiveness of the proposed method in 2D- FRFT features domain, we also implemented feature fusion with the state-of-the-art methods such as CCA, MCCA and DCAA with different 2D-FRFT features for the purpose of comparison. The recognition rates with different projected dimensions are shown in Figure 5 and the highest recognition results with mentioned methods are shown as Figure 6. Based on Figure 5 and Figure 6, it can be seen that the highest recognition rates were achieved with the proposed method on selecting discriminative features in 2D-FRFT features with DMCCA.

(7)

mathematical analysis, the experiments on visual emotion recognition using the method of DMCCA performed with the different number of 2D-FRFT features are shown as follows. In the experiments, it is worthwhile to point out that, as there are 32*32 dimensions in the visual recognition problem, we adopt the first 50 projected dimensions by the approach of DMCCA. The overall recognition rates are shown in Figure 3.

Furthermore, we also perform the highest recognition accuracy under different number of features, which can be shown as Table 2. Besides, in order to demonstrate the efficiency of the proposed method effectively, the relation between J( i) and highest recognition results is shown as Figure 4. It can be seen that the experimental results comply with our mathematical analysis presented in the previous section.

Figure.3 Emotion recognition experimental results of different number FRFT features Table 2 Highest emotion recognition results of different FRFT features

FRFT Features 0.8, 0.9 0.8,0.9,1.0

Highest results(%) 83.33 85.42

FRFT Features 0.7,0.8,0.9,1.0 0.3,0.7,0.8,0.9,1.0

Highest results(%) 90.67 84.38

Figure 4 Relation between highest recognition results and values of J( )ρi

Furthermore, in order to demonstrate the effectiveness of the proposed method in 2D- FRFT features domain, we also implemented feature fusion with the state-of-the-art methods such as CCA, MCCA and DCAA with different 2D-FRFT features for the purpose of comparison. The recognition rates with different projected dimensions are shown in Figure 5 and the highest recognition results with mentioned methods are shown as Figure 6. Based on Figure 5 and Figure 6, it can be seen that the highest recognition rates were achieved with the proposed method on selecting discriminative features in 2D-FRFT features with DMCCA.

0 5 10 15 20 25 30 35 40 45 50

30 40 50 60 70 80 90 100

CCA(0.8,0.9) CCA(0.9,1.0) DCCA(0.8,0.9) DCCA(0.9,1.0) MCCA(0.3,0.7,0.8,0.9,1.0) MCCA(0.7,0.8,0.9,1.0) MCCA(0.8,0.9,1.0) DMCCA(0.7,0.8,0.9,1.0)

Projected Dimensions Recognition

Results(%)

Figure 5 The overall emotion recognition experimental results with different feature fusion

0 10 20 30 40 50 60 70 80 90

100 CCA(0.8,0.9)

CCA(0.9,1.0) DCCA(0.8,0.9) DCCA(0.9,1.0) MCCA(0.8,0.9,1.0) MCCA(0.7,0.8,0.9,1.0) MCCA(0.3,0.7,0.8,0.9,1.0) DMCCA(0.7,0.8,0.9,1.0)

Recognition Results (% )

The highest recognition of different methods

Figure 6 The highest emotion recognition experimental results with different feature fusion

5 Conclusion

This paper has introduced a new approach for effectively selecting discriminative features in multimodal information fusion based the method of DMCCA. The proposed solution identifies the discriminative features from multiple features for information fusion. The effective combination of multi-feature data potentially provides a more complete and discriminatory description of the intrinsic characteristics of the pattern, and produce improved system performance. Experimental results on a visual based emotion recognition problem in FRFT domain demonstrate the proposed method selects more discriminative features effectively, improves the recognition performance, and is computationally efficient. Although we focus on an emotion recognition problem in FRFT domain in this paper, the proposed solution can also be applied to other multimodal or multi-feature related multimedia analysis problems. Another interesting and challenging topic will be investigating kernelized version of the DMCCA based on the Kernel canonical correlation analysis (KCCA) in order to obtain a more effective method to solve nonlinear fusion problems which are more prevalent in information fusion.

6 Acknowledgement

This work is supported by the National Natural Science Foundation of China, No.61071211 and the Canada Research Chair Program.

(8)

References

[LYRYAM-2010] L. Guan; Y. Wang; R. Zhang; Y. Tie; A. Bulzacki and M. T. Ibrahim:

Multimodal information fusion for selected multimedia applications. Int. J.

Multimedia Intell. Sec., vol. 1, no. 1, 2010; S. 5–32.

[AA-2003] A. Ross; A. K. JainL.: Information fusion in biometrics. Pattern Recognition.

Lett., vol. 24, no. 13, 2003; S. 2115–2125.

[JJDJ-2003] J. Yang; J.Y. Yang; D. Zhang; J.F. Lu.: Feature fusion: parallel strategy vs.

serial strategy. Pattern Recognition. Lett., vol. 36, no. 6, 2003; S. 1369–1381.

[XK-1999] X.T. Zhang; K.T. Fang: Multivariate Statistical Introduction. Sciences Press, Beijing, 1999.

[CP-2001] C. Fyfe; P. L. Lai: Kernel and nonlinear canonical correlation analysis. Int. J.

Neural Syst., vol. 10, 2001; S. 365–374.

[TSZJ-2007] T.K. Sun; S.C. Chen; Z. Jin; J.Y. Yang: Kernelized Discriminative Canonical Correlation Analysis. International Conference on Wavelet Analysis and Pattern Recognition, vol. 2, 2007; S. 1283-1287.

[A-2002] A. A. Nielsen: Multiset Canonical Correlations Analysis and Multispectral, Truly Multitemporal Remote Sensing Data. IEEE Trans. Image Processing, vol. 11, no. 3, Mar. 2002; S. 293–305.

[HG-2007] H. Bredin; G. Chollet: Audio-visual speech synchrony measure for talking- face Identity verification. in Proc. The IEEE International Conference on Acoustics, Speech, and Signal Processing, Honolulu, HI, vol. 2, Apr. 2007; S.

233–236.

[NTYV-2010] N. M. Correa; T. Adali; Y. Li; V. D. Calhoun: Canonical correlation analysis for data fusion and group ingerences: examining applications of medical imaging data. IEEE Signal Process. Mag., vol. 27, no. 4, 2010; S. 39–50.

[QSYPD-2005] Q. Sun; S. Zeng; Y. Liu; P. Heng; D. Xia: A new method of feature fusion and its application in image recognition. Pattern Recognition, vol. 36, no. 12, Dec. 2005; S. 2437–2448,.

[MYEA-2007] M. E. Sargin; Y. Yemez; E. Erzin; A. M. Tekalp: Audiovisual synchronization and fusion using canonical correlation analysis. IEEE Trans.

Multimedia, vol. 9, no. 7, Nov. 2007; S. 1396–1403.

[TWV-2009] T. Adali; W. Wang; V.D. Calhoun: Joint Blind Source Separation by Multiset Canonical Correlation Analysis. IEEE Trans. on Signal Processing, vol. 57, no. 10, Oct. 2009; S. 3918-3929.

[JID-2005] J. Via; I. Santamara; d J. Pérez: Canonical correlation analysis (CCA) algorithms for multiple data sets: application to blind SIMO equalization. The 13th European Signal Processing Conference, 2005; S. 1-4.

[LLEL-2012] L. Gao; L. Qi; E. Chen; Ling Guan: Discriminative Multiple Canonical Correlation Analysis For Multi-Feature Information Fusion. The IEEE International Symposium on Multimedia, 2012; S. 36-43.

[L-1994] L. B. Almeida: The fractional Fourier Transform and time frequency representation. IEEE Trans. on Signal Processing, 42(11) 1994; S. 3084–3091.

[HD-1993] H. M. Ozatakas; D. Mendlovic: Fourier transforms of fractional order and their optical interpretation. Optics Communications, vol. 101, 1993; S. 163–

169.

[A-1993] A. W. Lohmann: Image rotation, Wigner rotation, and the fractional Fourier transform. Opt. Soc. Am. A, 10(10), 1993; S. 2181-2186.

[LLEXL-2010] L. Gao; L. Qi; E. Chen; X. Mu; Ling Guan: Recognizing Human Emotional State Based on the Phase Information of the Two Dimensional Fractional Fourier Transform. Springer, Lecture Notes in Computer Science, vol 6298, 2010; S. 694-704.

[CMH-2000] C. Candan; M. Alper Kutay; H. M. Ozaktas: I The Discrete Fractional Fourier Transform. IEEE Trans. on Signal Processing 48(5), 2000; S. 1329–1337.

[SH-1998] S.C. Pei; H. Yeh: Two dimensional discrete fractional Fourier transform.

Signal Processing, vol. 67, 1998; S. 99–108.

[TMH-2003] T. Melzer; M. Reiter; Horst Bischof: Appearance models based on kernel canonical correlation analysis. Pattern Recognition, vol.36, 2003; S. 1961- 1971.

[YL-2008] Y. Wang; L. Guan: Recognizing human emotion from audiovisual signals.

IEEE Trans. on Multimedia, vol. 10, no. 5, 2008; S. 936-946.

Referenzen

ÄHNLICHE DOKUMENTE

Given that naive discriminative learning provides a reasonable approximation to the processing costs of Dutch inflected words, correctly predicting a form (exemplar) frequency

Fur- thermore, the principal component representing local syntactic and morphological diversity accounted for the majority of the variability in the response latencies in the

For instance, in Estonian, the regular locative case forms are based on a stem that is identical to the corresponding genitive form, singular or plural.. 1 The rules here are simple

Accordingly, given the disparity between what the Rumelhart and McClelland model (and the many models of in fl ection based on the same conceptual analy- sis that have followed

As one of the linear dimension reduction methods, linear discriminative analysis (LDA) [12] aims to find a class-preserving projection matrix G which can separate different classes

For learning the unique features of the two domains, we utilize a unique extractor for event based on Spiking neural networks to extract edge cues in the event domain which may

To further enhance the robustness of the modified MLA in zoom level 0, a combination of three landmarks (both eyes and chin) is applied by the landmark combination procedure

The DGHT com- bines the Generalized Hough Transform (GHT) with a discriminative training proce- dure to generate GHT shape models with individual positive and negative model