• Keine Ergebnisse gefunden

2.3 Related Work

2.3.2 High-Dimensional Methods

2.3.2.5 Simple Twoing (si2)

This last section of the related work introduces the idea of Frank and Hall, 2001. We will call this idea simple twoing (si2) adapting terminology from Breiman et al., 1984. This section will lead into the presentation of hierarchical twoing which is an extension of this idea.

Their key idea was to translate an ordinalL-level classification problem into L−1 binary classification problems. The binarization is achieved by partitioning the L levels into two groups preserving the ordering. As an example the 5 groups from the rectal cancer exampleTRG0,TRG1,TRG2,TRG3, andTRG4 would be partitioned into the four partitions

• {T RG0}, {T RG1, T RG2, T RG3, T RG4},

• {T RG0, T RG1}, {T RG2, T RG3, T RG4},

• {T RG0, T RG1, T RG2},{T RG3, T RG4}, and

• {T RG0, T RG1, T RG2, T RG3}, {T RG4}.

To predict the class of an unseen samplex, theseL−1 binary classifiers are used to assign class probabilities to each of theL classes. The class with the maximum class probability is then chosen as the (predicted) class of the unseen sample x. The assignment of class probabilities proceeds via the following scheme:

• P(l = 1) = 1−P(class(x)>1)

• P(l =i) = P(class(x)> i−1)−P(class(x)> i), 1< i < L

• P(l =L) = P(class(x)> L−1)

 Methods

That means that the class probability of the classes 1 andLeach depend on one binary classifier and the class probabilities of the levels in between depend on the two ’neighbouring’ classifiers.

The presented idea does not depend on a specific binary base learner, but any binary classifier can be trained and used as base learner as long as it is able to assign class probabilities. In their paper Frank and Hall use classification trees (the C4.5 implementation) as their base learner.

Depending on the chosen base learner simple twoing can scale well to high-dimensional problems.

. Related Work 

Method Reference

ordinal

wrapper (high-dimensional or non high-dimensional)

si2 Frank and Hall, 2001

hi2 Leha, Jung, and Beißbarth, 2013

high-dimensional

rpartOrdinal Archer and Mas, 2009; Archer, 2010

rpartScore Galimberti, Soffritti, and Maso, 2012

glmnetcr Archer and Williams, 2012

SVOREX Chu and Keerthi, 2007

ordinal KDA Sun et al., 2010

kknn Hechenbichler and Schliep, 2006

PCDOC anchez-Monedero et al., 2013

non high-dimensional

cumulative link model Agresti, 2010

continuation ratio model Cox, 1988; Agresti, 2010

non ordinal

high-dimensional

SVM classification Bennett and Campbell, 2000

SVM regression Vapnik, 1995

classification trees Breiman et al., 1984; Quinlan, 1986

penalization methods

- regularization Friedman, Hastie, and Tibshirani, 2010

- boosting uhlmann and Hothorn, 2007a

k nearest neighbours (kNN) Fix and Hodges, 1951; Cover and Hart, 1967

non high-dimensional

LDA, QDA Fisher, 1936

Hastie, Tibshirani, and Friedman, 2009

multinomial logistic regression McFadden, 1973; Agresti, 2002

(linear) regression Hastie, Tibshirani, and Friedman, 2009 TABLE 2.2 Systematic view on classification algorithms and their relation to ordinal response and high dimensional feature space. In this table, we distinguish at the top level specifically ordinal methods from more general methods. In each group we further split the methods into methods that can handle data withpn. The exception here are the wrapper methods si2 and hi2, as their ability to classify high-dimensional data depends on the chosen base learner.

This list only shows the most prominent methods, except for the ordinal and high-dimensional methods where the list is more detailed, as that class of methods is the main focus of this work.

 Methods

3 Material

In the previous chapter we discussed how to build and evaluate classifiers generally in the case of high-dimensional data when the response is on an ordinal scale. We also looked at existing methods to do such classification.

Before we introduce our own method hi2 and actually compare it with existing methods in chapter 4 we will introduce the data that we will use to do that comparison.

Both, simulated and real data, will be used and we will in this chapter first introduce the real data and present the simulation study in the second part.

3.1 Real Data

Data with ordinal response is surprisingly common. The reason for ordinal quantities usually lies in human scoring or grading (as the tumor regression grade TRG). But also composite measures often lead to ordinal quantities. An example is the AJCC stage which is essentially a composition of the TNM stage (Edge et al., 2011).

Four datasets are included in the classifier comparison:

• Treatment Response of Rectal Cancer Patients

This dataset has been introduced already in the introduction as it serves as the example dataset throughout this thesis. This dataset stems from a research group that the author was part of.

• Gene Expression in Neuroblastoma

This dataset was chosen as it is very similar in nature to the rectal cancer data but is quite well known in the scientific community.

 Material

• miRNA Expression in Breast Cancer

From this study not only mRNA expression but also miRNA expression data is publicly available. Here, we focus on the miRNA expression as an example of other sources for high dimensional data.

• Gene Expression in B-cell Acute Lymphocytic Leukemia

This data is available to R users (pre-packaged through bioconductor, Gentleman et al., 2005), which is why that dataset is extremely well studied. It is known that this data contain a lot of information and are, thus, ’easy’ to classify.

The evaluation of all real datasets was done via a 10-fold cross validation, where the samples were randomly assigned to one of 10 partitions.