• Keine Ergebnisse gefunden

Phylogenetic trees III Maximum Parsimony

N/A
N/A
Protected

Academic year: 2022

Aktie "Phylogenetic trees III Maximum Parsimony"

Copied!
54
0
0

Wird geladen.... (Jetzt Volltext ansehen)

Volltext

(1)

Phylogenetic trees III Maximum Parsimony

Gerhard Jäger

Words, Bones, Genes, Tools February 28, 2018

(2)

Background

(3)

Character-based tree estimation

distance-based tree estimation has several drawbacks:

very strong theoretical assumptions - e.g., all characters evolve at the same rate

Neighbor Joining and UPGMA produce good but sub-optimal trees no solid statistical justification for those algorithms

distances are black boxes — we get a tree, but we learn nothing about the history of individual characters

character-based tree estimation

estimates complete scenario (or distribution over scenarios) for each character

finds the tree that best explains the observed variation in the data (at least in theory, that is...)

(4)

Parsimony

(5)

Parsimony of a tree

background reading: Ewens and Grant (2005), 15.6 suppose a character matrix and a tree are given

parsimony score: minimal number of mutations that has to be assumed to explain the character values at the tips, given the tree

(6)

Parsimony of a tree

"head"Kopf kop

"head" head

"head" tête

"head" testa

"head" cap

"head"

(7)

Parsimony of a tree

"head"Kopf kop

"head" head

"head" tête

"head" testa

"head" cap

"head"

(8)

Parsimony of a tree

"head"Kopf kop

"head" head

"head" tête

"head" testa

"head" cap

"head"

"head"

?

?

?

?

?

(9)

Parsimony of a tree

"head"Kopf kop

"head" head

"head" tête

"head" testa

"head" cap

"head"

"head"*kop testa

"head"

(10)

Parsimony of a tree

"head"Kopf kop

"head" head

"head" tête

"head" testa

"head" cap

"head"

"head"*kop

*haubud-

"head"

testa

"head"

caput

"head"

*kaput-

"head"

(11)

Parsimony reconstruction

A C C

B

A B B

A B B

C

Parsimony = 2

(12)

Parsimony reconstruction

A C C

A B B

A

B

C

Parsimony = 3 A

A

(13)

Parsimony reconstruction

A C C

A B B

A C

Parsimony = 3

A

C

C

(14)

Weighted parsimony reconstruction

A C C

B

A B B

A B B

C Weighted

Parsimony = 3 Weight matrix

A B C

A 0 1 2

B 1 0 2

C 2 2 0

(15)

Weighted parsimony reconstruction

A C C

A B B

A

B

C A

A

Weighted

Parsimony = 4 Weight matrix

A B C

A 0 1 2

B 1 0 2

C 2 2 0

(16)

Weighted parsimony reconstruction

A C C

A B B

A C

Weighted Parsimony = 5

A

C

C

Weight matrix

A B C

A 0 1 2

B 1 0 2

C 2 2 0

(17)

Dynamic Programming (Sankoff Algorithm)

wp(mother, s) = X

d∈daughters

min

s0∈states(w(s, s0) +wp(d, s0))

A C C

A B B

0∞ ∞ 0∞ ∞ 0 ∞ ∞ 0 0 0

(18)

Dynamic Programming (Sankoff Algorithm)

wp(mother, s) = X

d∈daughters

min

s0∈states(w(s, s0) +wp(d, s0))

A C C

A B B

0∞ ∞ 0∞ ∞ 0 ∞ ∞ 0 0 0

0 2 4 4 4 0

(19)

Dynamic Programming (Sankoff Algorithm)

wp(mother, s) = X

d∈daughters

min

s0∈states(w(s, s0) +wp(d, s0))

A C C

A B B

0∞ ∞ 0∞ ∞ 0 ∞ ∞ 0 0 0 0

1 3

2 4 1 4

4 4 0 2 2

(20)

Dynamic Programming (Sankoff Algorithm)

wp(mother, s) = X

d∈daughters

min

s0∈states(w(s, s0) +wp(d, s0))

A C C

A B B

0∞ ∞ 0∞ ∞ 0 ∞ ∞ 0 0 0 0

1 3

4

2 4 1 4

4 4 0 2 2 3 5

(21)

Dynamic Programming (Sankoff Algorithm)

wp(mother, s) = X

d∈daughters

min

s0∈states(w(s, s0) +wp(d, s0))

A C C

A B B

0∞ ∞ 0∞ ∞ 0 ∞ ∞ 0 0 0 0

1 3

4

2 4 1 4

4 4 0 2 2 3 5

(22)

Dynamic Programming (Sankoff Algorithm)

wp(mother, s) = X

d∈daughters

min

s0∈states(w(s, s0) +wp(d, s0))

A C C

A B B

0∞ ∞ 0∞ ∞ 0 ∞ ∞ 0 0 0 0

1 3

4

2 4 1 4

4 4 0 2 2 3 5

(23)

Dynamic Programming (Sankoff Algorithm)

wp(mother, s) = X

d∈daughters

min

s0∈states(w(s, s0) +wp(d, s0))

A C C

A B B

0∞ ∞ 0∞ ∞ 0 ∞ ∞ 0 0 0 0

1 3

4

2 4 1 4

4 4 0 2 2 3 5

(24)

Dynamic Programming (Sankoff Algorithm)

wp(mother, s) = X

d∈daughters

min

s0∈states(w(s, s0) +wp(d, s0))

A C C

A B B

0∞ ∞ 0∞ ∞ 0 ∞ ∞ 0 0 0 0

1 3

4

2 4 1 4

4 4 0 2 2 3 5

(25)

Dynamic Programming (Sankoff Algorithm)

wp(mother, s) = X

d∈daughters

min

s0∈states(w(s, s0) +wp(d, s0))

A C C

A B B

0∞ ∞ 0∞ ∞ 0 ∞ ∞ 0 0 0 0

1 3

4

2 4 1 4

4 4 0 2 2 3 5

(26)

Searching for the best tree

total parsimony score of tree: sum over all characters

note: if weight matrix is symmetric, location of the root doesn’t matter

Sankoff algorithm efficiently computes parsimony score of a given tree goal: tree which minimizes parsimony score

no efficient way to find the optimal tree→ heuristic tree search

(27)

Searching the tree space

(28)

How many rooted tree topologies are there?

2 1 n=2

(29)

How many rooted tree topologies are there?

2 3

1 1 2 3

2 3 1

2 1 n=2

n=3

2 3

1

(30)

How many rooted tree topologies are there?

2 3

1

2 4 3 2 1 4 3 2 1 4 3 2 1 3 4 1

3 2 1 2

3 1

2 1 n=2

n=3

n=4

2 3

1

2 4 3 1

(31)

How many rooted tree topologies are there?

f(2) = 1

f(n+ 1) = (2n−3)f(n) f(n) = (2n−3)!

2n−2(n−2)!

2 1

3 3

4 15

5 105

6 945

7 10395

8 135135

9 2027025

10 34459425 11 654729075 12 13749310575 13 316234143225 14 7.9e+ 12 15 2.1e+ 14 16 6.1e+ 15 17 1.9e+ 17 18 6.3e+ 18 19 2.2e+ 20 20 8.2e+ 21 21 3.1e+ 23 22 1.3e+ 25 23 5.6e+ 26 24 2.5e+ 28 25 1.1e+ 30 26 5.8e+ 31 27 2.9e+ 33 28 1.5e+ 35 29 8.6e+ 36 30 4.9e+ 38 31 2.9e+ 40 32 1.7e+ 42 33 1.1e+ 44 34 7.2e+ 45 35 4.8e+ 47 36 3.3e+ 49 37 2.3e+ 51 38 1.7e+ 53 39 1.3e+ 55 40 1.0e+ 57

(32)

How many unrooted tree topologies are there?

2 31

n=3

(33)

How many unrooted tree topologies are there?

2 3 4

1

3 2 4

1

4 2 3

1

2 31

n=3 n=4

(34)

How many unrooted tree topologies are there?

2 5 43

1

2 3 45

1

2 4 53

1

25 34

1 5 2 34

1 3 5 24

1

24 35

1

3 2 54

1

3 4 25

1 5 3 42

1 4 5 23

1

4 2 35

1

4 3 52

1

45 23

1

5 4 23

1

2 3 4

1

3 2 4

1

4 2 3

1

2 31

n=3 n=4 n=5

(35)

How many unrooted tree topologies are there?

f(3) = 1

f(n+ 1) = (2n−3)f(n) f(n) = (2n−5)!

2n−3(n−3)!

3 1

4 3

5 15

6 105

7 945

8 10395

9 135135

10 2027025

11 34459425 12 654729075 13 13749310575 14 316234143225 15 7.90e+ 12 16 2.13e+ 14 17 6.19e+ 15 18 1.91e+ 17 19 6.33e+ 18 20 2.21e+ 20 21 8.20e+ 21 22 3.19e+ 23 23 1.31e+ 25 24 5.63e+ 26 25 2.53e+ 28 26 1.19e+ 30 27 5.84e+ 31 28 2.98e+ 33 29 1.57e+ 35 30 8.68e+ 36 31 4.95e+ 38 32 2.92e+ 40 33 1.78e+ 42 34 1.12e+ 44 35 7.29e+ 45 36 4.88e+ 47 37 3.37e+ 49 38 2.39e+ 51 39 1.74e+ 53 40 1.31e+ 55

(36)

Heuristic tree search

tree space is too large to do an exhaustive search ifn(number of taxa) is larger than 12 or so

heuristic search:

start with some tree topology (e.g., Neighbor-Joining tree) apply a bunch of local modifications to the current tree

if one of the modified tree has lower or equal parsimony, move to that tree

stop if no further improvement is possible

⇒ standard approach for optimization problems in computer science

(37)

Tree modifications

three tree modifications commonly in use:

1 Nearest Neighbor Interchange (NNI)

2 Tree Bisection and Reconnection(TBR)

3 Subtree Pruning and Regrafting (SPR)

local modifications are better than arbitrary moves in tree space because partial parsimony computations can be re-used in modified tree

(38)

Nearest Neighbor Interchange

(39)

Tree Bisection and Reconection

1 2 34

56 78

9 10 1 2 34

5 67

8 10

9 1 2 34

5 6

7 8

10

9

(40)

Subtree Pruning and Regrafting

1 2 34

5 6

7

8 10

9

1 2 34

56 78

9 10 1 2 34

5 6

7

8 10

9

(41)

Heuristic tree search

NNI is very local→ only O(n) possible moves

SPR and TBR are more aggressive→ O(n2)/O(n3) possible moves NNI search is comparatively fast, but prone to get stuck in local optima

(42)

Running example: SPR search with cognate data

parsimony=1984

Spanish Portuguese Hindi

Nepali Bengali Greek

Breton Welsh

Irish

Dutch German

English Sw

edish Danish

Icelandic Polish

Czech Russian Ukrainian

Bulgar ian

Lithuanian

French

Italian Romanian

Catalan

starting with Neighbor Joining tree . . .

(43)

Running example: SPR search with cognate data

parsimony=1979

Spanish Portuguese

Hindi Nepali Bengali Greek

Breton

Welsh Irish

Dutch German English

Sw edish Danish

Icelandic

Polish Czech Russian Ukr

ainian Bulgar Lithuanian ian

French Italian

Romanian Catalan

(44)

Running example: SPR search with cognate data

parsimony=1975

Spanish

Portuguese

Hindi Nepali Bengali Greek

Breton Welsh

Irish

Dutch German English

Sw edish Danish

Icelandic

Polish Czech Russian

Ukrainian Bulgar Lithuanian ian

French Italian Romanian

Catalan

(45)

Running example: SPR search with cognate data

parsimony=1973

Spanish Portuguese

Hindi Nepali Bengali Greek

Breton

Welsh Irish

Dutch Ger

man English

Sw edish

Danish Icelandic

Polish Czech Russian Ukr

ainian Bulgar Lithuanian ian

French Italian

Romanian Catalan

(46)

Running example: SPR search with cognate data

parsimony=1969

Spanish

Portuguese

Hindi Nepali Bengali

Greek Breton Welsh

Irish

Dutch Ger

man English

Sw edish

Danish Icelandic

Polish Czech Russian Ukr

ainian Bulgar Lithuanian ian

French Italian

Romanian Catalan

. . . Maximum Parsimony tree

(47)

Running example: SPR search with cognate data

there are actually 16 different trees with minimal parsimony score

Greek Irish

Welsh Breton

French Italian Catalan

Portuguese

Spanish

Romanian

Icelandic

Swedish

Danish

German

Dutch English

Hindi

Nepali Bengali

Ukrainian Russian Polish Czech Bulgarian

Lithuanian

(48)

MP tree for WALS characters

Bengali Breton

Irish Welsh Bulgarian

Greek Czech

Lithuanian

Polish

Russian

Ukrainian

Catalan

Romanian

Italian

Spanish

Portuguese

French Danish Swedish Icelandic Dutch German English

Hindi Nepali

(49)

MP tree for sound-concept characters

Greek Bulga

rian Russia

n Ukra inian

Polish Czech

Lithu

anian

Icela ndi

c

Swedish

Dani sh

Dutch

German

English

Catalan

Portug uese

Span ish

Ital ian Roman

ian French Breton Welsh Irish

Hindi Bengal Nepalii

(50)

Dollo parsimony

previous trees were estimated with a symmetric weight matrix if weights are asymmetric, location of the root matters extreme case: Dollo Parsimony

w(0→1) =∞ SpanishPortuguese

Hindi Nepali Bengali Greek

Breton Welsh Irish Dutch German

English Swedish

Danish Icelandic

Polish Czech Russian Ukrainian

Bulgarian Lithuanian

French Italian

Romanian

Catalan

(51)

Maximum Parsimony: Discussion

Once we have found the best tree (or, in any event, which is very close to the best tree), we can reconstruct ancestral states via the Sankoff algorithm

this allows to compute statistics about stability of characters, frequency and location of parallel changes etc.

much more informative than distance-based inference

(52)

Maximum Parsimony: Discussion

disadvantages of MP:

simulation studies: capacity to recover the true tree is decent but not overwhelming

possibility of multiple mutations on a single branch is not taken into consideration

all characters are treated equal; no discrimination between stable and volatile characters

ties are common, especially if you have few data values for weight matrix aread hoc

no real theoretical justification

Why should the true tree minimize the total number of mutations?

Rests on a valid intuition: Mutations are unlikely, so assuming fewer mutations increases the likelihood of the data.

Likelihood is not formally derived from a probabilistic modell though.

Next step: Maximum Likelihood tree estimation

(53)

Hands on

Install the software Paup*.

Go to the directory where you have the put the nexus files and type

> paup4 ielex.bin.nex

At Paup’s command prompt, type paup> hsearch.

Display tree with

paup> describetree /plot=phylo Save result with

paup> savetree format=newick file = ielex.mp.tre \ brlen=yes

Leave Paup* with paup> q

InstallDendroscope or FigTree and loadielex.mp.tre.

(54)

Ewens, W. and G. Grant (2005). Statistical Methods in Bioinformatics:

An Introduction. Springer, New York.

Referenzen

ÄHNLICHE DOKUMENTE

(single-label tree with additional node info, e.g. node type) Elements, attributes, and text values are nodes. DOM parsers load XML into main memory random access by traversing

insert new node v as k-th child of p Insert and delete are inverse edit operations (i.e., insert undoes delete and vice versa).

This directly implies two important features for mathemat- ical modeling: Firstly, it is a metric space containing all possible phylogenetic trees on n leaves, i.e., it is a

This improved predictive power especially for the aminoglycoside spectinomycin (blue point), where some of our candidate predictors had no effect on MIC (see Figure S6 and

In contrast to previous works that use smooth generalized cylinders to represent tree branches, our method generates realistic looking tree models with complex branching geometry

The R-tree [3] and the R*-tree [4], spatial access methods with a hierarchically structured directory that use minimum bounding rectangles (MBRs) as page regions, have primarily

For more realistic durations of cross-protection and larger values of ρ, cross-immunity alone cannot prevent branching events and suppression of the second mutant occurs only if

We now proceed to study the expressiveness, stability, and run-time performance of approximate tree kernels in real-world applications, namely supervised learning tasks dealing