Parsimony Likelihood Common Causes and Phylogenetic Inference Elliott

  • Slides: 62
Download presentation
Parsimony, Likelihood, Common Causes, and Phylogenetic Inference Elliott Sober Philosophy Department University of Wisconsin,

Parsimony, Likelihood, Common Causes, and Phylogenetic Inference Elliott Sober Philosophy Department University of Wisconsin, Madison 1

2 suggested uses of O’s razor • O’s razor should be used to constrain

2 suggested uses of O’s razor • O’s razor should be used to constrain the order in which hypotheses are to be tested. • O’s razor should be used to interpret the acceptability/support of hypotheses that have already been tested. 2

Pluralism about Ockham’s razor? • [Pre-test] O’s razor should be used to constrain the

Pluralism about Ockham’s razor? • [Pre-test] O’s razor should be used to constrain the order in which hypotheses are to be tested. • [Post-test] O’s razor should be used to interpret the acceptability/support of hypotheses that have already been tested. 3

these can be compatible, but… • [Pre-test] O’s razor should be used to constrain

these can be compatible, but… • [Pre-test] O’s razor should be used to constrain the order in which hypotheses are to be tested. • [Post-test] O’s razor should be used to interpret the acceptability/support of hypotheses that have already been tested. If pre-test O’s razor is “rejectionist, ” then post-test O’s razor won’t have a point. 4

these can be compatible, but… • [Pre-test] O’s razor should be used to constrain

these can be compatible, but… • [Pre-test] O’s razor should be used to constrain the order in which hypotheses are to be tested. • [Post-test] O’s razor should be used to interpret the acceptability/support of hypotheses that have already been tested. If the pre-test idea involves testing hypotheses one at time, then it views testing as noncontrastive. 5

within the post-test category of support/plausibility. . . • Bayesianism – compute posterior probs.

within the post-test category of support/plausibility. . . • Bayesianism – compute posterior probs. • Likelihoodism – compare likelihoods. • Frequentist model selection criteria like AIC – estimate predictive accuracy. 6

I am a pluralist about these broad philosophies … • Bayesianism – compute posterior

I am a pluralist about these broad philosophies … • Bayesianism – compute posterior probs • Likelihoodism – compare likelihoods • Frequentist model selection criteria like AIC – estimate predictive accuracy. 7

I am a pluralist about these broad philosophies … • Bayesianism – compute posterior

I am a pluralist about these broad philosophies … • Bayesianism – compute posterior probs • Likelihoodism – compare likelihoods • Frequentist model selection criteria like AIC – estimate predictive accuracy. Not that each is okay as a global thesis about all scientific inference … 8

I am a pluralist about these broad philosophies … • Bayesianism – compute posterior

I am a pluralist about these broad philosophies … • Bayesianism – compute posterior probs • Likelihoodism – compare likelihoods • Frequentist model selection criteria like AIC – estimate predictive accuracy. But I do think that each has its place. 9

Ockham’s Razors* Different uses of O’s razor have different justifications and some have none

Ockham’s Razors* Different uses of O’s razor have different justifications and some have none at all. * “Let’s Razor Ockham’s Razor, ” in D. Knowles (ed. ), Explanation and Its Limits, Cambridge University Press, 1990, 73 -94. 10

Parsimony and Likelihood In model selection criteria like AIC and BIC, likelihood and parsimony

Parsimony and Likelihood In model selection criteria like AIC and BIC, likelihood and parsimony are conflicting desiderata. AIC(M) = log[Pr(Data│L(M)] - k 11

Parsimony and Likelihood In model selection criteria like AIC and BIC, likelihood and parsimony

Parsimony and Likelihood In model selection criteria like AIC and BIC, likelihood and parsimony are conflicting desiderata. In other settings, parsimony has a likelihood justification. 12

the Law of Likelihood Observation O favors H 1 over H 2 iff Pr(O│H

the Law of Likelihood Observation O favors H 1 over H 2 iff Pr(O│H 1) > Pr(O│H 2) 13

a Reichenbachian idea Salmon’s example of plagiarism E 1 E 2 C [Common Cause]

a Reichenbachian idea Salmon’s example of plagiarism E 1 E 2 C [Common Cause] E 1 E 2 C 1 C 2 [Separate Causes] 14

a Reichenbachian idea Salmon’s example of plagiarism E 1 E 2 C [Common Cause]

a Reichenbachian idea Salmon’s example of plagiarism E 1 E 2 C [Common Cause] E 1 E 2 C 1 C 2 [Separate Causes] 15 more parsimonious

Reichenbach’s argument IF (i) (iii) (iv) A cause screens-off its effects from each other

Reichenbach’s argument IF (i) (iii) (iv) A cause screens-off its effects from each other All probabilities are non-extreme (≠ 0, 1) a particular parameterization of the CC and SC models cause/effect relationships are “homogenous” across branches. THEN Pr[Data │Common Cause] > Pr[Data │Separate Causes]. 16

parameters and homogeneity E 1 E 2 p 1 p 2 C [Common Cause]

parameters and homogeneity E 1 E 2 p 1 p 2 C [Common Cause] E 1 p 1 C 1 E 2 p 2 C 2 [Separate Causes] 17

Reichenbach’s argument IF (i) (iii) (iv) A cause screens-off its effects from each other.

Reichenbach’s argument IF (i) (iii) (iv) A cause screens-off its effects from each other. All probabilities are non-extreme parameterization of the CC and SC models. cause/effect relationships are “homogenous” across branches. THEN Pr[Data │Common Cause] > Pr[Data │Separate Causes]. The more parsimonious hypothesis has the higher likelihood. 18

Reichenbach’s argument IF (i) (iii) (iv) A cause screens-off its effects from each other.

Reichenbach’s argument IF (i) (iii) (iv) A cause screens-off its effects from each other. All probabilities are non-extreme parameterization of the CC and SC models. cause/effect relationships are “homogenous” across branches. THEN Pr[Data │Common Cause] > Pr[Data │Separate Causes]. Parsimony and likelihood are ordinally equivalent. 19

Some differences with Reichenbach • I am comparing two hypotheses. • I’m not using

Some differences with Reichenbach • I am comparing two hypotheses. • I’m not using R’s Principle of the Common Cause. • I take the evidence to be the matching of the students’ papers, not their “correlation. ” 20

empirical foundations for likelihood ≈ parsimony (i) (iii) (iv) A cause screens-off its effects

empirical foundations for likelihood ≈ parsimony (i) (iii) (iv) A cause screens-off its effects from each other. All probabilities are non-extreme parameterization of the CC and SC models. cause/effect relationships are “homogenous” across branches. By adopting different assumptions, you can arrange for CC to be less likely than SC. Now likelihood and parsimony conflict! 21

empirical foundations for likelihood ≈ parsimony (i) (iii) (iv) A cause screens-off its effects

empirical foundations for likelihood ≈ parsimony (i) (iii) (iv) A cause screens-off its effects from each other. All probabilities are non-extreme parameterization of the CC and SC models. cause/effect relationships are “homogenous” across branches. Note: the R argument shows that these are sufficient for likelihood ≈ parsimony, not that they are necessary. 22

Parsimony in Phylogenetic Inference Two sources: ─ Willi Hennig ─ Luigi Cavalli-Sforza and Anthony

Parsimony in Phylogenetic Inference Two sources: ─ Willi Hennig ─ Luigi Cavalli-Sforza and Anthony Edwards Two types of inference problem: ─ find the best tree “topology” ─ estimate character states of ancestors 23

1. Which tree topology is better? H C (HC)G G H C G H(CG)

1. Which tree topology is better? H C (HC)G G H C G H(CG) MP: (HC)G is better supported than H(CG) by data D if and only if (HC)G is a more parsimonious explanation of D than H(CG) is. 24

An Example of a Parsimony Calculation 1 H 1 C 0 G 1 H

An Example of a Parsimony Calculation 1 H 1 C 0 G 1 H 1 C 0 0 (HC)G H(CG) 0 G 25

2. What is the best estimate of the character states of ancestors in an

2. What is the best estimate of the character states of ancestors in an assumed tree? 1 H 1 C 1 G A=? 26

2. What is the best estimate of the character states of ancestors in an

2. What is the best estimate of the character states of ancestors in an assumed tree? 1 H 1 C 1 G A=? MP says that the best estimate is that A=1. 27

Maximum Likelihood H C (HC)G G H C G H(CG) ML: (HC)G is better

Maximum Likelihood H C (HC)G G H C G H(CG) ML: (HC)G is better supported than H(CG) by data D if and only if Pr[D│(HC)G] > Pr[D│H(CG)]. 28

Maximum Likelihood H C (HC)G G H C G H(CG) ML: (HC)G is better

Maximum Likelihood H C (HC)G G H C G H(CG) ML: (HC)G is better supported than H(CG) by data D if and only if Pr. M[D│(HC)G] > Pr. M[D│H(CG)]. ML is “model dependent. ” 29

the present situation in evolutionary biology • MP and ML sometimes disagree. • The

the present situation in evolutionary biology • MP and ML sometimes disagree. • The standard criticism of MP is that it assumes that evolution proceeds parsimoniously. • The standard criticism of ML is that you need to choose a model of the evolutionary process. 30

When do parsimony and likelihood agree? • (Ordinal Equivalence) For any data set D

When do parsimony and likelihood agree? • (Ordinal Equivalence) For any data set D and any pair of phylogenetic hypotheses H 1 and H 2, Pr(D│H 1) > Pr(D│H 2) iff H 1 is a more parsimonious explanation of D than H 2 is. 31

When do parsimony and likelihood agree? • (Ordinal Equivalence) For any data set D

When do parsimony and likelihood agree? • (Ordinal Equivalence) For any data set D and any pair of phylogenetic hypotheses H 1 and H 2, Pr. M(D│H 1) > Pr. M(D│H 2) iff H 1 is a more parsimonious explanation of D than H 2 is. • Whether likelihood agrees with parsimony depends on the probabilistic model of evolution used. 32

When do parsimony and likelihood agree? • (Ordinal Equivalence) For any data set D

When do parsimony and likelihood agree? • (Ordinal Equivalence) For any data set D and any pair of phylogenetic hypotheses H 1 and H 2, Pr. M(D│H 1) > Pr. M(D│H 2) iff H 1 is a more parsimonious explanation of D than H 2 is. • Whether likelihood agrees with parsimony depends on the probabilistic model of evolution used. • Felsenstein (1973) showed that the postulate of very low rates of evolution suffices for ordinal equivalence. 33

Does this mean that parsimony assumes that rates are low? • NO: the assumptions

Does this mean that parsimony assumes that rates are low? • NO: the assumptions of a method are the propositions that must be true if the method correctly judges support. 34

Does this mean that parsimony assumes that rates are low? • NO: the assumptions

Does this mean that parsimony assumes that rates are low? • NO: the assumptions of a method are the propositions that must be true if the method correctly judges support. • Felsenstein showed that the postulate of low rates suffices for ordinal equivalence, not that it is necessary for ordinal equivalence. 35

Tuffley and Steel (1997) • T&S showed that the postulate of “no-commonmechanism” also suffices

Tuffley and Steel (1997) • T&S showed that the postulate of “no-commonmechanism” also suffices for ordinal equivalence. • “no-common-mechanism” means that each character on each branch is subject to its own drift process. 36

the two probability models of evolution Felsenstein Tuffley and Steel • Rates of change

the two probability models of evolution Felsenstein Tuffley and Steel • Rates of change are low, but not necessarily equal. • Rates of change can be high. • Drift not assumed: Pr(i j) and Pr(j i) may differ. • Drift is assumed: Pr(i j) = Pr(j i) 37

How to use likelihood to define what it means for parsimony to assume something

How to use likelihood to define what it means for parsimony to assume something • The assumptions of parsimony = the propositions that must be true if parsimony correctly judges support. • For a likelihoodist, parsimony correctly judges support if and only if parsimony is ordinally equivalent with likelihood. • Hence, for a likelihoodist, parsimony assumes any proposition that follows from ordinal equivalence. 38

A Test for what Parsimony does not assume Model M ordinal equivalence A where

A Test for what Parsimony does not assume Model M ordinal equivalence A where A = what parsimony assumes 39

A Test for what Parsimony does not assume Model M ordinal equivalence A where

A Test for what Parsimony does not assume Model M ordinal equivalence A where A = what parsimony assumes • If model M entails ordinal equivalence, and M entails proposition X, X may or may not be an assumption of parsimony. 40

A Test for what Parsimony does not assume Model M ordinal equivalence A where

A Test for what Parsimony does not assume Model M ordinal equivalence A where A = what parsimony assumes • If model M entails ordinal equivalence, and M entails proposition X, X may or may not be an assumption of parsimony. • If model M entails ordinal equivalence, and M does not entail proposition X, then X is not an assumption of parsimony. 41

applications of the negative test • T&S’s model does not entail that rates of

applications of the negative test • T&S’s model does not entail that rates of change are low; hence parsimony does not assume that rates are low. • F’s model does not assume neutral evolution; hence parsimony does not assume neutrality. 42

How to figure out what parsimony does assume? • Find a model that forces

How to figure out what parsimony does assume? • Find a model that forces parsimony and likelihood to disagree about some example. • Then, if parsimony is right in what it says about the example, the model must be false. 43

Example #1 Task: Infer the character state of the MRCA of species that all

Example #1 Task: Infer the character state of the MRCA of species that all exhibit the same state of a quantitative character. 10 10 … 10 10 A=? The MP estimate is A=10. When is A=10 the ML estimate? And when is it not? 44

Answer 10 10 … 10 10 A=? ML says that A=10 is the best

Answer 10 10 … 10 10 A=? ML says that A=10 is the best estimate (and thus agrees with MP) if there is neutral evolution or selection is pushing each lineage towards a trait value of 10. 45

Answer 10 10 … 10 10 A=? ML says that A=10 is the best

Answer 10 10 … 10 10 A=? ML says that A=10 is the best estimate (and thus agrees with MP) if there is neutral evolution or selection is pushing each lineage towards a trait value of 10. ML says that A=10 is not the best estimate (and thus disagrees with MP) if (*) selection is pushing all lineages towards a single trait value different from 10. 46

Answer 10 10 … 10 10 A=? ML says that A=10 is the best

Answer 10 10 … 10 10 A=? ML says that A=10 is the best estimate (and thus agrees with MP) if there is neutral evolution or selection is pushing each lineage towards a trait value of 10. ML says that A=10 is not the best estimate (and thus disagrees with MP) if (*) selection is pushing all lineages towards a single trait value different from 10. So: Parsimony assumes, in this problem, that (*) is false. 47

Example #2 Task: Infer the character state of the MRCA of two species that

Example #2 Task: Infer the character state of the MRCA of two species that exhibit different states of a dichotomous character. 1 0 A=? A=0 and A=1 are equally parsimonious. When are they equally likely? And when are they unequally likely? 48

Answer 1 0 A=? ML agrees with MP that A=0 and A=1 are equally

Answer 1 0 A=? ML agrees with MP that A=0 and A=1 are equally good estimates if the same neutral process occurs in the two lineages. 49

Answer 1 0 A=? ML agrees with MP that A=0 and A=1 are equally

Answer 1 0 A=? ML agrees with MP that A=0 and A=1 are equally good estimates if the same neutral process occurs in the two lineages. ML disagrees with MP if (*) the same selection process occurs in both lineages. 50

Answer 1 0 A=? ML agrees with MP that A=0 and A=1 are equally

Answer 1 0 A=? ML agrees with MP that A=0 and A=1 are equally good estimates if the same neutral process occurs in the two lineages. ML disagrees with MP if (*) the same selection process occurs in both lineages. So: Parsimony assumes, in this problem, that (*) is false. 51

Conclusions about phylogenetic parsimony ≈ likelihood • The assumptions of parsimony are the propositions

Conclusions about phylogenetic parsimony ≈ likelihood • The assumptions of parsimony are the propositions that must be true if parsimony correctly judges support. 52

Conclusions about phylogenetic parsimony ≈ likelihood • The assumptions of parsimony are the propositions

Conclusions about phylogenetic parsimony ≈ likelihood • The assumptions of parsimony are the propositions that must be true if parsimony correctly judges support. • To find out what parsimony does not assume, use the test described [M ordinal equivalence A]. 53

Conclusions about phylogenetic parsimony ≈ likelihood • The assumptions of parsimony are the propositions

Conclusions about phylogenetic parsimony ≈ likelihood • The assumptions of parsimony are the propositions that must be true if parsimony correctly judges support. • To find out what parsimony does not assume, use the test described [M ordinal equivalence A]. • To find out what parsimony does assume, look for examples in which parsimony and likelihood disagree, not for models that ensure that they agree. 54

Conclusions about phylogenetic parsimony ≈ likelihood • The assumptions of parsimony are the propositions

Conclusions about phylogenetic parsimony ≈ likelihood • The assumptions of parsimony are the propositions that must be true if parsimony correctly judges support. • To find out what parsimony does not assume, use the test described [M ordinal equivalence A]. • To find out what parsimony does assume, look for examples in which parsimony and likelihood disagree, not for models that ensure that they agree. • Maybe parsimony’s assumptions vary from problem to problem. 55

broader conclusions • underdetermination: O’s razor often comes up when the data don’t settle

broader conclusions • underdetermination: O’s razor often comes up when the data don’t settle truth/falsehood or acceptance/rejection. 56

broader conclusions • underdetermination: O’s razor often comes up when the data don’t settle

broader conclusions • underdetermination: O’s razor often comes up when the data don’t settle truth/falsehood or acceptance/rejection. • reductionism: when O’s razor has authority, it does so because it reflects some other, more fundamental, desideratum. 57

broader conclusions • underdetermination: O’s razor often comes up when the data don’t settle

broader conclusions • underdetermination: O’s razor often comes up when the data don’t settle truth/falsehood or acceptance/rejection. • reductionism: when O’s razor has authority, it does so because it reflects some other, more fundamental, desideratum. [But there isn’t a single global justification. ] 58

broader conclusions • underdetermination: O’s razor often comes up when the data don’t settle

broader conclusions • underdetermination: O’s razor often comes up when the data don’t settle truth and falsehood. • reductionism: when O’s razor has authority, it does so because it reflects some other, more fundamental, desideratum. • two questions: When parsimony has a precise meaning, we can investigate: What are its presuppositions? What suffices to justify it? 59

A curiosity: in the R argument, to get a difference in likelihood, the hypotheses

A curiosity: in the R argument, to get a difference in likelihood, the hypotheses should not specify the states of the causes. E 1 E 2 p 1 p 2 C [Common Cause] E 1 p 1 C 1 E 2 p 2 C 2 [Separate Causes] 60

Example #0 Task: Infer the character state of the MRCA of species that all

Example #0 Task: Infer the character state of the MRCA of species that all exhibit the same state of a dichotomous character. 1 1 … 1 1 A=? The MP inference is that A=1. When is A=1 the ML inference? 61

Example #0 Task: Infer the character state of the MRCA of species that all

Example #0 Task: Infer the character state of the MRCA of species that all exhibit the same state of a dichotomous character. 1 1 … 1 1 A=? The MP inference is that A=1. When is A=1 the ML inference? Answer: when lineages have finite duration and the process is Markovian. It doesn’t matter whether selection or drift is the process at work. 62