Hostname: page-component-6bf8c574d5-xtvcr Total loading time: 0 Render date: 2025-02-20T22:59:43.791Z Has data issue: false hasContentIssue false

On the Evidential Import of Unification

Published online by Cambridge University Press:  01 January 2022

Rights & Permissions [Opens in a new window]

Abstract

There are two senses in which a hypothesis may be said to unify evidence: (1) ability to increase the mutual information of a set of evidence statements; (2) explanation of commonalities in phenomena by positing a common origin. On Bayesian updating, only Mutual Information Unification contributes to incremental support. Defenders of explanation as a confirmatory virtue that makes independent contribution must appeal to some relevant difference between humans and Bayesian agents. I argue that common origin unification has at best a limited heuristic role in confirmation. Finally, Reichenbachian common cause hypotheses are shown to be instances of Mutual Information Unification.

Type
Research Article
Copyright
Copyright © The Philosophy of Science Association

1. Introduction

Myrvold (Reference Myrvold2003) identified what was described therein as “one interesting sense” in which a theory can unify phenomena. This consists of the ability of the theory to render distinct phenomena informative (or more informative) about each other. Call this Mutual Information Unification (MIU). This sense lends itself nicely to a probabilistic explication, and it can be shown that unification in this sense contributes to incremental evidential support of the theory by the phenomena unified.

There is another sense of unification, having to do with hypotheses that posit a common origin for the phenomena in question, be it a common cause or some other type of explanation. Call this Common Origin Unification (COU). As emphasized by Lange (Reference Lange2004) and Schupbach (Reference Schupbach2005), the two senses are logically independent; neither is a necessary or a sufficient condition for the other, even though, in a number of interesting cases, they are concomitants of each other.

In this article, the respective roles of these two notions of unification in connection with the bearing of evidence on a theory are discussed. There are, of course, other questions one might ask, and other roles for a notion of unification to play apart from contributing to confirmation. Having a common explanation for disparate phenomena can contribute to deeper understanding, which is one goal of scientific research. Insofar as it contributes to such understanding, COU may play the role of a cognitive value.Footnote 1 As such, it can play a legitimate role in questions such as that of which research program to pursue; a theory might be regarded as more worthy of development on account of its potential for affording understanding.Footnote 2 This is a different matter from the question at issue in this article, which is whether unification ought to be regarded as contributing to the evidential support of a theory by the phenomena unified.

On the question of the respective roles of these two notions of unification in theory confirmation, on a Bayesian analysis, the answer is clear: MIU contributes to incremental evidential support, and there is no scope, within Bayesian updating, for COU to add to the evidential support of the theory (see sec. 4).

There may be some who do not take this to settle the normative issue and who will maintain that, despite the Bayesian verdict, we ought to take explanatory power of a hypothesis as a confirmatory virtue. Advocates of such a view would have to reject the idea that we should take consideration of a Bayesian agent updating via conditionalization as normative for those of us who are not such agents. This presents a challenge for such advocates. If it is rational, or reasonable, or otherwise well and good for us to do what is impossible for a Bayesian agent updating its credences via conditionalization, that is, to take COU to be something that makes a contribution to evidential support above and beyond what it contributes to MIU, then this must be grounded in some relevant difference between us and Bayesian agents. It is incumbent on an explanationist to give an account of what that difference is.

In what follows, these points are first illustrated by means of a simple example that, despite its artificiality, shares some salient features with cases of actual scientific interest. Next, in section 3, are presented the probabilistic measures of MIU introduced in Myrvold (Reference Myrvold2003), and in section 4 their impact on evidential support is exhibited. In section 5 are outlined possible reactions to the Bayesian verdict regarding the respective confirmatory roles of the two types of unification. In section 6 the question is addressed whether there is still a role for COU to play in hypothesis assessment, in assessing priors rather than in assessing incremental evidential support (the answer is no). Finally, in section 7 it is shown how Reichenbachian common causes fit into the schema of MIU.

2. Two Kinds of Unification

Consider the following toy example, of no use except for introducing the issues at hand, although it does share some salient features with a multitude of real-world cases of genuine scientific interest. You are about to be presented with two data streams, S 1 and S 2, each of which will be sequences of 10 heads or tails. You know that they have been produced by coin flipping, but you are not sure of exactly the procedure used or whether the coin or coins involved are fair.

Suppose that you have nonzero credences in both of the following hypotheses:

H 1: A fair coin was flipped 10 times, and the results of this series of coin flips are reported in both data streams.

H 2: Two fair coins were flipped 10 times each, and each data stream reports the results of one of these series of coin flips.

I invite you to consider the effect of the evidence on these two hypotheses. That evidence consists of specification of the two data streams:

S 1: HHHTTHTHHT

S 2: HHHTTHTHHT

Let E 1 be the proposition that S 1 is the string given above, and E 2 the corresponding proposition about S 2.

Now, if you have nonnegligible prior credence that the strings might have been produced by radically unfair coins, E 1 and E 2 might boost your confidence in the fairness of the coins, and hence conditionalizing on each of E 1 and E 2, separately, might boost your credence in both H 1 and H 2. But, when taken together, E 1 and E 2 strongly favor H 1 over H 2.

There are two features of this example that I would like to draw your attention to. The first feature is that H 1, if true, renders E 1 informative about what data stream S 2 will be. Conditional on H 1, knowing E 1 permits one to anticipate the truth of E 2. That is, H 1 exhibits MIU with respect to the evidence set {E1,E2}. A hypothesis has this property, with respect to a set of evidential propositions, if conditionalizing on that hypothesis increases the mutual informativeness of the set. Obviously, this is the sort of thing that comes in degrees. In our toy example, conditional on H 1, knowledge of E 1 permits one to anticipate, with certainty, all details of E 2. In more interesting cases the increase of informativeness will be less than maximal. Probabilistic measures of degree of this sort of unification will be introduced below.

The second feature is that H 1 posits a common origin of the two data streams and thus is ripe to be the subject of what Janssen (Reference Janssen2002) has called a COI story, for Common Origin Inference. In addition to MIU, H 1 exhibits COU.

The two concepts are of a manifestly very different sort. One belongs to a cluster of concepts involving information, states of knowledge, and the like; the other is related to concepts of cause and explanation.Footnote 3 As already mentioned, they are logically independent. A hypothesis can posit a common origin for two (or more) evidential propositions without making them mutually informative about each other, as the propositions could be about independent aspects of their posited common origin; thus, we can have COU without MIU. Furthermore, once two or more evidential propositions are known, that is, have been absorbed into one’s background knowledge, they are no longer informative about each other, although any common origin they might have remains, and again we have COU without MIU. One can also trivially construct hypotheses that exhibit MIU without COU. With respect to our toy example, consider the hypothesis

H 3: Two fair coins were flipped 10 times each, each data stream reports the results of one of these series of coin flips, and the results of each series of flips just happened to be the same.

This hypothesis, if true, also renders one data stream informative about the other. Of course, before the evidence, one would expect credence in H 3 to be low, lower than credence in H 2 by a factor of 1,024.

Although artificial, our toy example has a multitude of parallels in actual science. One is the case of heliocentric v geocentric world systems, discussed by Janssen (Reference Janssen2002), Myrvold (Reference Myrvold2003), and Henderson (Reference Henderson2014). The analog of H 1 is what was called hC in Myrvold (Reference Myrvold2003), that is, the heliocentric hypothesis that all planets have circular or nearly circular orbits centered at or near the sun, and the analog of H 2 is the bare-bones geocentric hypothesis hP, which posits that, for each planet, there is a deferent circle centered near the earth and that the planet travels on an epicycle whose center travels on the deferent, with no assumption made about any connections between the motions of different planets or between planetary motions and the motion of the sun. The analog of H 3 is the geocentric hypothesis conjoined with the sun-planet parallelism condition; this is the hypothesis h SP, or the strengthened Ptolemaic hypothesis.

One can find analogs in cases in which a hypothesis turns disparate, prima facie unrelated phenomena into agreeing measurements of some theoretical parameter. The classic case is Perrin’s argument for the existence of atoms. Perrin (Reference Perrin1913, sec. 119; Reference Perrin and Hammick1916, sec. 120) adduces 13 distinct phenomena that, on the atomic hypothesis, count as measurements of Avogadro’s number. The analog of H 1 is that atoms exist, and hence there is a common origin explanation of the agreement of these measurements; the analog of H 2 would be the hypothesis that matter is continuously divisible, and the analog of H 3 would be the hypothesis that adds to H 2 the stipulation that Perrin’s 13 phenomena yield values that just happen to agree within experimental error, even though they are not agreeing measurements of any physically meaningful quantity. Another example is the quantum hypothesis, which turns disparate phenomena into agreeing measurements of Planck’s constant (see Kao Reference Kao2015).

3. Probabilistic Measures of Unification

Consider a Bayesian agent whose credences are represented by a probability function Cr. We define the mutual information of a pair of propositions, {p1,p2}, relative to background b, byFootnote 4

(1)I(p1,p2b)=log2(Cr(p2p1b)Cr(p2b))=log2(Cr(p1p2b)Cr(p1b)Cr(p2b)).

If p 1 and p 2 are probabilistically independent on b, then I(p1,p2b) is zero; it is positive if conditionalizing on one boosts credence in the other, negative, if conditionalizing on one lowers credence in the other.

For a larger set, p={p1,p2,,pn}, we add up the information yielded by p 1 about p 2, the information yielded by p 1p 2 about p 3, and so on, up to the information about pn yielded by the conjunction of all the others.Footnote 5

(2)I(p1,,pnb)=I(p1,p2b)+I(p1p2,p3b)++I(p1pn1,pnb)=k=1n1I(i=1kpi,pk+1b).

Although the form of (2) does not make this obvious, this quantity is independent of the order in which the elements of the set p are taken, and we have

(3)I(p1,,pnb)=log2(Cr(p1p2pnb)Cr(p1b)Cr(p2b)Cr(pnb))=log2(Cr(i=1npib)Πi=1nCr(pib)).

With a slight abuse of notation, we will write I(pb) for I(p1,,pnb). We will also drop, as irrelevant, the base of the logarithm, since changing base is only a matter of a constant multiplicative factor.

The mutual information I(pb) is the logarithm of a quantity that appears in Keynes (1921, sec. 14.8) as the coefficient of dependence, with an attribution to unpublished work by W. E. Johnson.Footnote 6 It was called a measure of similarity by Wayne (Reference Wayne1995) and Myrvold (Reference Myrvold1996) and taken by Shogenji (Reference Shogenji1999) as a measure of coherence of a set of propositions.

We will say that a hypothesis h MIUnifies a set e={e1,,en}, relative to background b, if and only if

(4)I(ehb)>I(eb).

This suggests a way to measure the degree to which a hypothesis MIUnifies a set of evidential propositions.Footnote 7

(5)MIU1(e;hb)=I(ehb)I(eb).

We might also be interested in whether a hypothesis does a better job of unifying a set of propositions than its negation. Define

(6)MIU2(e;hb)=MIU1(e;hb)MIU1(e;h¯b)=I(ehb)I(eh¯b).

The two are not ordinally equivalent and, indeed, need not agree as to sign.

Suppose a hypothesis h unifies a body of evidence, relative to background b. That is, suppose the evidence is more mutually informative conditional on hb than on b alone. Then MIU1(e;hb) is positive. But whether MIU2(eb) is negative or positive depends on whether h¯ unifies the evidence more. If I(e;h¯b) is greater than I(e;hb), then, even if MIU1(e;hb) is positive, MIU2(e;hb) is negative. In fact, all four combinations of signs of MIU 1 and MIU 2 are possible, although it is easy to show that, unless e 1 and e 2 are, when taken individually, oppositely relevant to h (i.e., unless one of them is positively relevant and the other negatively relevant), if MIU1(e1,e2;hb) is positive, MIU2(e1,e2;hb) is also positive. See the appendix for details.

Both of these quantities are special cases of a comparative measure of unification:

(7)MIUc(e;h1,h2b)=I(eh1b)I(eh2b).

On McGrew’s account of consilience, h 1 is said to be more consilient than h 2 with respect to e to the extent that I(eh1b)>I(eh2b), or, equivalently, to the extent that MIUc(e;h1,h2b)>0 (McGrew Reference McGrew2003, 562).

Readers are asked to kindly refrain from engaging in a battle of the intuitions over whether MIU 1 or MIU 2 is the One True Measure of degree of unification. They are simply measuring different things, and if you have intuitions that are incompatible with properties that one or another of these quantities possesses, then your intuitions are about some other concept.Footnote 8

4. The Evidential Value of Unification

To some readers, it might seem obvious that what counts when it comes to confirmation is COU, with MIU being a poor cousin that hardly merits the illustrious family surname. A view of this sort is expressed by Marc Lange, who writes,

the examples I have given suggest that insofar as theories that unify in the stronger,Footnote 9 ontological-explanatory sense derive greater support in virtue of the unification they achieve, they do so not solely in virtue of their achieving unification in the weaker, creating-mutual-positive relevance sense. The stronger sense of unification is epistemically significant. In the case of the light-quantum hypothesis, hC and hL both supply unity in the weaker sense, but Einstein took hL to receive greater support from the phenomena than hC by virtue of hL’s unifying those phenomena in an ontological-explanatory sense. (Reference Lange2004, 212)

Here hL is Einstein’s light quantum hypothesis, and hC is the hypothesis that hL is false but nevertheless, by sheer coincidence, light behaves as if it were quantized. According to Lange, hL receives greater support from the phenomena unified than does hC.

It is not entirely clear whether incremental or absolute support is meant, where incremental support has to do with an increase in credibility lent to a hypothesis by the evidence, and absolute support with the credibility of the hypothesis, taking all known considerations into account. If absolute, this suggests that the case of hC is analogous to that of our toy example’s H 3, which is accorded a low prior because it posits an improbable coincidence. But, if the claim is to be a counterexample to anything in Myrvold (Reference Myrvold2003), incremental support must be what is meant. Let us therefore consider the position that, when it comes to incremental support, it is COUnification, not MIUnification, that counts.

A Bayesian analysis renders the opposite verdict: when it comes to incremental support of a hypothesis, it is MIUnification, rather than COUnification, that matters. One popular measure of the degree to which an evidential proposition e lends incremental confirmation to a hypothesis h, relative to background b, is the ratio of posterior probability of h to its prior probability. This is, of course, ordinally equivalent to its logarithm. Let us define

(8)R(h;eb)=log(Cr(heb)Cr(hb)).

Another is the ratio of the posterior odds of h to its prior odds, or, equivalently, the logarithm of this, called weight of evidence by Good (Reference Good1950). Define

(9)W(h;eb)=log(Cr(heb)/Cr(h¯eb)Cr(hb)/Cr(h¯b))=log(Cr(eh)Cr(eh¯)).

As Myrvold (Reference Myrvold2003) pointed out, on either way of measuring incremental confirmation, we have a contribution of unification to confirmation.Footnote 10 The incremental support, as measured by R, of h by e can be decomposed into a sum of increments due to the individual members of e, plus an additional term that is the degree of MIUnification (positive or negative) of e by h, as measured by MIU 1.

(10)R(h;i=1neib)=i=1nR(h;eib)+MIU1(e;hb).

The result for W takes the same form, with MIU 2 in place of MIU 1.

(11)W(h;i=1neib)=i=1nW(h;eib)+MIU2(e;hb).

These relations can be readily verified by the reader.

Since the MIU term is not the only contribution to the increment of confirmation, it would be incorrect to gloss these results as saying that hypotheses that are more unifying receive more confirmation. Although it would not be incorrect to say that ceteris paribus, a hypothesis that achieves a higher degree of MIUnification of the evidence is accorded greater incremental support, this is strictly weaker than what is conveyed in equations (10) and (11), and there is no advantage in making the ceteris paribus claim when it is a trivial matter to say how things stand when all else is not equal.

Imagine, now, a Bayesian agent that had numerical credences, which it updated by conditionalizing on new items of evidence.Footnote 11 Then, depending on how we measured degree of incremental confirmation, the confirmational boost accorded to h by a set e of evidential propositions would be given by (10) or (11). In each case the additional confirmational boost, beyond that attributable to the items of evidence taken singly, is given by the MIUnification term.

Applied to our toy example: the fact that H 1 and H 3 make E 1 and E 2 informative about each other is reflected in the likelihoods, Cr(E1E2H1) and Cr(E1E2H3), which are higher than Cr(E1E2H2) by a factor of 1,024. Thus, relative to H 2, credence in H 1 and H 3 is boosted:

(12)Cr(H1E1E2)Cr(H1)=Cr(H3E1E2)Cr(H3)=1,024×Cr(H2E1E2)Cr(H2).

It does not follow, of course, that H 3 gets final credence comparable to that of H 1. Since H 3 posits an improbable coincidence, it is accorded a lower prior probability, lower than that of H 2 by a factor of 1,024; the additional confirmational boost it receives is just enough to bring it up to posterior credence equal to that of H 2 (which, of course, must be the case, since, given the evidence, H 3 is true if and only if H 2 is).

There is a close parallel between this case and the case of geocentric v heliocentric world systems, and also the case of the light quantum, considered by Lange. In the case of planetary motion, on both the heliocentric hypothesis and the strengthened Ptolemaic hypothesis H SP, features of one planet’s apparent motion are informative about features of others’ (see Janssen [Reference Janssen2002] and Myrvold [Reference Myrvold2003] for discussion). In the case of the heliocentric hypothesis, HC, these have a common origin in the motion of our vantage point as observers on earth; for H SP, they are the consequence of the posited sun-planet parallelism. Against a background that includes little or no information about observed planetary motions, both of these get a confirmational boost from the celestial phenomena, due to the MIU component of incremental confirmation. It does not follow that they end up with equal posterior credence. Arguably, H SP, on that background, should be accorded markedly lower prior credence than the bare-bones geocentric hypothesis HP, as it posits a relation that HP by itself would not lead one to anticipate. The two hypotheses HC and H SP get the same incremental confirmation on the evidence. Therefore, posterior credence in HC will be markedly higher than posterior credence in H SP unless prior credence in HC is markedly lower than prior credence in HP.

Something similar can be said in regard to Lange’s case of the light quantum hypothesis. Let us grant that the light quantum hypothesis plays a unificatory role. Lange asserts that Einstein took the observed phenomena to lend greater support to the light quantum hypothesis than the hypothesis that, by sheer coincidence, all observable phenomena are as if the light quantum hypothesis is true. The suggestion is that such a judgment is the right one, given the evidence available to Einstein in 1905. In order for this claim to be relevant to the issue at hand, this must mean that the phenomena lend greater incremental support to the light-quantum hypothesis than to the coincidence hypothesis. One might also regard the hypothesis of coincidence as so implausible as to be dismissed out of hand. But this would mean according it a low prior, which is consistent with the Bayesian account of the virtue of unification.

5. Possible Reactions to the Bayesian Verdict

Bayesian updating leaves no room for an additional confirmatory boost to be attached to hypotheses with greater explanatory power; the contribution to incremental support comes via the MIUnification term. There is a tension between this Bayesian verdict and the thought that COUnification should play a role in incremental confirmation above and beyond its contribution to MIUnification. We have here an exact parallel with van Fraassen’s argument against those who would take explanatory power of a theory to yield an extra confirmatory boost, beyond that yielded by conditionalization on the evidence (van Fraassen Reference van Fraassen1989, 166–69).

One reaction might be to downplay the distinction, focusing on cases in which explanationist and Bayesian judgments agree. One might be tempted to declare that hypotheses that provide ‘lovelier’ explanations are precisely those that bestow higher likelihood on a hypothesis. This is not tenable as a general thesis. Although, in many interesting cases, explanation and likelihood go together, the connection is not so tight that they never come apart. The interesting question is what the explanationist will say about the cases in which they do come apart.

One possible reaction, in my opinion the correct one, is to use the Bayesian verdict to correct any intuitions one might have that are in tension with it. The ability of a theory to unify disparate phenomena by positing a common origin plays a confirmatory role only insofar as the posited common origin renders distinct phenomena informative (or more informative) about each other. A temptation to assign it a stronger role in confirmation might be ascribed, in part, to a conflation of distinct questions (a conflation encouraged by philosophers’ overuse of the phrase “theory choice,” a phrase that conflates distinct sorts of choices). Certainly, a hypothesis’s power to explain, if true, can contribute to making it worthwhile to pursue a project of developing a theory that includes that hypothesis, and it can contribute to the value of accepting the hypothesis, if true; we should only be wary of thinking that everything that contributes to making a hypothesis pursuit worthy also lends it greater credibility. The temptation might also be ascribed, in part, to not distinguishing between incremental confirmation and overall credibility in the light of all evidence. The most obvious examples that exhibit MIUnification without COUnification are those, such as our H 3, that achieve it by brute fiat, by tacking on an improbable conjunct, and we rightly regard these as implausible.

This suggests one way in which an explanationist might retrench; the import of COUnification might be relegated to informing priors. While, certainly, common-origin considerations sometimes play a role in assessing prior credibility, I am skeptical that anything beyond a very limited role can be defended; more on this in the next section.

The only other avenue of defense for an advocate of an explanationist thesis would be to deny that considerations of how a Bayesian agent would update have normative force for the judgments of human scientists. A defense along these lines of thought would have to ground it in some relevant difference between us and Bayesian agents. We are certainly different from Bayesian agents in a number of ways. We do not have precise numerical degrees of belief; our judgments about how likely or unlikely a hypothesis is tend to be vague. Moreover, as an abundance of empirical evidence shows, routinely our qualitative judgments of the relative credibility of various propositions are not even compatible with the existence of numerical credences satisfying the axioms of probability, and our changes in credences are often not in accord with Bayesian conditionalization.

The usual understanding of facts of this sort is that they are due to cognitive limitations and that some of them can be understood as resulting from usually reliable heuristics, of the sort that any agent with limited cognitive capacities would be well advised to employ as an alternative to spending excessive time on cogitation. In taking such limitations into account, one does not ipso facto abandon the domain of normativity for descriptive psychology. From a decision-theoretic point of view, deployment of such heuristics can be regarded as rational behavior for a cognitively limited agent. This involves what Good (Reference Good, Godambe and Sprott1971, Reference Sprott, Harper and Hooker1976) called Type II Rationality: decision making that takes into account the cost in time and cognitive effort of the act of deliberation.

Peter Lipton has offered a limited defense of explanationism along these lines. We are often not very good, he notes, at judging likelihoods correctly.

My thought is this. In many real life situations, the calculation that the Bayesian formula would have us make does not, in its bare form, meet the requirement of epistemic effectiveness: it is not a recipe we can readily follow. … My suggestion is that explanatory considerations of the sort to which Inference to the Best Explanation appeals are often more accessible than those probabilistic principles to the inquirer on the street or in the laboratory, and provide an effective surrogate for certain components of the Bayesian calculation. On this proposal, the resulting transition of probabilities in the face of new evidence might well be just as the Bayesian says, but the process that actually brings about the change is explanationist. (Lipton Reference Rackover2004, 113–14; see also Lipton Reference Lipton, Hon and Rackover2001, 110–11)

On such a view, when a judgment needs to be made on the fly, it is better to invoke an explanationist heuristic than to spend time thinking through likelihoods; this will, one hopes, provide judgments that are not too far off, either most of the time or in the most significant cases. Although Lipton suggests that the division of labor between Bayesian and explanationism maps onto the distinction between normative and descriptive accounts, he also uses language that suggests that we cognitively limited agents are well advised to employ explanationist considerations as a surrogate for doing a Bayesian calculation: “explanatory considerations help us to perform what is in effect a Bayesian calculation” (Lipton Reference Rackover2004, 120). This suggests that considerations of Type II rationality are in play.

Using a heuristic of this sort as a surrogate for a considered evaluation of likelihoods carries with it a risk of error, in those cases in which COU and MIU come apart. Presumably, Lipton would agree that, in such cases, if an accurate appraisal of the import of the evidence matters, one should correct the explanationist judgment by reference to the Bayesian one. On Lipton’s view, the role of explanationist considerations is severely constrained.

Can a stronger defense of explanationism be mounted? It is doubtful. Since such a defense would have to be grounded in some difference between cognitively limited humans and Bayesian agents, it is hard to see any role for explanationist consideration beyond the limited heuristic role envisaged by Lipton.

6. A Prior Preference for Unifying Hypotheses?

We have considered cases (in the toy example, H 1 and H 3; in the case of planetary motion, HC and H SP; and in the light quantum case, hL and hC) in which each of a pair of hypotheses possesses the same ability to render items of evidence informationally relevant to one another, but they do so in different ways. In each of these cases one does it by virtue of positing a common origin for prima facie unrelated phenomena, the other, by brute fiat, in positing an unexplained correlation between the phenomena. In each of these cases, the hypothesis that involves a common origin is, arguably, less implausible than the one that posits brute coincidence.

One might be tempted to generalize, positing that, whenever we have MIU without COU, there will be a corresponding hypothesis that achieves precisely the same MIUnification via COUnification, and we should accord much less prior credence to the hypothesis that exhibits MIU without COU than to the one that achieves it via COU. This would mean that there is a role for COU, not in incremental confirmation but in setting priors.

Anything so sweeping would be a mistake, I think. There are patterns in the world of all sorts, some due to some sort of common origin, some not. We should not demand that a common origin be found for every similarity between two phenomena. Given any pattern in the phenomena, however, it will be possible to cook up an artificial MIUnifying hypothesis. We ought not seek a common origin lurking behind every such hypothesis.

Perhaps, then, the generalization should be that, when we do have a pair of hypotheses that both induce the same informational relevance relations among a body of phenomena, one doing it via COUnification and the other by brute fiat, we should attach higher prior credence to the COUnifying hypothesis. This is still too sweeping. When we have a case of two hypotheses h 1 and h 2 of roughly equal prior credibility and create a third h 3 by tacking on to h 2 some conjunct with low prior plausibility, then, indeed, in such a case we should place lower credence in h 3 than in h 1. But not all cases will be like that, and a COUnifying hypothesis might be deemed implausible on other grounds. Take, for example, Ptolemy’s attitude toward heliocentric hypotheses. Since Ptolemy recognized that in the observed phenomena there is a connection between the apparent motion of the sun and that of the other planets, he was in a position to appreciate the COUnifying power of heliocentrism. But, since he accepted Aristotelian physics for terrestrial phenomena, he thought that terrestrial phenomena ruled out a diurnal rotation of the earth (see Ptolemy Reference Toomer1984, bk. I, sec. 7); for him, it was reasonable to place low credence in heliocentric theories that posited such a rotation.

One can exhibit plenty of hypothesis pairs in which the less unifying, less explanatory hypothesis has less prior credibility because the less explanatory hypothesis posits an implausible coincidence. But the emphasis should be on the credibility-diminishing role of coincidence, rather than any prior conviction that nature is unified. What H 3, the strengthened Ptolemaic hypothesis, and Lange’s hC have in common is that, in each case, we have a hypothesis to which is tacked on some additional condition that one would not expect to hold in the absence of evidence that it does, and hence we have a hypothesis that ought to be accorded low prior credence. Rather than a sweeping preference for COUnification, I suggest that the methodological adage that underwrites low prior credence in such hypotheses is

Place little prior credence in things you take to be improbable.

This is, I hope, unobjectionable! It is, of course, utterly empty, but I am skeptical that anything stronger could be defended as a maxim of more than very limited scope.

It would be a mistake to raise this bland but unobjectionable maxim into a global rejection of hypotheses that posit coincidences. Improbable things do happen, after all. Moreover, in some cases it is reasonable to accept hypotheses that posit an improbable coincidence. The evidence available to you in the toy example strongly suggests a common cause. But if you were to obtain strong evidence that the two data streams were the result of independent tosses of two fair coins, then it would be reasonable to accord high credence to H 3. For a real-world case: Ptolemy propounded a geocentric system with an unexplained sun-planet parallelism because he thought he had strong evidence to rule out hypotheses that involved a moving earth.

7. Unification and Reichenbachian Common Causes

Among unifying hypotheses are those that posit a Reichenbachian common cause to explain some observed statistical correlation (Reichenbach Reference Reichenbach1956, sec. 19). This type of hypothesis fits well within the schema of the Bayesian account of unification, but, since this might not be obvious, it is worth showing how it fits.

Consider two sequences of propositions, {Ai,i=1,,n} and {Bi,i=1,n}. Given such sequences, let n(A) be the number of true instances of the Ais, and let f(A)=n(A)/n be the relative frequency of true instances of the Ais. Define f(B) and f(AB) similarly. Let E 1 be a proposition expressing which of the Ais are true and which are false. For example, in our toy example, Ai could be the proposition that the ith element of S 1 is heads, and E 1 would be

A1A2A3A¯4A¯5A6A¯7A8A9A¯10.

Let E 2 be the evidence statement specifying the B-sequence.

A statistically significant difference between f(AB) and the product f(A)f(B) is thought to call for explanation. A Reichenbachian Common Cause of an observed correlation between A and B is a third sequence Ci that screens off their correlation. That is,

(13)Pr(AiBiCi)=Pr(AiCi)Pr(BiCi);Pr(AiBiC¯i)=Pr(AiC¯i)Pr(BiC¯i).

A hypothesis that posits a common cause of this sort, if it leads one to expect correlations close to those observed, clearly can be supported by evidence in which there is an observed statistical correlation between two sequences of events. Such a hypothesis can be a MIUnifying hypothesis, in the sense of making the evidence statements E 1 and E 2 mutually informative.

This might seem paradoxical. A common cause screens off the correlations between the Ais and Bis; how can it be that, at the same time, there is a confirmational boost associated with rendering them informative about each other?

The answer to this is that the hypothesized common causes Ci screen off the correlations, but a hypothesis Hcc that posits that there are common causes of the right sort can render the truth or falsity of Ai informative about the truth or falsity of Bi, and hence render E 1 and E 2 mutually informative. That is, a hypothesis that there is a common cause of the right sort will lead one to expect correlations between the Ais and Bis, and so count as MIUnifying with respect to the evidence set {E1,E2}, relative to a background against which the observed correlations are unexpected.

Moreover, each event Ci can count as a common origin of Ai and Bi. Let Hcc be some hypotheses according to which there exists a sequence {Ci} satisfying (13). Suppose that, on the supposition of Hcc, Ci is a probability raiser for both Ai and Bi, as a cause should be, and suppose that, according to Hcc, for each i, Ci and C¯i both have nonzero probability. Then, even though, for each Ci, the truth or falsity of Ci screens off informational relations between Ai and Bi, the supposition of Hcc leads one to expect correlations between the Ais and the Bis.

(14)Pr(AiBiHcc)>Pr(AiHcc)Pr(BiHcc).

Let us now see in more detail how this works. We consider the bearing of the statistical evidence stemming from observation of the A-sequence and the B-sequence on members of a family of hypotheses, each of which posits the existence of a Reichenbachian common cause. For simplicity, we consider only hypotheses on which distinct Ais are independent and identically distributed, as are {Bi} and {AiBi}. The statistical data can be accounted for on a hypothesis positing Cis that are also independently and identically distributed. Any hypothesis positing a common cause of this sort can be characterized by five parameters:

(15)p=Pr(Ci),a1=Pr(AiCi),a0=Pr(AiC¯i),b1=Pr(BiCi),b0=Pr(BiC¯i).

Probabilities for the Ais, Bis, conditional on a hypothesis of this sort, are

(16)Pr(AiHcc)=pa1+(1p)a0,Pr(BiHcc)=pb1+(1p)b0,

and their covariance is

(17)Cov(Ai,BiHcc)=Pr(AiBiHcc)Pr(AiHcc)Pr(BiHcc)=p(1p)(a1a0)(b1b0).

As pointed out by Reichenbach (Reference Reichenbach1956, 159–61), and as can be readily seen from (17), if p ∈ (0, 1) and a 1a 0 and b 1b 0 are both positive, then, conditional on the hypothesis Hcc, the Ais are positively correlated with the Bis. Obviously, the same conclusion follows if a 1a 0 and b 1b 0 are both negative; also, the Ais are negatively correlated with the Bis if a 1a 0 and b 1b 0 have opposite sign, and they are uncorrelated if the Cis are irrelevant to either the Ais or the Bis, that is, if a 1 = a 0 or b 1 = b 0.Footnote 12 The family of all such hypotheses, thus, includes as a special case those that posit no common cause for Ai and Bi.

We inquire into the degree of support lent to common-cause hypotheses, with various values of the parameters, by the pair {E 1, E 2}. Let Hcc be some hypothesis of the form considered above. We have, from (10),

(18)R(Hcc;E1E2)=R(Hcc;E1)+R(Hcc;E2)+MIU1({E1,E2};Hcc).

Since we are interested in comparing degrees of support for different hypotheses on a fixed body of evidence, it is useful to compare log-likelihoods, as, for two different hypotheses, the differences between their R-values will be the same as the differences between the respective log-likelihoods. The log-likelihoods can be partitioned in a manner parallel to our partitioning of R:

(19)logPr(E1E2Hcc)=logPr(E1Hcc)+logPr(E2Hcc)+I(E1,E2Hcc).

The first two terms of this are

(20)logPr(E1Hcc)=n(A)logPr(AiHcc)+n(A¯)logPr(A¯iHcc);logPr(E2Hcc)=n(B)logPr(BiHcc)+n(B¯)logPr(B¯iHcc).

These are maximized by a hypothesis Hcc that has Pr(AiHcc) = f(A) and Pr(BiHcc) = f(B). That is, these terms are largest for hypotheses that posit probabilities for the Ais and Bis that are equal to the observed relative frequencies.

The mutual information of E 1 and E 2, conditional on a hypothesis Hcc, is

(21)I(E1,E2Hcc)=n(AB)I(Ai,BiHcc)+n(AB¯)I(Ai,B¯iHcc)+n(A¯B)I(A¯i,BiHcc)+n(A¯B¯)I(A¯i,B¯iHcc).

Once Pr(AiHcc) and Pr(BiHcc) are fixed, this is maximized by taking

(22)Pr(AiBi)=f(AB).

Thus, in the expression (19) for the log-likelihood, we see that the first two terms reward hypotheses whose probabilities for Ai and Bi are close to the observed relative frequencies of these, and the last term, which corresponds to unification in the Mutual Information sense, rewards hypotheses with theoretical correlations close to the observed statistical correlations. What goes for log-likelihoods goes also for the evidential support R. Thus, when there is a difference between f(AB) and f(A)f(B), a common-cause hypothesis on which this difference is expected, by virtue of appropriate values of the parameters, counts as a MIUnifying hypothesis and thereby achieves greater support.

For example, consider a case in which we have two sequences {Ai}, {Bi}, with a significant positive correlation between them: f(AB) is much larger than f(A)f(B). Consider two hypotheses, Hcc and Hcc, which posit the existence of sequences {Ci} and {Ci}, respectively, such that

(23)Pr(AiHcc)=Pr(AiHcc)f(A);Pr(BiHcc)=Pr(BiHcc)f(B).

Suppose, now that, Hcc correctly predicts the correlations, but Hcc does not. That is, Pr(AiBiHcc) is close to f(AB), but Pr(AiBiHcc) is not. In such a case we will have

(24)MIU1({E1,E2};Hcc)>MIU1({E1,E2};Hcc).

Thus, for appropriate values of the parameters, the hypothesis Hcc affords MIUnification to the evidence set {E 1, E 2}, even though, in individual cases, the supposition Ci does not render Ai informative about Bi.

This does not prevent Ci from being regarded as a common origin of Ai and Bi. To take an example used by Lange in section 3 of his article, suppose that we take the clinical evidence to establish that some disease C can cause symptoms A and B. Then, if we observe A and B in some patient, this will raise our credence that C also occurs in that patient, even if the symptoms A and B are independent, conditional on C. In such a case, the support provided by the symptoms A and B to the hypothesis that the patient has disease C is just the sum of the supports given to the hypothesis by the individual items by themselves.

Lange raises the question whether we should place more credence in a hypothesis that posits a single disease than in one that posits two independent origins of the symptoms A and B. Suppose there are two other diseases D 1 and D 2, such that A but not B is a symptom of D 1 and B but not A is a symptom of D 2, and suppose further that the chance that a patient with D 1 exhibits symptom A is the same as that of a patient with Cand that the chance that a patient with D 2 exhibits symptom B is the same as that of a patient with C. Then, upon observation of both symptoms, the confirmational boost afforded to the hypothesis that the patient has C is the same as the boost afforded to the hypothesis that the patient has both D 1 and D 2. The issue then comes down to priors. Is the joint occurrence of D 1 and D 2 much rarer than the occurrence of C? If the answer is yes—as would be the case if the three diseases are equally rare, and D 1 and D 2 uncorrelated—then we should place more credence in the hypothesis that the patient has C. If not—if the disease C is so rare and D 1 and D 2 so common that more patients contract both D 1 and D 2 than C—then our credences should favor the two-disease hypothesis. It would clearly be a mistake for one’s credences to favor the C-hypothesis merely on the basis of a preference for common origin explanations.

8. Conclusion

MIU is not the same as common origin explanation and is neither a necessary nor a sufficient condition for a hypothesis to play an explanatory role. Nevertheless, in a host of interesting cases, MIUnification is a concomitant of common origin explanation. Moreover, when a hypothesis that renders an otherwise puzzling coincidence comprehensible by providing a common origin explanation does receive an incremental confirmational boost from a body of evidence, beyond that provided by the individual items of evidence, that boost stems from MIUnification.

So, at least, is the verdict delivered by a Bayesian analysis; there is no room in Bayesian conditionalization for an extra confirmatory boost that is due to COU. A proponent of an explanationist thesis, to the effect that we ought to take hypotheses that involve common origin explanations to receive greater incremental support than hypotheses that achieve the same degree of MIU without explanation, should be in a position to explain why what is impossible for a Bayesian agent is rational for us. As we have seen, there is a limited heuristic role for considerations of COU, based on considerations of Type II rationality. It is doubtful whether any stronger explanationist thesis can be defended.

Appendix

Appendix Given a probability function Pr and propositions h, e 1, e 2, define

(A1)U1=Pr(e1e2h)Pr(e1h)Pr(e2h)Pr(e1)Pr(e2)Pr(e1e2);
(A2)U2=Pr(e1e2h¯)Pr(e1h¯)Pr(e2h¯)Pr(e1)Pr(e2)Pr(e1e2).

Then we have

(A3)MIU1(e1,e2;h)=logU1;
(A4)MIU2(e1,e2;h)=log(U1U2).

Thus, MIU 1(e 1, e 2; h) is positive if and only if U 1 > 1, negative if and only if U 1 < 1, and zero if and only if U 1 = 1, and MIU 2(e 1, e 2; h) is positive if and only if U 1 > U 2, negative if and only if U 1 < U 2, and zero if and only if U 1 = U 2.

We want to show that each of the following four alternatives can be realized by some probability function.

  1. 1. MIU 1 > 0 and MIU 2 > 0; that is, U 1 > 1 and U 1 > U 2.

  2. 2. MIU 1 > 0 and MIU 2 < 0; that is, 1 < U 1 < U 2.

  3. 3. MIU 1 < 0 and MIU 2 > 0; that is, U 2 < U 1 < 1.

  4. 4. MIU 1 < 0 and MIU 2 < 0; that is, U 1 < 1 and U 1 < U 2.

It is easy to show (see lemma 1, below) that, if either e 1 or e 2 is irrelevant to h, then if U 1 > 1, U 2 < 1 and vice versa. Thus, it is easy to construct examples that satisfy conditions 1 and 4. Take Pr(e 1h) = Pr(e 1). Then, on an any probability function with U 1 > 1, we will have U 2 < 1 < U 1, and condition 1 will be satisfied. Similarly, if Pr(e 1h) = Pr(e 1), on any probability function with U 1 < 1, we will have U 1 < 1 < U 2, and condition 4 will be satisfied.

For condition 2, we need to have both U 1 and U 2 greater than 1. As is shown in lemma 1, below, this is possible only if e 1 and e 2 are relevant to h in opposite directions, that is, only if R(h; e 1) and R(h; e 2) have opposite signs. Here is one way to do it. Take, for simplicity, Pr(h) = Pr(e 1) = Pr(e 2) = 1/2, and take Pr(e 1e 2) = 1/4. Take Pr(e 1h) = 0.7, Pr(e 2h) = 0.3, and Pr(e 1e 2h) = 0.24. The reader can readily verify that these are consistent and that they determine the full probability function on Boolean combinations of {h, e 1, e 2}. In particular, they entail that Pr(e1h¯)=0.3, Pr(e2h¯)=0.7, and Pr(e1e2h¯)=0.26. We thus have U 1 = 24/21 and U 2 = 26/21, satisfying the desired conditions.

For condition 3, we can take the probability assignment described in the previous paragraph and create a new one by interchanging e 2 and e¯2. We have, once again, Pr(h) = Pr(e 1) = Pr(e 2) = 1/2, Pr(e 1e 2) = 1/4, Pr(e 1h) = 0.7, and Pr(e1h¯)=0.3. We also have Pr(e 2h) = 0.7, and Pr(e 1e 2h) = 0.46. These further entail that Pr(e2h¯)=0.3, and Pr(e1e2h¯)=0.04. We thus have U 1 = 46/49, and U 2 = 4/9, and so U 2 < U 1 < 1, and condition 4 is satisfied.

Having shown that all four alternatives are possible, we now prove the lemma alluded to above.

Lemma 1. Let {h, e 1, e 2} be logically independent propositions, and let Pr be a probability function on the Boolean algebra generated by this set. We assume that the denominators of the relevant fractions are nonzero and define U 1 and U 2 as above.

  • a) If Pr(he 1) = Pr(h) or Pr(he 2) = Pr(h), then if U 1 > 1, U 2 < 1 and vice versa.

  • b) If U 1 and U 2 are both less than one, then either both e 1 and e 2 are positively relevant to h or they are both negatively relevant to h.

  • c) If U 1 and U 2 are both greater than one, then one of {e 1, e 2} is positively relevant to h, and the other negatively relevant.

Proof. Let

(A5)p=Pr(h);q=Pr(h¯)=1p;α1=Pr(h|e1)/Pr(h);α2=Pr(h|e2)/Pr(h);β1=Pr(h¯|e1)/Pr(h¯);β2=Pr(h¯|e1)/Pr(h¯).

This allows us to write

(A6)U1=1α1α2Pr(e1e2h)Pr(e1e2);U2=1β1β2Pr(e1e2h¯)Pr(e1e2).

Once p, α1, α2, β1, and β2 are fixed, this yields a constraint on U 1 and U 2:

(A7)pα1α2U1+qβ1β2U2=1.

It is convenient to write this in terms of a weighted average of U 1 and U 2. Define

(A8)w1=pα1α2pα1α2+qβ1β2;w2=qβ1β2pα1α2+qβ1β2.

Then (A7) becomes

(A9)w1U1+w2U2=1pα1α2+qβ1β2,

with w 1 and w 2 both nonnegative, and

(A10)w1+w2=1.

It is instructive to rewrite the right-hand side of (A9) using the fact that pα1 + qβ1 = pα2 + qβ2 = 1. A bit of algebraic manipulation yields

(A11)w1U1+w2U2=1pq(α1β1)(α2β2)pα1α2+qβ1β2.

From (A11) it is readily apparent that, if either e 1 or e 2 is irrelevant to h—that is, if α1 = β1 or α2 = β2—then

(A12)w1U1+w2U2=1,

and in such a case, if U 1 > 1, then U 2 < 1, and vice versa. If we want to construct a case in which U 1 and U 2 are both greater than one, this requires the right-hand side of (A11) to be greater than one, which means that α1 − β1 and α2 − β2 must have opposite sign: one of {e 1, e 2} must be positively relevant to h, and the other negatively relevant. If we want to construct a case in which U 1 and U 2 are both less than one, then α1 − β1 and α2 − β2 must have the same sign: e 1 and e 2 are either both positively relevant or both negatively relevant to h. QED

Footnotes

I thank Michel Janssen, Marc Lange, Bill Harper, and Molly Kao for helpful discussions. I am grateful to Clark Glymour for raising the question, addressed in sec. 7, of how common-cause explanations fit into the framework. This work was supported, in part, by a grant from the Social Sciences and Humanities Research Council of Canada.

1. I am grateful to Michel Janssen for making this suggestion. See Myrvold (Reference Myrvold2011), and references therein, on the subject of how to incorporate cognitive values into a Bayesian framework.

2. Cf. Salmon (Reference Salmon, Hon and Rackover2001, 130): “the scientist might say that Halley’s hypothesis is worth pursuing, not because it is more likely to be true, but because, if it should turn out to be true, it would be extremely valuable in terms of informational content.”

3. Similar remarks apply to probabilistic measures of explanatory power such as those proposed by Popper (Reference Popper1954, Reference Popper1959), Good (Reference Good1960), Schupbach and Sprenger (Reference Schupbach and Sprenger2011), and Crupi and Tentori (Reference Crupi and Tentori2012). Glymour (Reference Glymour2015) has argued that it would be a grave mistake to take any of these probabilistic notions as an explication of explanatory power. This seems to be generally accepted by recent authors; Schupbach and Sprenger, for example, are clear that what is proposed is a measure of strength of explanation between propositions bearing an antecedently identified explanatory relation to each other.

4. A note on notation. We use concatenation for conjunction, and the overbar for the negation of p. We use boldface letters to denote sets of propositions. Note that these are sets and are not replaceable by a single proposition that is their conjunction. Thus, {p 1, p 2} is not the same set as {p 1p 2, T}, where T is the logically true proposition, although the conjunction of their members is the same. This matters because we will be concerned with the mutual informativeness of members of a set of propositions; p 1 and p 2 may be mutually informative, although the logically true proposition is not informative about their conjunction or anything else.

5. Obviously, a single number cannot capture all the informational relations there could be between elements of a set of more than two members. This would require a specification of all I(q, q′∣b), where q and q′ range over all conjunctions of elements of p. But it is this quantity that will be useful for the purposes at hand.

6. I am indebted to Brössel (Reference Brössel2015) for pointing this out.

7. This quantity is the logarithm of a quantity that was referred to as an “interaction term” in Myrvold (Reference Myrvold1996) and is called focused correlation in Wheeler (Reference Wheeler2009), Schlosshauer and Wheeler (Reference Schlosshauer and Wheeler2011), and Wheeler and Scheines (Reference Wheeler and Scheines2013). What we are calling MIU 1 was called U (for unification) in Myrvold (Reference Myrvold2003): MIU 2 was discussed therein, although not given its own name.

8. And if your intuitions find it repugnant to use the word “unification” in connection with either of these, then feel free to use a different word.

9. This is a slip; the two senses are, as Lange emphasizes, logically independent.

10. Equation (10) corresponds to (6) of Myrvold (Reference Myrvold1996) and to (12) of Myrvold (Reference Myrvold2003); (11) corresponds to (13) of Myrvold (Reference Myrvold2003). Closely related results appear already in Keynes (Reference Keynes1921, 151–54); in particular, our eq. (11) is essentially the same as Keynes’s (48).

11. I say “it” because a being with precise numerical credences would be far from human.

12. These probabilistic facts were familiar in the statistical literature well before Reichenbach’s use of them; see Yule (Reference Yule1911, secs. 4.6–7).

References

Brössel, P. 2015. “Keynes’s Coefficient of Dependence Revisited.” Erkenntnis 80:521–53.CrossRefGoogle Scholar
Crupi, V., and Tentori, K.. 2012. “A Second Look at the Logic of Explanatory Power with Two Novel Representation Theorems.” Philosophy of Science 79:365–85.CrossRefGoogle Scholar
Glymour, C. 2015. “Probability and the Explanatory Virtues.” British Journal for the Philosophy of Science 66:591604.CrossRefGoogle Scholar
Good, I. J. 1950. Probability and the Weighing of Evidence. London: Griffin.Google Scholar
Good, I. J. 1960. “Weight of Evidence, Corroboration, Explanatory Power, Information and the Utility of Experiments.” Journal of the Royal Statistical Society B 22:319–22.Google Scholar
Good, I. J. 1971. “Twenty-Seven Principles of Rationality.” In Foundations of Statistical Inference, ed. Godambe, V. P. and Sprott, D. A., 123–27. Toronto: Holt, Rinehart & Winston. Repr. in I. J. Good, Good Thinking: The Foundations of Probability and Its Applications (Minneapolis: University of Minnesota Press, 1983), 15–19.Google Scholar
Sprott, D. A. 1976. “The Bayesian Influence; or, How to Sweep Subjectivism under the Carpet.” In Foundations of Probability Theory, Statistical Inference, and Statistical Theories of Science, ed. Harper, W. L. and Hooker, C., 2:125–74. Dordrecht: Reidel. Repr. in I. J. Good, Good Thinking: The Foundations of Probability and Its Applications (Minneapolis: University of Minnesota Press, 1983), 22–55.Google Scholar
Henderson, L. 2014. “Bayesianism and Inference to the Best Explanation.” British Journal for the Philosophy of Science 65:687715.CrossRefGoogle Scholar
Janssen, M. 2002. “COI Stories: Explanation and Evidence in the History of Science.” Perspectives on Science 10:457522.CrossRefGoogle Scholar
Kao, M. 2015. “Unification and the Quantum Hypothesis in 1900–1913.” Philosophy of Science 82:12001210.CrossRefGoogle Scholar
Keynes, J. M. 1921. A Treatise on Probability. London: Macmillan.Google Scholar
Lange, M. 2004. “Bayesianism and Unification: A Reply to Wayne Myrvold.” Philosophy of Science 71:205–15.CrossRefGoogle Scholar
Lipton, P. 2001. “Is Explanation a Guide to Inference? A Reply to Wesley C. Salmon.” In Explanation: Theoretical Approaches and Applications, ed. Hon, G. and Rackover, S. S., 93120. Dordrecht: Kluwer.CrossRefGoogle Scholar
Rackover, S. S. 2004. Inference to the Best Explanation. 2nd ed. London: Routledge.Google Scholar
McGrew, T. 2003. “Confirmation, Heuristics, and Explanatory Reasoning.” British Journal for the Philosophy of Science 54:553–67.CrossRefGoogle Scholar
Myrvold, W. C. 1996. “Bayesianism and Diverse Evidence: A Reply to Andrew Wayne.” Philosophy of Science 63:661–65.CrossRefGoogle Scholar
Myrvold, W. C. 2003. “A Bayesian Account of the Virtue of Unification.” Philosophy of Science 70:399423.CrossRefGoogle Scholar
Myrvold, W. C. 2011. “Epistemic Values and the Value of Learning.” Synthese 87:547–68.Google Scholar
Perrin, J. 1913. Les Atomes. Paris: Alcan.Google Scholar
Perrin, J. 1916. Atoms. Trans. Hammick, D. L.. New York: Van Nostrand.Google Scholar
Popper, K. R. 1954. “Degree of Confirmation.” British Journal for the Philosophy of Science 5:143–49.CrossRefGoogle Scholar
Popper, K. R. 1959. The Logic of Scientific Discovery. New York: Basic.Google Scholar
Ptolemy. 1984. Ptolemy’s Almagest. Trans. Toomer, G. J.. London: Duckworth.Google Scholar
Reichenbach, H. 1956. The Direction of Time. Berkeley: University of California Press.CrossRefGoogle Scholar
Salmon, W. C. 2001. “Reflections of a Bashful Bayesian: A Reply to Peter Lipton.” In Explanation: Theoretical Approaches and Applications, ed. Hon, G. and Rackover, S. S., 121–35. Dordrecht: Kluwer.Google Scholar
Schlosshauer, M., and Wheeler, G.. 2011. “Focused Correlation, Confirmation, and the Jigsaw Puzzle of Variable Evidence.” Philosophy of Science 78:376–92.CrossRefGoogle Scholar
Schupbach, J. N. 2005. “On a Bayesian Analysis of the Virtue of Unification.” Philosophy of Science 72:594607.CrossRefGoogle Scholar
Schupbach, J. N., and Sprenger, J.. 2011. “The Logic of Explanatory Power.” Philosophy of Science 78:105–27.CrossRefGoogle Scholar
Shogenji, T. 1999. “Is Coherence Truth Conducive?Analysis 59:338–45.CrossRefGoogle Scholar
van Fraassen, B. 1989. Laws and Symmetry. Oxford: Oxford University Press.CrossRefGoogle Scholar
Wayne, A. 1995. “Bayesianism and Diverse Evidence.” Philosophy of Science 62:111–21.CrossRefGoogle Scholar
Wheeler, G. 2009. “Focused Correlation and Confirmation.” British Journal for the Philosophy of Science 60:79100.CrossRefGoogle Scholar
Wheeler, G., and Scheines, R.. 2013. “Coherence and Confirmation through Causation.” Mind 122:135–70.CrossRefGoogle Scholar
Yule, G. U. 1911. An Introduction to the Theory of Statistics. London: Griffin.CrossRefGoogle Scholar