How Not to Detect Design

A review of The Design Inference: Eliminating Chance Through Small Probabilities.
William A. Dembski. Cambridge University Press, Cambridge, UK, 1998. 272 pp.


Branden Fitelson
Christopher Stephens
Elliott Sober†‡
Department of Philosophy,
University of Wisconsin, Madison

As every philosopher knows, “the design argument” concludes that God exists from premisses that cite the adaptive complexity of organisms or the lawfulness and orderliness of the whole universe. Since 1859, it has formed the intellectual heart of creationist opposition to the Darwinian hypothesis that organisms evolved their adaptive features by the mindless process of natural selection. Although the design argument developed as a defense of theism, the logic of the argument in fact encompasses a larger set of issues. William Paley saw clearly that we sometimes have an excellent reason to postulate the existence of an intelligent designer. If we find a watch on the heath, we reasonably infer that it was produced by an intelligent watchmaker. This design argument makes perfect sense. Why is it any different to claim that the eye was produced by an intelligent designer? Both critics and defenders of the design argument need to understand what the ground rules are for inferring that an intelligent designer is the unseen cause of an observed effect.

Dembski’s book is an attempt to clarify these ground rules. He proposes a procedure for detecting design and discusses how it applies to a number of mundane and nontheological examples, which more or less resemble Paley’s watch. Although the book takes no stand on whether creationism is more or less plausible than evolutionary theory, Dembski’s epistemology can be evaluated without knowing how he thinks it bears on this highly charged topic. In what follows, we will show that Dembski’s account of design inference is deeply flawed. Sometimes he is too hard on hypotheses of intelligent design; at other times he is too lenient. Neither creationists nor evolutionists nor people who are trying to detect design in nontheological contexts should adopt Dembski’s framework.

The Explanatory Filter

Dembski’s book provides a series of representations of how design inference works. The exposition starts simple and grows increasingly complex. However, the basic pattern of analysis can be summarized as follows. Dembski proposes an “explanatory filter” (37), which is a procedure for deciding how best to explain an observation E:

(1) There are three possible explanations of E -- Regularity, Chance, and Design. They are mutually exclusive and collectively exhaustive. The problem is to decide which of these explanations to accept.

(2) The Regularity hypothesis is more parsimonious than Chance, and Chance is more parsimonious than Design. To evaluate these alternatives, begin with the most parsimonious possibility and move down the list until you reach an explanation you can accept.

(3) If E has a high probability, you should accept Regularity; otherwise, reject Regularity and move down the list.

(4) If the Chance hypothesis assigns E a sufficiently low probability and E is “specified,” then reject Chance and move down the list; otherwise, accept Chance.

(5) If you have rejected Regularity and Chance, then you should accept Design as the explanation of E.

The entire book is an elaboration of the ideas that comprise the Explanatory Filter.1 Notice that the filter is eliminativist, with the Design hypothesis occupying a special position.

We have interpreted the Filter as sometimes recommending that you should accept Regularity or Chance. This is supported, for example, by Dembski’s remark (38) that “if E happens to be an HP [a high probability] event, we stop and attribute E to a regularity.” However, some of the circumlocutions that Dembski uses suggest that he doesn't think you should ever “accept” Regularity or Chance.2 The most you should do is “not reject” them. Under this alternative interpretation, Dembski is saying that if you fail to reject Regularity, you can believe any of the three hypotheses, or remain agnostic about all three. And if you reject Regularity, but fail to reject Chance, you can believe either Chance or Design, or remain agnostic about them both. Only if you have rejected Regularity and Chance must you accept one of the three, namely Design. Construed in this way, a person who believes that every event is the result of Design has nothing to fear from the Explanatory Filter -- no evidence can ever dislodge that opinion. This may be Dembski's view, but for the sake of charity, we have described the Filter in terms of rejection and acceptance.

The Caputo Example

Before discussing the filter in detail, we want to describe Dembski’s treatment of one of the main examples that he uses to motivate his analysis (9-19,162-166). This is the case of Nicholas Caputo, who was a member of the Democratic party in New Jersey. Caputo’s job was to determine whether Democrats or Republicans would be listed first on the ballot. The party listed first in an election has an edge, and this was common knowledge in Caputo’s day. Caputo had this job for 41 years and he was supposed to do it fairly. Yet, in 40 out of 41 elections, he listed the Democrats first. Caputo claimed that each year he determined the order by drawing from an urn that gave Democrats and Republicans the same chance of winning. In spite of his protestations, Caputo was brought up on charges and the judges found against him. They rejected his claim that the outcome was due to chance, and were persuaded that he had rigged the results. The ordering of names on the ballots was due to Caputo’s intelligent design.

In this story, the hypotheses of Chance and Intelligent Design are prominent. But what of the first alternative, that of Regularity? Dembski (11) says that this can be rejected because our background knowledge tells us that Caputo probably didn’t innocently use a biased process. For example, we can rule out the possibility that Caputo, with the most honest of intentions, spun a roulette wheel in which 00 was labeled “Republican” and all the other numbers were labeled “Democrat.” Apparently, we know before we examine Caputo’s 41 decisions that there are just two possibilities -- he did the equivalent of tossing a fair coin (Chance) or he intentionally gave the edge to his own party (Design).

There is a straightforward reason for thinking that the observed outcomes favor Design over Chance. If Caputo had allowed his political allegiance to guide his arrangement of ballots, you’d expect Democrats to be listed first on all or almost all of the ballots. However, if Caputo did the equivalent of tossing a fair coin, the outcome he obtained would be very surprising. This simple analysis also can be used to represent Paley’s argument about the watch (Sober 1993). The key concept is likelihood. The likelihood of a hypothesis is the probability it confers on the observations; it is not the probability that the observations confer on the hypothesis. The likelihood of H relative to E is Pr(E|H), not Pr(H|E). Chance and Design can be evaluated by comparing their likelihoods, relative to the same set of observations. We do not claim that likelihood is the whole story, but surely it is relevant.

The reader will notice that the Filter does not use this simple likelihood analysis to help decide between Chance and Design. The likelihood of Chance is considered, but the likelihood of Design never is. Instead, the Chance hypothesis is evaluated for properties additional to its likelihood. Dembski thinks it is possible to reject Chance and accept Design without asking what Design predicts. Whether the Filter succeeds in showing that this possible is something we’ll have to determine.

The Three Alternative Explanations

Dembski defines the Regularity hypothesis in different ways. Sometimes it is said to assert that the evidence E is noncontingent and is reducible to law (39, 53); at other times it is taken to claim that E is a deterministic consequence of earlier conditions (65, 146n5); and at still other times, it is supposed to say that E was highly probable, given some earlier state of the world (38). The Chance Hypothesis is taken to assign to E a lower probability than the Regularity Hypothesis assigns (40). The Design Hypothesis is said to be the complement of the first two alternatives. As a matter of stipulation, the three hypotheses are mutually exclusive and collectively exhaustive (36).

Dembski emphasizes that design need not involve intelligent agency (8-9, 36, 60, 228-229). He regards design as a mark of intelligent agency; intelligent agency can produce design, but he seems to think that there could be other causes as well. On the other hand, Dembski says that “the explanatory filter pinpoints how we recognize intelligent agency (66)” and his section 2.4 is devoted to showing that design is reliably correlated with intelligent agency. Dembski needs to supply an account of what he means by design and how it can be caused by something other than intelligent agency.3 His vague remark (228-229) that design is equivalent to "information” is not enough. Dembski quotes Dretske (1981) with approval, as deploying the concept of information that the design hypothesis uses. However, Dretske’s notion of information is, as Dembski points out, the Shannon-Weaver account, which describes a probabilistic dependency between two events labeled source and receiver. Hypotheses of mindless chance can be stated in terms of the Shannon-Weaver concept. Dembski (39) also says that the design hypothesis isn't “characterized by probability.”

Understanding what “regularity,” “chance,” and “design” mean in Dembski’s framework is made more difficult by some of his examples. Dembski discusses a teacher who finds that the essays submitted by two students are nearly identical (46). One hypothesis is that the students produced their work independently; a second hypothesis asserts that there was plagiarism. Dembski treats the hypothesis of independent origination as a Chance hypothesis and the plagiarism hypothesis as an instance of Design. Yet, both describe the matching papers as issuing from intelligent agency, as Dembski points out (47). Dembski says that context influences how a hypothesis gets classified (46). How context induces the classification that Dembski suggests remains a mystery.

The same sort of interpretive problem attaches to Dembski’s discussion of the Caputo example. We think that all of the following hypotheses appeal to intelligent agency: (i) Caputo decided to spin a roulette wheel on which 00 was labeled “Republican” and the other numbers were labeled “Democrat;” (ii) Caputo decided to toss a fair coin; (iii) Caputo decided to favor his own party. Since all three hypotheses describe the ballot ordering as issuing from intelligent agency, all, apparently, are instances of Design in Dembski’s sense. However, Dembski says that they are examples, respectively, of Regularity, Chance, and Design.

The Parsimony Ordering

Dembski says that Regularity is a more parsimonious hypothesis than Chance, and that Chance is more parsimonious than Design (38-39). He defends this ordering as follows:

Note that explanations that appeal to regularity are indeed simplest, for they admit no contingency, claiming things always happen that way. Explanations that appeal to chance add a level of complication, for they admit contingency, but one characterized by probability. Most complicated are those explanations that appeal to design, for they admit contingency, but not one characterized by probability (39).

Here Dembski seems to interpret Regularity to mean that E is nomologically necessary or that E is a deterministic consequence of initial conditions. Still, why does this show that Regularity is simpler than Chance? And why is Chance simpler than Design? Even if design hypotheses were “not characterized by probability,” why would that count as a reason? But, in fact, design hypotheses do in many instances confer probabilities on the observations. The ordering of Democrats and Republicans on the ballots is highly probable, given the hypothesis that Caputo rigged the ballots to favor his own party. Dembski supplements this general argument for his parsimony ordering with two examples (39). Even if these examples were convincing,4 they would not establish the general point about the parsimony ordering.

It may be possible to replace Dembski’s faulty argument for his parsimony ordering with a different argument that comes close to delivering what he wants. Perhaps determinism can be shown to be more parsimonious than indeterminism (Sober 1999a) and perhaps explanations that appeal to mindless processes can be shown to be simpler than explanations that appeal to intelligent agency (Sober 1998). But even if this can be done, it is important to understand what this parsimony ordering means. When scientists choose between competing curves, the simplicity of the competitors matters, but so does their fit-to-data. You don’t reject a simple curve and adopt a complex curve just by seeing how the simple curve fits the data and without asking how well the complex curve does so. You need to ask how well both hypotheses fit the data. Fit-to-data is important in curve-fitting because it is a measure of likelihood; curves that are closer to the data confer on the data a higher probability than curves that are more distant. Dembski’s parsimony ordering, even if correct, makes it puzzling why the Filter treats the likelihood of the Chance hypothesis as relevant, but ignores the likelihoods of Regularity and Design.

Why Regularity is Rejected

As just noted, the Explanatory Filter evaluates Regularity and Chance in different ways. The Chance hypothesis is evaluated in part by asking how probable it says the observations are. However, Regularity is not evaluated by asking how probable it says the observations are. The filter starts with the question, “Is E a high probability event (38)?” This doesn’t mean “is E a high probability event according to the Regularity hypothesis?” Rather, you evaluate the probability of E on its own. Presumably, if you observe that events like E occur frequently, you should say that E has a high probability and so should conclude that E is due to Regularity. If events like E rarely occur, you should reject Regularity and move down the list. However, since a given event can be 5 described in many ways, any event can be made to appear common, and any can be made to appear rare.

Dembski’s procedure for evaluating Regularity hypotheses would make no sense if it were intended to apply to specific hypotheses of that kind. After all, specific Regularity hypotheses (e.g., Newtonian mechanics) are often confirmed by events that happen rarely -- the return of a comet, for example. And specific Regularity hypotheses are often disconfirmed by events that happen frequently. This suggests that what gets evaluated under the heading of “Regularity” are not specific hypotheses of that kind, but the general claim that E is due to some regularity or other. Understood in this way, it makes more sense why the likelihood of the Regularity hypothesis plays no role in the Explanatory Filter. The claim that E is due to some regularity or other, by definition, says that E was highly probable, given antecedent conditions.

It is important to recognize that the Explanatory Filter is enormously ambitious. You don’t just reject a given Regularity hypothesis; you reject all possible Regularity explanations (53). And the same goes for Chance -- you reject the whole category; the Filter “sweeps the field clear” of all specific Chance hypotheses (41, 52-53). We doubt that there is any general inferential procedure that can do what Dembski thinks the Filter accomplishes. Of course, you presumably can accept “E is due to some regularity or other” if you accept a specific regularity hypothesis. But suppose you have tested and rejected the various specific regularity hypotheses that your background beliefs suggest. Are you obliged to reject the claim that there exists a regularity hypothesis that explains E? Surely it is clear that this does not follow.

The fact that the Filter allows you to accept or reject Regularity without attending to what specific Regularity hypotheses predict has some peculiar consequences. Suppose you have in mind just one specific regularity hypothesis that is a candidate for explaining E; you think that if E has a regularity-style explanation, this has got to be it. If E is a rare type of event, the Filter says to conclude that E is not due to Regularity. This can happen even if the specific hypothesis, when conjoined with initial condition statements, predicts E with perfect precision. Symmetrically, if E is a common kind of event, the Filter says not to reject Regularity, even if your lone specific Regularity hypothesis deductively entails that E is false. The Filter is too hard on Regularity, and too lenient.

The Specification Condition

To reject Chance, the evidence E must be “specified.” This involves four conditions -- CINDE, TRACT, DELIM, and the description D* that you use to delimit E must have a low probability on the Chance hypothesis. We consider these in turn.

CINDE
Dembski says several times that you can’t reject a Chance hypothesis just because it says that what you observe was improbable. If Jones wins a lottery, you can’t automatically conclude that there is something wrong with the hypothesis that the lottery was fair and that Jones bought just one of the 10,000 tickets sold. To reject Chance, further conditions must be satisfied. CINDE is one of them.

CINDE means conditional independence. This is the requirement that Pr(E | H & I) = Pr(E | H), where H is the Chance hypothesis, E is the observations, and I is your background knowledge. H must render E conditionally independent of I. CINDE requires that H capture everything that your background beliefs say is probabilistically relevant to the occurrence of E.

CINDE is too lenient on Chance hypotheses -- it says that their violating CINDE suffices for them to be accepted (or not rejected). Suppose you want to explain why Smith has lung cancer (E). It is part of your background knowledge (I) that he smoked cigarettes for thirty years, but you are considering the hypothesis (H) that Smith read the works of Ayn Rand and that this helped bring about his illness. To investigate this question, you do a statistical study and discover that smokers who read Rand have the same chance of lung cancer as smokers who do not. This study allows you to draw a conclusion about Smith -- that Pr(E | H&I) = Pr(E | not-H &I). Surely this equality is evidence against the claim that E is due to H. However, the filter says that you can’t reject the causal claim, because CINDE is false -- Pr(E | H&I) [is not equal to] Pr(E | H).6

TRACT and DELIM
The ideas examined so far in the Filter are probabilistic. The TRACT condition introduces concepts from a different branch of mathematics – the theory of computational complexity. TRACT means tractability – to reject the Chance hypothesis, it must be possible for you to use your background information to formulate a description D* of features of the observations E. To construct this description, you needn’t have any reason to think that it might be true. For example, you could satisfy TRACT by obtaining the description of E by “brute force” – that is, by producing descriptions of all the possible outcomes, one of which happens to cover E (150- 151). Whether you can produce a description depends on the language and computational framework used. For example, the evidence in the Caputo example can be thought of as a specific sequence of 40 Ds and 1 R. TRACT would be satisfied if you have the ability to generate all of the following descriptions: “0 Rs and 41 Ds,” “1 R and 40 Ds,” “2 Rs and 39 Ds,” ... “41 Rs and 0 Ds.”

Whether you can produce these descriptions depends on the character of the language you use (does it contain those symbols or others with the same meaning?) and on the computational procedures you use to generate descriptions (does generating those descriptions require a small number of steps, or too many for you to perform in your lifetime?). Because tractability depends on your choice of language and computational procedures, we think that TRACT has no evidential significance at all. Caputo’s 41 decisions count against the hypothesis that he used a fair coin, and in favor of the hypothesis that he cheated, for reasons that have nothing to do with TRACT. The relevant point is simply that Pr(E|Chance) << Pr(E|Design). This fact is not relative to the choice of language or computational framework.

The DELIM condition, as far as we can see, adds nothing to TRACT. A description D*, generated by one’s background information, “delimits” the evidence E just in case E entails D*. In the Caputo case, TRACT and DELIM would be satisfied if you were able to write down all possible sequences of D’s and R’s that are 41 letters long. They also would be satisfied by generating a series of weaker descriptions, like the one just mentioned. In fact, just writing down a tautology satisfies TRACT and DELIM (165). On the assumption that human beings are able to write down tautologies, we conclude that these two conditions are always satisfied and so play no substantive role in the Filter.

Do CINDE, TRACT, and DELIM “Call the Chance Hypothesis into Question”?
Dembski argues that CINDE, TRACT and DELIM, if true, “call the chance hypothesis H into question.” We quote his argument in its entirety:

The interrelation between CINDE and TRACT is important. Because I is conditionally independent of E given H, any knowledge S has about I ought to give S no knowledge about E so long as --- and this is the crucial assumption --- E occurred according to the chance hypothesis H. Hence, any pattern formulated on the basis of I ought not give S any knowledge about E either. Yet the fact that it does in case D delimits E means that I is after all giving S knowledge about E. The assumption that E occurred according to the chance hypothesis H, though not quite refuted, is therefore called into question (147).

Dembski then adds:

To actually refute this assumption, and thereby eliminate chance, S will have to do one more thing, namely, show that the probability P(D* | H), that is, the probability of the event described by the pattern D, is small enough (147).

We'll address this claim about the impact of low probability later.

To reconstruct Dembski's argument, we need to clarify how he understands the conjunction TRACT & DELIM. Dembski says that when TRACT and DELIM are satisfied, your background beliefs I provide you with “knowledge” or “information” about E (143, 147). In fact, TRACT and DELIM have nothing to do with informational relevance understood as an evidential concept. When I provides information about E, it is natural to think that Pr(E | I) [is not equal to] Pr(E); I provides information because taking it into account changes the probability you assign to E. It is easy to see how TRACT & DELIM can both be satisfied by brute force without this evidential condition's being satisfied. Suppose you have no idea how Caputo might have obtained his sequence of D's and R's; still, you are able to generate the sequence of descriptions we mentioned before. The fact that you can generate a description which delimits (or even matches) E does not ensure that your background knowledge provides evidence as to whether E will occur. As noted, generating a tautology satisfies both TRACT and DELIM, but tautologies don't provide information about E.

Even though the conjunction TRACT & DELIM should not be understood evidentially (i.e., as asserting that Pr[E | I] [is not equal to] Pr[E]), we think this is how Dembski understands TRACT & DELIM in the argument quoted. This suggests the following reconstruction of Dembski's argument:

(1) CINDE, TRACT, and DELIM are true of the chance hypothesis H and the agent S.

(2) If CINDE is true and S is warranted in accepting H (i.e., that E is due to chance), then S should assign Pr(E | I) = Pr(E).

(3) If TRACT and DELIM are true, then S should not assign Pr(E | I) = Pr(E).

-----------------------

(4) Therefore, S is not warranted in accepting H.

Thus reconstructed, Dembski's argument is valid. We grant premiss (1) for the sake of argument. We've already explained why (3) is false. So is premiss (2); it seems to rely on something like the following principle:

(*) If S should assign Pr(E|H&I) = p and S is warranted in accepting H, then S should assign Pr(E|I) = p.

If (*) were true, (2) would be true. However, (*) is false. For (*) entails

If S should assign Pr(H|H) = 1.0 and S is warranted in accepting H, then S should assign Pr(H) = 1.0.

Justifiably accepting H does not justify assigning H a probability of unity. Bayesians warn against assigning probabilities of 1 and 0 to any proposition that you might want to consider revising later. Dembski emphasizes that the Chance hypothesis is always subject to revision.

It is worth noting that a weaker version of (2) is true:

(2*) If CINDE is true and S should assign Pr(H)=1, then S should assign Pr(E | I) = Pr(E).

One then can reasonably conclude that

(4*) S should not assign Pr(H) = 1.

However, a fancy argument isn’t needed to show that (4*) is true. Moreover, the fact that (4*) is true does nothing to undermine S's confidence that the Chance hypothesis H is the true explanation of E, provided that S has not stumbled into the brash conclusion that H is entirely certain. We conclude that Dembski's argument fails to “call H into question.”

It may be objected that our criticism of Dembski's argument depends on our taking the conjunction TRACT & DELIM to have probabilistic consequences. We reply that this is a charitable reading of his argument. If the conjunction does not have probabilistic consequences, then the argument is a nonstarter. How can purely non-probabilistic conditions come into conflict with a purely probabilistic condition like CINDE? Moreover, since TRACT and DELIM, sensu strictu, are always true (if the agent's side information allows him/her to generate a tautology), how could these trivially satisfied conditions, when coupled with CINDE, possibly show that H is questionable?

The Improbability Threshold

The Filter says that Pr(E | Chance) must be sufficiently low if Chance is to be rejected. How low is low enough? Dembski’s answer is that Pr(E(n) | Chance) < ½ , where n is the number of times in the history of the universe that an event of kind E actually occurs (209, 214- 217). As mentioned earlier, if Jones wins a lottery, it does not follow that we should reject the hypothesis that the lottery was fair and that he bought just one of the 10,000 tickets sold. Dembski thinks the reason this is so is that lots of other lotteries have occurred. If p is the probability of Jones’ winning the lottery if it is fair and he bought one of the 10,0000 tickets sold, and if there are n such lotteries that ever occur, then the relevant probability to consider is Pr(E(n) | Chance) = 1 - (1-p) . If n is large enough this quantity can be greater than ½, even though p is n very small. As long as the probability exceeds ½ that Smith wins lottery L2, or Quackdoodle wins lottery L3, or ... or Snerdley wins lottery Ln, given the hypothesis that each of these lotteries was fair and the individuals named each bought one of the 10,000 tickets sold, we shouldn’t reject the Chance hypothesis about Jones.

Why is ½ the relevant threshold? Dembski thinks this follows from the Likelihood Principle (190-198). As noted earlier, that principle states that if two hypotheses confer different probabilities on the same observations, the one that entails the higher probability is the one that is better supported by those observations. Dembski thinks this principle solves the following prediction problem. If the Chance hypothesis predicts that either F or not-F will be true, but says that the latter is more probable, then, if you believe the Chance hypothesis and must predict whether F or not-F will be true, you should predict not-F. We agree that if a gun were put to your head, that you should predict the option that the Chance hypothesis says is more probable if you believe the Chance hypothesis and this exhausts what you know that is relevant. However, this doesn’t follow from the likelihood principle. The likelihood principle tells you how to evaluate different hypotheses by seeing what probabilities they confer on the observations. Dembski’s prediction principle describes how you should choose between two predictions, not on the basis of observations, but on the basis of a theory you already accept; the theory says that one prediction is more probable, not that it is more likely.

Even though Dembski’s prediction principle is right, it does not entail that you should reject Chance if Pr(E(n) | Chance) < ½ and the other specification conditions are satisfied. Dembski thinks that you face a “probabilistic inconsistency” (196) if you believe the Chance hypothesis and the Chance hypothesis leads you to predict not-F rather than F, but you then discover that E is true and that E is an instance of F. However, there is no inconsistency here of any kind. Perfectly sensible hypotheses sometimes entail that not-F is more probable than F; they can remain perfectly sensible even if F has the audacity to occur.

An additional reason to think that there is no “probabilistic inconsistency” here is that H and not-H can both confer an (arbitrarily) low probability on E. In such cases, Dembski must say that you are caught in a "probabilistic inconsistency" no matter what you accept. Suppose you know that an urn contains either 10% green balls or 1% green balls; perhaps you saw the urn being filled from one of two buckets (you don’t know which), whose contents you examined. Suppose you draw 10 balls from the urn and find that 7 are green. From a likelihood point of view, the evidence favors the 10% hypothesis. However, Dembski would point out that the 10% hypothesis predicted that most of the balls in your sample would fail to be green. Your observation contradicts this prediction. Are you therefore forced to reject the 10% hypothesis? If so, you are forced to reject the 1% hypothesis on the same grounds. But you know that one or the other hypothesis is true. Dembski’s talk of a “probabilistic inconsistency” suggests that he thinks that improbable events can’t really occur -- a true theory would never lead you to make probabilistic predictions that fail to come true.

Dembski’s criterion is simultaneously too hard on the Chance hypothesis, and too lenient. Suppose there is just one lottery in the whole history of the universe. Then the Filter says you should reject the hypothesis that Jones bought one of 10,000 tickets in a fair lottery, just on the basis of observing that Jones won (assuming that CINDE and the other conditions are satisfied). But surely this is too strong a conclusion. Shouldn’t your acceptance or rejection of the Chance hypothesis depend on what alternative hypotheses you have available? Why can’t you continue to think that the lottery was fair when Jones wins it? The fact that there is just one lottery in the history of the universe hardly seems relevant. Dembski is too hard on Chance in this case. To see that he also is too lenient, let’s assume that there have been many lotteries, so that Pr(E(n) | Chance) > ½. The Filter now requires that you not reject Chance, even if you have reason to consider seriously the Design hypothesis that the lottery was rigged by Jones’ cousin, Nicholas Caputo. We think you should embrace Design in this case, but the Filter disagrees. The flaw in the Filter’s handling of both these examples traces to the same source. Dembski evaluates the Chance hypothesis without considering the likelihood of Design.

We have another objection to Dembski’s answer to the question of how low Pr(E(n) | Chance) must be to reject Chance. How is one to decide which actual events count as “the same” with respect to what the Chance hypothesis asserts about E? Consider again the case of Jones and his lottery. Must the other events that are relevant to calculating E(n) be lotteries? Must exactly 10,000 tickets have been sold? Must the winners of the other lotteries have bought just one ticket? Must they have the name “Jones?” Dembski’s E(n) has no determinate meaning.

Dembski supplements his threshold of Pr(E(n)|Chance) < ½ with a separate calculation (209). He provides generous estimates of the number of particles in the universe (1080 ), of the duration of the universe (1025 seconds), and of the number of changes per second that a particle can experience (1045 ). From these he computes that there is a maximum of 10150 specified events in the whole history of the universe. The reason is that there can’t be more agents than particles, and there can’t be more acts of specifying than changes in particle state.7 Dembski thinks it follows that if the Chance hypothesis assigns to any event that occurs a probability lower than 1/[(2)10150 ], that you should reject the Chance hypothesis (if CINDE and the other conditions are satisfied). This is a fallacious inference. The fact that there are no more than 10150 acts of specifying in the whole history of the universe tells you nothing about what the probabilities of those specified events are or should be thought to be. Even if sentient creatures manage to write down only N inscriptions, why can’t those creatures develop a well confirmed theory that says that some actual events have probabilities that are less than 1/(2N)?

Conjunctive, Disjunctive, and Mixed Explananda

Suppose the Filter says to reject Regularity and that TRACT, CINDE and the other conditions are satisfied, so that accepting or rejecting the Chance hypothesis is said to depend on whether Pr(E(n) | Chance) < ½. Now suppose that the evidence E is the conjunction E1&E2&...& Em. It is possible for the conjunction to be sufficiently improbable on the Chance hypothesis that the Filter says to reject Chance, but that each conjunct is sufficiently probable according to the Chance hypothesis that the Filter says that Chance should be accepted. In this case, the Filter concludes that Design explains the conjunction while Chance explains each conjunct. For a second example, suppose that E is the disjunction E1v E2 v ... v Em. Suppose that the disjunction is sufficiently probable, according to the Chance hypothesis, so that the Filter says not to reject Chance, but that each disjunct is sufficiently improbable that the Filter says to reject Chance. The upshot is that the Filter says that each disjunct is due to Design though the disjunction is due to Chance. For a third example, suppose the Filter says that E1 is due to Chance and that E2 is due to Design. What will the Filter conclude about the conjunction E1&E2? The Filter makes no room for “mixed explanations” -- it cannot say that the explanation of E1&E2 is simply the conjunction of the explanations of E1 and E2.

Rejecting Chance as a Category Requires A Kind of Omniscience

Although specific chance hypotheses may confer definite probabilities on the observations E, this is not true of the generic hypothesis that E is due to some chance hypothesis or other. Yet, when Dembski talks of “rejecting Chance” he means rejecting the whole category, not just the specific chance hypotheses one happens to formulate. The Filter’s treatment of Chance therefore applies only to agents who believe they have a complete list of the chance processes that might explain E. As Dembski (41) says, “... before we even begin to send E through the Explanatory Filter, we need to know what probability distribution(s), if any, were operating to produce the event.” Dembski’s epistemology never tells you to reject Chance if you do not believe you have considered all possible chance explanations.

Here Dembski is much too hard on Design. Paley reasonably concluded that the watch he found is better explained by postulating a watchmaker than by the hypothesis of random physical processes. This conclusion makes sense even if Paley admits his lack of omniscience about possible Chance hypotheses, but it does not make sense according to the Filter. What Paley did was compare a specific chance hypothesis and a specific design hypothesis without pretending that he thereby surveyed all possible chance hypotheses. For this reason as well as for others we have mentioned, friends of Design should shun the Filter, not embrace it.

Concluding Comments

We mentioned at the outset that Dembski does not say in his book how he thinks his epistemology resolves the debate between evolutionary theory and creationism.8 Still, it is abundantly clear that the overall shape of his epistemology reflects the main pattern of argument used in “the intelligent design movement.” Accordingly, it is no surprise that a leading member of this movement has praised Dembski’s epistemology for clarifying the logic of design inference (Behe 1996, pp. 285-286). Creationists frequently think they can establish the plausibility of what they believe merely by criticizing the alternatives (Behe 1996; Plantinga 1993, 1994; Phillip Johnson, as quoted in Stafford 1997, p. 22). This would make sense if two conditions were satisfied. If those alternative theories had deductive consequences about what we observe, one could demonstrate that those theories are false by showing that the predictions they entail are false. If, in addition, the hypothesis of intelligent design were the only alternative to the theories thus refuted, one could conclude that the design hypothesis is correct. However, neither condition obtains. Darwinian theory makes probabilistic, not deductive, predictions. And there is no reason to think that the only alternative to Darwinian theory is intelligent design.

When prediction is probabilistic, a theory cannot be accepted or rejected just by seeing what it predicts (Royall 1997, ch. 3). The best you can do is compare theories with each other. To test evolutionary theory against the hypothesis of intelligent design, you must know what both hypotheses predict about observables (Fitelson and Sober 1998, Sober 1999b). The searchlight therefore must be focused on the design hypothesis itself. What does it predict? If defenders of the design hypothesis want their theory to be scientific, they need to do the scientific work of formulating and testing the predictions that creationism makes (Kitcher 1984, Pennock 1999). Dembski’s Explanatory Filter encourages creationists to think that this responsibility can be evaded. However, the fact of the matter is that the responsibility must be faced.

References

Behe, M. (1996) Darwin’s Black Box. New York: Free Press.

Dembski, W. (1998), “Intelligent Design as a Theory of Information,” unpublished manuscript. Reprinted electronically at the following web site: http://www.arn.org/docs/dembski/.

Dretske, F. (1981): Knowledge and the Flow of Information. Cambridge, MA: MIT Press.

Fitelson, B. and Sober, E. (1998): “Plantinga’s Probability Arguments Against Evolutionary Naturalism.” Pacific Philosophical Quarterly 79: 115-129.

Kitcher, P. (1984): Abusing Science -- the Case Against Creationism. Cambridge, MA: MIT Press.

Pennock, R. (1999): Tower of Babel. Cambridge, MA: MIT Press.

Plantinga, A. (1993): Warrant and Proper Function. Oxford: Oxford University Press.

--------------- (1994): “Naturalism Defeated.” unpublished manuscript.

Royall, R. (1997): Statistical Evidence -- a Likelihood Paradigm. London: Chapman and Hall.

Sober, E. (1993): Philosophy of Biology. Boulder, CO: Westview Press.

----------- (1998): “Morgan’s Canon.” In C. Allen and D. Cummins (eds.), The Evolution of Mind, Oxford University Press, pp. 224-242.

----------- (1999a): “Physicalism from a Probabilistic Point of View.” Philosophical Studies, forthcoming.

----------- (1999b): “Testability.” Proceedings and Addresses of the American Philosophical Association, forthcoming.

Stafford, T. (1997): “The Making of a Revolution.” Christianity Today. December 8, pp. 16-22.

Notes

*. Received April, 1999; revised May, 1999.
†. Send requests for reprints to Elliott Sober, Philosophy Department, University of Wisconsin, Madison, WI 53706.
‡. We thank William Dembski and Philip Kitcher for comments on an earlier draft.

1. Dembski (48) provides a deductively valid argument form in which “E is due to design” is the conclusion. However, Dembski’s final formulation of “the design inference” (221-3) deploys an epistemic version of the argument, whose conclusion is “S is warranted in inferring that E is due to design.” One of the premisses of this latter argument contains two layers of epistemic operators; it says that if certain (epistemic) assumptions are true, then S is warranted in asserting that “S is not warranted in inferring that E did not occur according to the chance hypothesis.” Dembski claims (223) that this convoluted epistemic argument is valid, and defends this claim by referring the reader back to the quite different, nonepistemic, argument presented on p. 48. This establishes nothing as to the validity of the (official) epistemic rendition of “the design inference.”

2. For example, he says that “to retain chance a subject S must simply lack warrant for inferring that E did not occur according to the chance hypothesis H (220).”

3. Dembski (1998) apparently abandons the claim that design can occur without intelligent agency; here he says that after regularity and chance are eliminated, what remains is the hypothesis of an intelligent cause.

4. In the first example, Dembski (39) says that Newton’s hypothesis that the stability of the solar system is due to God’s intervention into natural regularities is less parsimonious than Laplace’s hypothesis that the stability is due solely to regularity. In the second, he compares the hypothesis that a pair of dice is fair with the hypothesis that each is heavily weighted towards coming up 1. He claims that the latter provides the more parsimonious explanation of why snake-eyes occurred on a single roll. We agree with Dembski’s simplicity ordering in the first example; the example illustrates the idea that a hypothesis that postulates two causes R and G is less parsimonious than a hypothesis that postulates R alone. However, this is not an example of Regularity versus Design, but an example of Regularity&Design versus Regularity alone; in fact, it is an example of two causes versus one, and the parsimony ordering has nothing to do with the fact that one of those causes involves design. In Dembski’s second example, the hypotheses differ in likelihood, relative to the data cited; however, if parsimony is supposed to be a different consideration from fit-to-data, it is questionable whether these hypotheses differ in parsimony.

5. Dembski incorrectly applies his own procedure to the Caputo example when he says (11) that the regularity hypothesis should be rejected on the grounds that background knowledge makes it improbable that Caputo in all honesty used a biased device. Here Dembski is describing the probability of Regularity, not the probability of E.

6. Strictly speaking, CINDE requires that Pr(E | H&J) = Pr(E | J), for all J such that J can be “generated” by the side information I (145). Without going into details about what Dembski means by “generating,” we note that this formulation of CINDE is logically stronger than the one discussed above. This entails that it is even harder to reject chance hypotheses than we suggest in our cancer example.

7. Note the materialistic character of Dembski’s assumption here.

8. Dembski has been more forthcoming about his views in other manuscripts. The interested reader should consult the following web site: http://www.arn.org/docs/dembski/.

Copyright 2000 Branden Fitelson, Christopher Stephens, Elliott Sober. All rights reserved. International copyright secured.
File Date: 7.05.00