Home| Letters| Links| RSS| About Us| Contact Us

On the Frontline

What's New

Table of Contents

Index of Authors

Index of Titles

Index of Letters

Mailing List


subscribe to our mailing list:



SECTIONS

Critique of Intelligent Design

Evolution vs. Creationism

The Art of ID Stuntmen

Faith vs Reason

Anthropic Principle

Autopsy of the Bible code

Science and Religion

Historical Notes

Counter-Apologetics

Serious Notions with a Smile

Miscellaneous

Letter Serial Correlation

Mark Perakh's Web Site

email this article to a friend printer-friendly format download format voice your opinion Digg It

The "Algorithm Room":
Can the "Design Inference"
Catch a Cheater?

By Wesley R. Elsberry

Posted May 6, 2002

William A. Dembski's writings claim that algorithms cannot produce Complex Specified Information (CSI), but intelligent agents can. A recent posting of Dembski's introduced qualifiers to CSI, so that we now have "apparent CSI" and "actual CSI". Dembski categorizes as "apparent CSI" those solutions which meet the formerly given criteria of CSI, but which are produced via evolutionary computation. This is contrasted with "actual CSI", in which a solution meets the CSI criteria and which an intelligent agent produces. See The Anti-Evolutionists: William A. Dembski and follow the link for "Explaining Specified Complexity".

Dembski is also fond of both practical and hypothetical illustrations to make his points. I'd like to propose a hypothetical illustration to explore the utility of the "apparent CSI"/"actual CSI" split.

Let's say that we have an intelligent agent in a room. The room is equipped with all sorts of computers and tomes on algorithms, including the complete works of Knuth. We'll call this the "Algorithm Room". We pass a problem whose solution would meet the criteria of CSI into the room (say, a 100-city tour of the TSP or perhaps the Zeller congruence applied to many dates). Enough time passes that our intelligent agent could work the problem posed from first principles by hand without recourse to references or other resources. The correct solution is passed out of the room, with an statement from the intelligent agent that no computational or reference assistance was utilized. Under those circumstances, we pay our intelligent agent at a high consultant rate. But if our intelligent agent simply used the references or computers, he would get paid at the lowly computer operator rate. We suspect that our intelligent agent not only utilized the references or computers to accomplish the task, but that he also used the time thus freed up to do some light reading, like "Once is Not Enough".

There are four broad categories of possible explanation of the solution that was passed back out of the "Algorithm Room". First, our intelligent agent might have employed chance, throwing dice to come up with the solution, and then waiting an appropriate period to pass the solution out. Given that the solution actually did solve the problem passed in, we can be highly confident that this category of explanation is not the actual one. Second, our intelligent agent might have ignored every resource of the "Algorithm Room" and spent the entire time working out the solution from the basic information provided with the problem (distances between cities or dates in question). Third, our intelligent agent might have gone so far as to look up and apply, via pencil and paper, some appropriate algorithm taken from one of the reference books. In this case, the sole novel intelligent action on our agent's part was looking up the algorithm. Essentially, our agent utilized himself as a computer. Fourth, our intelligent agent might simply have fed the basic data into one of the computers and run an algorithm to pop out the needed solution. Again, the intelligent agent's deployment of intelligence stopped well short of being applied to produce the actual solution to the problem at hand.

Because we suspect cheating, we wish to distinguish between a solution that is the result of the third or fourth categories of action, and a solution that is the result of the second category of action of our intelligent agent. We have only the attributes of the provided solution to the problem to go upon. Can we make a determination as to whether cheating happened or not?

Dembski's article, "Explaining Specified Complexity", critiques a specific evolutionary algorithm. Dembski does not dispute that the solution represents CSI, but categorizes the result as "apparent CSI" because the specific algorithm critiqued must necessarily produce it. Dembski then claims that this same critique applies to all evolutionary algorithms, and Dembski includes natural selection within that category.

The question all this poses is whether Dembski's analytical processes bearing upon CSI can, in the absence of further information from inside the "Algorithm Room", decide whether the solution received was actually the work of the intelligent agent (and thus "actual CSI") or the product of an algorithm falsely claimed to be the work of the intelligent agent (and thus "apparent CSI")?

If Dembski's analytical techniques cannot resolve the issue of possible cheating in the "Algorithm Room", how does he hope to resolve the issue of whether certain features of biology are necessarily the work of an intelligent agent or agents? If Dembski has no solution to this dilemma, the Design Inference is dead.

Wesley R. Elsberry is a student in Wildlife & Fisheries Sciences, Tx A&M U.