Uncommon Descent Serving The Intelligent Design Community

Two forthcoming peer-reviewed pro-ID articles in the math/eng literature

Share
Facebook
Twitter
LinkedIn
Flipboard
Print
arroba Email

The publications page at EvoInfo.org has just been updated. Two forthcoming peer-reviewed articles that Robert Marks and I did are now up online (both should be published later this year).*

——————————————————-

“Conservation of Information in Search: Measuring the Cost of Success”
William A. Dembski and Robert J. Marks II

Abstract: Conservation of information theorems indicate that any search algorithm performs on average as well as random search without replacement unless it takes advantage of problem-specific information about the search target or the search-space structure. Combinatorics shows that even a moderately sized search requires problem-specific information to be successful. Three measures to characterize the information required for successful search are (1) endogenous information, which measures the difficulty of finding a target using random search; (2) exogenous information, which measures the difficulty that remains in finding a target once a search takes advantage of problem-specific information; and (3) active information, which, as the difference between endogenous and exogenous information, measures the contribution of problem-specific information for successfully finding a target. This paper develops a methodology based on these information measures to gauge the effectiveness with which problem-specific information facilitates successful search. It then applies this methodology to various search tools widely used in evolutionary search.

[ pdf draft ]

——————————————————-

“The Search for a Search: Measuring the Information Cost of Higher Level Search”
William A. Dembski and Robert J. Marks II

Abstract: Many searches are needle-in-the-haystack problems, looking for small targets in large spaces. In such cases, blind search can stand no hope of success. Success, instead, requires an assisted search. But whence the assistance required for a search to be successful? To pose the question this way suggests that successful searches do not emerge spontaneously but need themselves to be discovered via a search. The question then naturally arises whether such a higher-level “search for a search” is any easier than the original search. We prove two results: (1) The Horizontal No Free Lunch Theorem, which shows that average relative performance of searches never exceeds unassisted or blind searches. (2) The Vertical No Free Lunch Theorem, which shows that the difficulty of searching for a successful search increases exponentially compared to the difficulty of the original search.

[ pdf draft ]

—————

*For obvious reasons I’m not sharing the names of the publications until the articles are actually in print.

Comments
JayM 270, I thought I'd check back on this. You've missed my points completely and are reduced to repeating yourself. The only way you can continue your assertion of "uniform probability distributions are not valid for biology" is to ignore specific cases where they are and then refer back to cases where they are not. Humorously, you then say that I haven't addressed your points yet I did not attempt to address them in the first place and took pains to explain why: they are valid within their limited scope as I've already said multiple times! If you don't get it at this point I don't know what else to say. Others are understanding my meaning. Although, it's obvious you haven't read No Free Lunch or other explanations of calculating CSI (or FCSI) since Dembski did exactly what you demanded in #255 and again in #270. You also made a gross error by referring to the entire genome instead of the specified sequence. So at this point I'm forced to conclude that you do not comprehend what you're attempting to criticize. Now my copy of No Free Lunch is somewhere in storage. Dembski spent pages going over the flagellum, but the result was 10^-1170 or around 3800+ informational bits (no calculator on hand so that # may be way off). As I said in 267 "the Explanatory Filter can take multiple types of inputs (which also makes it susceptible to GIGO and thus falsification). Two are (a) the encoded digital object and (b) hypothetical indirect pathways that lead to said objects." Now if MET mechanisms can indeed traverse an indirect pathway to the flagellum then Dembski's calculations amount to Garbage In. So, essentially what you're demanding of Dembski is that he accept what is in question as already verified: to assume that an indirect pathway for every object exists when doing calculations. That premise would make Dembski's position untenable. But since no one can even come up with a potential indirect pathway then how can we calculate otherwise? BTW, based upon the flagellum research of the time Dembski ran his calculations based upon the assumed IC core of 30 proteins out of the total 42. It's possible that less are required for the IC core, but even if halved that would not substantially change the outcome since it'd still be well over 500 informational bits. This is also why Dembski and Behe's positions are inter-dependent. As Behe acknowledged long ago IC as an indicator for intelligent agency is only sustainable if all potential indirect pathways are infeasible, which is where Dembski's work comes in. And Behe is trying to research the limits of MET mechanisms, which would validate Dembski's assumption. But like you I prefer Behe's line of inquiry since it's based on things more tangible and easier to comprehend. On a side note, here's one foresighted mechanism. And the specific ID-compatible hypotheses I was referring to were variants of front-loading and other hypotheses about foresighted mechanisms (viral, epigenetics, etc.). I cannot help it if you are ignorant about the full breadth of ID thought but please do not attempt to argue with me on this basis alone. Patrick
Jerry [273] "I just pointed out that it is hard to use an example of one variant of a species morphing into another variant of the same species as proof of Darwinian macroevolution." Questions of species don't change the fact that three single point mutations caused gross changes in the body plan of teosinte AND a new tissue in the covering of the kernels. Do micro evolution three times and presto, you've got macro evolution. "...no one here said teosinte came from corn when it was obvious that it is the reverse." I misinterpreted your statement in [269]: "The example of teosinte to corn is an example of devolution since the gene pool of teosinte probably contains the gene pool of corn except for a mutation or two." But, since this is now the topic of discussion, it would be begging the question to discuss it. djmullen
rna, These are interesting examples. And should be brought up at the appropriate time in the new Durston thread. If you do not, then I will. It could mean that I have misinterpreted what Durston has said, always a possibility, or there could be some other explanation. Read Durston's comments at https://uncommondesc.wpengine.com/intelligent-design/mathematically-defining-functional-information-in-biology/#comment-303307 Thank you for the information. jerry
Jerry @271
Maybe this will come out in Kirk Durston’s new thread when it appears but just what are the implications if the distribution is not uniform. Is it a big deal or is it closer to a nit that someone throws out to divert the discussion away from the obvious by posing a pseudo objection?
Excellent question. This should get the conversation back on topic. If the probability distribution of solutions generated by MET mechanisms is not uniform then the NFL theorems and Dr. Dembski's two new papers are not applicable, as written, to biological systems. In fact, MET predicts and observations confirm that the solutions generated by MET mechanisms do not form a uniform distribution, very far from it in fact. Further, the search space itself is not uniform -- there are clusters of viability. Given that, Dr. Dembski's papers cannot be said to be "pro ID" without further work to tie his results specifically to real world biological systems. The uniform probability distribution is also important in the calculation of CSI, but the discussion above has already explained that. JJ JayM
Jerry #271# "A more interesting question seems to be can any change or any distribution of changes ever lead to a protein reaching one of these other isolated protein islands of foldable proteins?" This has been investigated and there are cases where very few changes are necessary to connect two different islands of foldable proteins. One reported artificial example for a small protein (56 amino acids) showed that only seven mutations where necessary to reach a different stable! fold and a new functionality (Proc. Natl. Acad. Sci. USA, Vol. 104, No. 29 pages 11963-11968, Orban, Bryan & coworkers). An even more extrem example is the so called Arc-repressor protein where two mutations are necessary to change the fold and a single muation results in a protein existing in an equilibrium between two different folds. (Sauer & coworkers, Nature structural biology 2000, Vol. 7, page 1129-1132) For folds of RNA there is one example where an RNA can adopt two totally different folds associated with two totally different functions. Thus, the islands of possible protein folds are not necessarily isolated. Furthermore, one should not forget that not all functional proteins must be folded to function but many fold only upon interaction with a ligand. This means that even partially unfolded proteins might not be non-functional. The cited papers should be open access since they are more then a year old. rna
djmullen, I am stopping this conversation. Some others can waste their time with you. You do not know what ID is about and seem only to impose your own ill conceived perceptions. So you can have the last word till you start dealing with reality and what people say. For example, I know all the problems with the definition of species you bring up and they have been discussed here before and I have read about the problem in pro-Darwinian and anti ID books as well as with pro ID books and discussions and so it is a non issue. Wolves and dogs theoretically parted millions of years ago but still remain the same species by biological textbooks. My example of the Chihuahua and the wolf was meant to point out the absurdity of your example. I just pointed out that it is hard to use an example of one variant of a species morphing into another variant of the same species as proof of Darwinian macroevolution. Especially when it looks like the mechanism is probably artificial selection and not natural selection. And if it was by natural selection, ID accepts that example with open arms. So what point are you making by first choosing it and then defending it. And you defend it badly because no one here said teosinte came from corn when it was obvious that it is the reverse. It seems the indigenous people of the Americas started the process several thousand years ago probably first by accident and then maybe they got smart. So what is your point. Your discussion seems pointless. Don't bring up trivial examples that ID is wholly in sync with and expect us to take you seriously. Nothing you said so far has any merit except maybe in your own mind. Adios. jerry
jerry [269] I see you've found some of the problems you run into when trying to use the species concept in an evolved world. "Species" is an idea straight from the ancient Babylonian world and its Old Testament variations. An Intelligent Designer God designs and builds animals and plants separately and they grow separately and you expect to always be able to clearly separate different species from each other. But in an evolved world, a single population starts to split and you have two populations that share some attributes and not others and you have real trouble separating one from the other. Hence we have the problem of teosinte and corn and wolves and chihuahuas. Are they a single species or two? Scientists wrestled with this idea for quite a long while before finally, rather grudgingly, settling on the ability to cross fertilize as the definition of species. Our official, man-made definition of species is any group of animals that can mate and produce fertile offspring. Under evolutionary conditions, this marks the point where some key genes have diverged enough so they don't match up when mated and the offspring either dies or some key part of it's reproduction process is inoperative. It works in the sense that from that point on the two populations are forever separated and will always evolve separately, but, as you've shown, that definition has a lot of problems. Another problem is that it's useless with non-sexually reproducing organisms and that includes the vast majority of all living creatures, which are single celled. You and I, for instance, have more bacterial cells in us (mostly in the digestive tract) than human cells. Under evolution, there's just no hard and fast division between two populations that are evolving apart short of reproductive incompatability, even though one is very much needed. For instance, how do you call it when two plants have 99.9999+% identical genomes, but one is a clump of grass and one is corn? Or one is a chihuahua and one is a wolf? You can say they're both the same species, but if you plant teosinte instead of corn, your family is going to starve and if a chihuahua goes for your throat it's very annoying, but if a wolf goes for your throat, you're dead. I guess one way to know if ID or evolution is right is to ask how easy it is to separate species and so far the data points entirely towards evolution. I'm afraid teosinte is NOT mutated corn. We have archeological specimens of teosinte plants and pollen going back a long, long ways, but corn doesn't even start to appear until about 9,000 years ago and it's a very crude, teosinte-like version of corn. Incidently, all of the early differences are due to natural selection. Humans changed the teosinte environment in a big way when they started eating it and planting it. Once people started saving their best plants for seed and eating the lesser plants, it put natural selection on steroids and the plan transformed from teosinte, a barely edible grass, into corn, a highly desireable and nutritious plant to humans, in a few thousand years. "When we refer to no natural examples of FCSI ever being formed by nature we do not include life since life is the topic under discussion." My mind just boggles at this. When we refer to no natural examples of FCSI ever being formed by nature, we do not include life, although life is a part of nature, since life is the topic under discussion." Since the question is, "Does life produce FCSI?" you seem to have constructed an impregnable position. A veritable Maginot line of logic. "... the changes we have seen are all trivial in terms of genomic change and no complex new functional capabilities have ever been demonstrated." It's certainly true that the changes we see in teosinte to corn are "trivial in terms of genomic change". There are about 2 billion DNA base-pairs in the corn genome and if we have three point mutations, that's only 3/2,000,000,000ths of the corn genome. And yet just those three changes make the plant taller and narrower and concentrate the ears near the stalk, the ears get larger and the coating on the kernels gets softer. "Trivial" changes in the genome can obviously make very significant changes in the phenotype and it's the phenotype that feeds you or rips your throat out. As for "no complex new functional capabilities have ever been demonstrated", I'd say that just those three mutations produce functional changes that are complex enough to turn a barely edible weed into something that can start to feed the entire New World. How many base-pairs do you think mutated to turn teosinte into corn? I'm betting it's way under a thousand, which, as you say, is "trivial in terms of genomic change". I'm glad you brought up bat sonar. I read something about that a few years ago. This is from memory and I don't have a citation handy, but since it's the topic under discussion I don't see why I have to provide one. Don't want to beg the question, after all. For starters, you and I can already do a crude form of echolocation. If you shout, "Hello!" and hear a fainter "Hello!" a few seconds later, you know you're some distance from a large reflecting object, such as a cliff. If you say, "Hi!" and hear an almost immediate series of echos, you're in a gymnasium or concert hall sized room. If you say, "Hi!" and instantly get a load of echos in return, you're in a small room. If you hear a lot of high frequencies in the echo, the surfaces are hard, if not they're soft. You can tell direction pretty well too, by using the already existing ability of your ears and brain to tell what direction a sound is coming from. Blind people get very good at this, by the way. Now if you're a small flying mammal hunting insects in dim light, what kind of changes can help you find more insects via echo-location? Well, for starters you can shout, "Hi!" louder. This will detect insects farther away. You can make your shouts shorter, too, which makes your distance measurements more accurate. Or, like some bats, you can "chirp" where you send out a loud cry that sweeps from a low to high frequency. That helps you sort out the mass of echos that come back to you. If you hear low and high sounds, you'll know that the low pitched sounds are coming from a greater distance away because they were uttered first. Better ears will help too. If the shape of your outer ears changes, they can make sounds coming from different directions have phase relationships that increase your ability to tell direction. Your face can change too, to make those phase relationships clearer. (That's why bats tend to have faces like gargoyles - it helps them determine directions.) If you develop a really, really loud chirp or yelp, you can have a problem with your voice damaging your hearing. (I've heard some bats have such loud calls that they can stun insects with them.) Contracting muscles in the middle ear while you shout can clamp the auditory bones, protecting them from damage. I'm sure there are many more ways to improve echo location. Now can you tell us which, if any, of these changes are impossible to produce? And remember before answering that all of them and more exist in contemporary bats and other echo-locators. djmullen
Maybe this will come out in Kirk Durston's new thread when it appears but just what are the implications if the distribution is not uniform. Is it a big deal or is it closer to a nit that someone throws out to divert the discussion away from the obvious by posing a pseudo objection? If one had 60 seconds to explain the implications of the likely distribution to the President of the US, what would that conversation sound like. Is it a big ho hum or does it have serious implications for anything? A more interesting question seems to be can any change or any distribution of changes ever lead to a protein reaching one of these other isolated protein islands of foldable proteins? jerry
Patrick @267
JayM 261 First You only succeed in continuing the "non-uniform argument" by changing the metric to things that seem nonsensical. You might as well have said, "I ignore what you just said, I ignore the pointed questions, and I prefer to focus on aspects of the discussion that are already acknowledged as being valid."
Frankly, I could leverage the same charge against you, with far more supporting evidence. You have responded with great verbosity, yet have not even begun to address a single one of my arguments. Allow me to demonstrate:
Now back to the main disagreement. Which is that uniform probability distributions do not apply to biology at ALL in regards to MET. In fact, I would expect non-uniformity to be more complete in its breadth if some ID-compatible hypotheses are true.
Why? What specific hypotheses do you mean?
But to be fair we’re assuming MET and Darwinism as a starting point in evaluating its claims.
My understanding as well is that they are equally likely, but MET mechanisms do not create genomes from scratch so the assumption of a uniform probability distribution across the entire genome is invalid.
I’m struggling to even comprehend your objection. Non-foresighted variation is produced by copying errors. How does the mere act of replication cause these copying errors at any particular point(s) in the genome to become unequally probable/non-uniform?
MET mechanisms do not assert that every location in a genome is likely to change, as would be required by a uniform probability distribution. In fact, MET mechanisms are very conservative. Genomes change very, very little from generation to generation. Further, only those changes that are viable are preserved and only those that are viable and result in surviving offspring are propagated. This is far from a uniform distribution in genome space.
Depending on the form of variation, and the error correction (human versus bacterium) they’re certainly unequal in size or scope, but that does not change this observation.
What observation? The fact is that MET mechanisms result in child populations differing very, very little from their parent populations in terms of allele frequency. There is not a uniform distribution of potential changes.
And at this point natural selection (the localized search you’re so focused on) is irrelevant since it’s after the fact and may not even be a component at all in calculations.
I have no idea what point you're trying to make here. Natural selection operates on the non-uniform distribution of changes, probabilistically preserving those with greater fitness. This mechanism is completely at odds with the assumption of a uniform probability distribution of solutions.
Now if we have foresighted variation that seems to be outside of the bounds of MET (and said existence was predicted by ID proponents). Although some insist that the non-foresighted mechanisms of MET somehow produce the foresighted mechanisms we’ve observed.
What foresighted mechanisms have been observed? I have seen nothing in the ID or peer-reviewed mainstream literature to suggest that foresight has been rigorously documented.
This is why Dr. Dembski’s papers, however interesting, do not appear to be directly supportive of ID theory.
And the ID theory of Dembski and Marks conflicts with micro-evolution how?
Not at all. In fact, the papers do not describe any "ID theory" at all and do not tie their assumptions and conclusions to any real or hypothesized biological mechanisms.
Again, I’d rather Dembski and Marks make the connection to biology specific but my assumption is that their focus is narrow and they’re largely constraining their argument to specific cases relevant to (a) macro-evolution via the unguided processes of MET and (b) OOL.
However, they do not do this. Their arguments do not address the claims of MET at all, as far as I can see. That's why these two papers cannot be said to support ID as they stand.
This would still not be a uniform probability distribution, because the probabilities of small changes given MET mechanisms is much greater than that of large changes.
You’re changing the metric again.
No, I am not. A uniform probability distribution requires that all changes be equally likely so that subsequent populations are exploring the whole genome space. That is not the claim of MET and is not what is observed in real biological systems. Any mathematics that presumes a uniform probability distribution, as do the NFL theorems, is not applicable to biological systems.
Yes, small changes are more probable.
Thank you. You have admitted that MET mechanisms do not result in the uniform probability distribution required by the two papers under discussion.
But your point is only relevant to the localized search. What exactly about the OCCURRENCE of unguided variation makes small or large changes to any specific location of the genome unequally probable?
We're not talking about "unguided variation" in general, we're talking about MET mechanisms in particular. They do not result in uniform probability distributions of genomes within genome space.
Here’s the point related to biology. I forget the informational bits for the flagellum as calculated by Dembski (isn’t that in No Free Lunch…it’s been years since I read it?). But if anyone submits a reasonable hypothesis for an indirect pathway composed of steps which are less than 500 informational bits EACH then Dembski’s design inference for that particular biological object is rendered invalid.
This is "ID of the gaps." We need to prove that it couldn't happen, not sit back and challenge scientists to prove it couldn't. MET mechanisms demonstrably generate complexity by funneling information from the environment to the genomes of subsequent populations. So far, evolutionary biologists haven't demonstrated something as complex as a flagellum evolving, but they have shown the evolution of some structures and pathways. Until we show real limits to those mechanisms, the idea that many small steps result in a large journey is at least credible.
In addition, the calculation of CSI uses the length of the genome as a factor in determining how "specified" a sequence is. That is not logical given the small changes possible with MET mechanisms.
The calculation of CSI in biology does??
Yes.
Can you be more specific in what you mean (or provide a reference)? Or did you mean the "length[complexity] of the specified sequence"?
Exactly.
Dembski’s specification is very general in its application. Which leads to the problem of subjective specifications. Art is subjective and requires an intelligent agent to identify it (although I suppose it can be argued that realist landscapes can be objectively specified on a comparative basis). But biology contains objective specifications independent of the observer based upon the machine functionality. That’s why people refer to FCSI in regards to biology.
None of this addresses the problem I pointed out, namely that the CSI calculation assumes creation from scratch, which is not the claim of MET. Further, people refer to FCSI, but no one in the ID camp has yet demonstrated how to calculate it objectively for real world biological organisms or structures.
Viable regions are clearly clustered or you and I wouldn’t be here. We each differ from both of our parents, and yet we live. That demonstrates that there are multiple connected points in genome space that are viable. The question, of course, is how large these regions are and how connected.
You make my point for me, which is all I meant by saying "'can' be clustered".
If they are clustered, then MET mechanisms can easily keep subsequent populations in the "sweet spot", so all the calculations that assume a uniform distribution of solutions in the genome space are not applicable. If we want to make ID theory truly scientific, we have to address what is known and provide better predictions of what is not known. That means finding where the real "edge of evolution" lies, not attempting to apply invalid assumptions like uniform probability distributions to observed phenomena to which they clearly do not apply. JJ JayM
"But teosinte into corn?" Poor choice since both are the same species. Just as a Chihuahuas and a wolves are the same species. You are confusing physical difference with macro evolution of complex new functional capabilities which all take place at the genome level. It is also not clear which differences between teosinte and corn are due to artificial selection and which to natural selection. The example of teosinte to corn is an example of devolution since the gene pool of teosinte probably contains the gene pool of corn except for a mutation or two. It is unlikely you could get teosinte from corn but it may happen and would be an interesting experiment. When we refer to no natural examples of FCSI ever being formed by nature we do not include life since life is the topic under discussion. Let me lay out the logic of the argument for you. Excluding life there is no example of FCSI forming by law or chance. That is nature does not create FCSI. You can not point to life because that is begging the question. Intelligent activity produces FCSI all the time. Now if one set of causes has not shown the capability of producing the outcome while another set of causes does it all the time it seem reasonable to say that the reason for the FCSI could be due to the activity that does it all the time and not to the one that has never done it. Notice I did not say absolutely only that it is reasonable that it originated this way. One cannot do it and another can. So which one would you include as a potential cause. Think how unreasonable it would be to say that the process that has never done it is the only that can be considered as the cause while the process that does it all the time cannot be considered. How does one justify that logic? Now once FCSI is there and there are mechanisms for changing it, then one has to look at these mechanism and see what power they might have to create new FCSI through this change mechanism. So ID does not deny that these change mechanisms may generate new FCSI. It is just that the changes we have seen are all trivial in terms of genomic change and no complex new functional capabilities have ever been demonstrated. In other words not only does the gene pool have to be significantly increased but this increase along with other elements that were there previously must coordinate a system with new capabilities that did not exist before. Think bat sonar, insect wings, avian oxygen processing system for some starters not a a morphing of teosinte to corn or a wolf to a Chihuahua. jerry
JayM [231] "Your overall message, while somewhat brutally stated..." "Not all ID proponents are young earth creationists." I have to apologize to you and the others who have pointed this out. I didn't mean to say that only Young Earth Creationists are clinging to the old "God in Six Days" theory or that all IDists are also YECs, although I gather that most of them are. Put it down to writing using stolen moments on a netbook with a tiny screen. What I really meant is even more brutal: ID is the only part of the Old Testament theory that has even a faint patina of intellectual respectability today. The "God in Six Days Theory" incorporated a young earth, simultaneous creation of plans and animals, a separate creation of man and an intelligent creator. The Old Earth was dead long before "Origin of Species" was published. If I remember right, Reverend William Buckland, Oxford professor of geology and fervent Christian believer finally gave up and admitted that Genesis could not be reconciled with Geology circa 1835. The reverand spent decades trying to reconcile the two, he was a very strong believer, very intelligent and very knowledgable about both the Bible and geology and his pronouncement pretty well killed the young earth part of the theory. Simultaneous creation of plants and animals and separate creation of man were lost later, though I'd have to dig out the books to tell you when and how. The only thing left of the Old Theory is an intelligent being doing the creating. This idea is still very faintly respectable in that you can say that God is using evolution to do the heavy lifting and only giving it occasional tweeks or that God put enough info into the system around the Origin of Life to do the guiding. Front loading has lost almost all respectability with the discovery that life is billions of years old and that information that's not preserved by natural selection doesn't last nearly that long. The divine tweaking of evolution hasn't exactly been disproven yet, but it God is doing that, he's making it look just like evolution. Anyway, that's my reading of the present state of the "God in Six Days" theory. jerry [234] "No natural process has ever been seen creating specification..." Look at my examples of teosinte to corn evolution. Natural processes - point mutations - change the branch structure, make the ears bigger and move the ears closer to the stem. Three mutations, three specifications. Natural processes DO create specification. CJYman - Forget all the math coming from Dembski and Marks. They are making an elemental error. They are trying to prove that it's impossible to land in the target area of the genome (what I've been calling the "sweet spot") from the outside. This target area is the part of all possible genomes which contains DNA strings that will construct and operate an organism that is both alive and successfully reproducing. IF YOU'RE NOT IN THE SWEET SPOT, YOU'RE EITHER DEAD OR STERILE! Dembski and Marks are saying that if you're dead or sterile, you're not going to evolve into life and I have to agree with them there, but it's a complete non-sequitar because nobody outside of ID is saying that anything evolves into a part of the genome that supports life. Instead, the First Living Thing was in some not-too-improbable portion of the sweet spot and all of that organisms descendants have stayed in it. All of this NFLT and active information and what-not are just wrong. Embarassingly wrong. Paul Giem [236] "It is true that we haven’t found one, but to assume that the Garden of Eden, or even part of it, could survive the Flood seems to be stretching it." We have found traces of some things that existed prior to The Flood and after it. Egypt, for instance. Those Egyptians were building the pyramids during the time the flood is supposed to happen. The world wide Flood is another Babylonian myth that has been disproven. As for the fossils, it took about a century or more just to get believers to admit that yes, they really were the remains of extinct animals. Even the concept of extinction was hard to believe. Stupid design is stupid if you or I or some other mere human, can think of a better way to do it. There's a nerve in the giraffe that goes from the brain all the way down that long neck, loops around a bone and then travels all the way back up that long neck to the voice box, which is situated just below the brain. I would have run it direct. I think you would have too, along with just about everybody else. We're intelligent enough to do that. Presumably, an intelligent designer would have done the same. But not mindless Darwinian evolution - it's not smart enough to break the nerve and rejoin it on the right side of that bone, so as the neck lengthened, the nerve lengthened twice as much. Dumb for an intelligent designer, but about what you'd expect to see from time to time with mindless evolution. (Referring to the point mutation that invigorates Apolipoprotein AI milano) "...this misses the point. If one is given wild-type protein, one can get to milano protein with one mutation." "Now if you can create the wild-type protein with only a series of beneficial single mutations from some other protein, we will be quite impressed." Okay, look at the teosinte-corn evolution example I gave in [230]. Changing one gene makes the ears grow bigger and grow on short branches. Changing a second gene makes the kernels grow more rows of seeds. A third mutation softens the hard kernel casings, making them easier to eat. Three mutations, which occurred over a period of a few thousand years, made the ears larger, restricted them to growing near the stem and softened them to make them more edible. More mutations eventually turned teosinte, which looks like a clump of grass and whose ears are tiny, hard to harvest, rock hard and barely edible into all the varieties of modern corn. Better yet, this evolutionary event is under widespread study today. We have the original plant (it still grows wild in Mexico), we have archeological samples from various times during its evolution (don't know if we can get DNA from them, though) and we have all the modern corn plants. Even better, a method has been discovered that helps find the mutations that caused these changes. There are several labs hard at work finding and reporting these mutations and we will soon have a road map showing how teosinte became corn and approximately when and in what order the mutations occurred that did it. We're making new tissues here! And new traits! And we're going way beyond Behe's so-called "edge of evolution". And all in under ten thousand years! You wonder where the wild-type proteins came from? Well, with about three and a half billion years of evolution and over 500 million years of multi-cell evolution, I don't think there's any real mystery of how they were produced. We don't know the details, but we know the process and we have an example of its extreme power to create new forms of life. Micro-evolution becomes macro-evolution. You just have to let it run long enough and your clump of grass becomes corn. gpuccio [238]: (Regarding Apolipoprotein A-1 milano) "...its protective effect in vivo is usually assumed, but not really proved..." The mutant protein was discovered when people noticed that the descendants of an 1800's couple didn't have heart attacks. This happy condition stuck out like a sore thumb and sent medical people looking for the reason. A-1 milano turned out to be the reason. Yes, the original protein also removes cholesterol. But A-1 milano has a much stronger effect, strong enough to prevent heart attacks. So far as I know, no adverse affects have been seen and the family has been studied since the '80s. CJYman [240] "All you need is a specified pattern (many biological patterns are specified by functional events), an understanding of the probabilistic resources available (max. number of bit operations available to life on earth based on mutation rates), an approximation of specified patterns of same length (research in progress), and the measurement of “shannon information” (provided by Hubert Yockey and others). There is no assumption of anything to measure for a specification." The "measurement" comes when the new organism tries to make a living with its new DNA. If the organism lives and successfully reproduces, it passes, if not it dies and we don't worry about it until the next time this particular mutation occurs. This process is normally called "natural selection". jerry [243] "ID accepts everything that could be classified as micro evolution and your examples were within micro evolution." "Certainly new functions and capabilities have developed but while important for the species, medicine or other areas they are trivial for evolutionary biology which must explain how the very complicated information network in a genome arose and changed over time to specify these novel complex functional capabilities. As an aside, no one has been able to explain it to us here or in any written medium that we are aware of." What kind of explanation would you be looking for? A step by step, mutation by mutation explanation of how every base-pair in modern DNA came to be, starting with the first living organism? I think you're going to be pretty safe from challenges like that. But teosinte into corn? Ten years from now, we should be able to hand you a list of every mutation that converted the former into the latter and the approximate order they occurred in. Will that be enough? That's a pretty big change, after all. You basically can't eat teosinte without crushing the kernels with a rock to get at the edible innards. The organization of the plants is quite different - corn is tall and narrow with the ears near the shaft, teosinte is branched and the tiny ears are scattered all over the place. We've got different tissues being created, too. Teosinte has a hard kernel, corn has a very soft coating - two different tissues. Why is this not macro-evolution? And if they're not, could you please tell us what IS macro-evolution? gpuccio [249] "3) As the current explanations for that CSI in biological information by non design models are completely unsatisfying and empirically inacceptable" Translation: You don't believe CSI can be manufactured by non-intelligent processes. But that's just a belief. I believe differently. What facts can you cite to support your belief? What would it take to convince you that CSI could be manufactured non-intelligently? How many base-pairs do we have to manufacture to convince you? djmullen
gpuccio
And it is obvious, at least to me, that in the second case there are all the reasons to assume (and maintain) the hypothesis of a quasi-uniform distribution.
Exactly. We're only considering specific cases, not MET, since it obviously succeeds in certain cases, or all of biology in general. JayM 261 First You only succeed in continuing the "non-uniform argument" by changing the metric to things that seem nonsensical. You might as well have said, "I ignore what you just said, I ignore the pointed questions, and I prefer to focus on aspects of the discussion that are already acknowledged as being valid." Like:
It is impossible to get out of the “sweet spot” using MET mechanisms, because only offspring in that region survive to form the next generation. This is why I think this is such a rich area for ID research. It’s a clear prediction of the limitations of MET mechanisms. If the islands of viability are not connected, all the way back to the origin of life 3.8 billion years ago, MET doesn’t explain the evidence.
Agreed 100%. You say that "I am starting to feel a bit like a Behe-ist in a camp of Dembski-ists". And personally I prefer Behe's approach since it's grounded in real observations and Darwinists cannot retreat into hypothetical la-la land. But for the sake of readers I do not let this argument rest since Dembski does have a good point as well. Now back to the main disagreement. Which is that uniform probability distributions do not apply to biology at ALL in regards to MET. In fact, I would expect non-uniformity to be more complete in its breadth if some ID-compatible hypotheses are true. But to be fair we're assuming MET and Darwinism as a starting point in evaluating its claims.
My understanding as well is that they are equally likely, but MET mechanisms do not create genomes from scratch so the assumption of a uniform probability distribution across the entire genome is invalid.
I'm struggling to even comprehend your objection. Non-foresighted variation is produced by copying errors. How does the mere act of replication cause these copying errors at any particular point(s) in the genome to become unequally probable/non-uniform? Depending on the form of variation, and the error correction (human versus bacterium) they're certainly unequal in size or scope, but that does not change this observation. And at this point natural selection (the localized search you're so focused on) is irrelevant since it's after the fact and may not even be a component at all in calculations. Now if we have foresighted variation that seems to be outside of the bounds of MET (and said existence was predicted by ID proponents). Although some insist that the non-foresighted mechanisms of MET somehow produce the foresighted mechanisms we've observed. Whatever. In any case, these foresighted mechanisms have only been observed in limited instances, but as I've already noted if they operate in all aspects of biology that would support ID, not MET. You then restate the basis of MET in regards to micro-evolution which I fully comprehend. I'm NOT attacking this as I already stated in the 2nd paragraph of comment #258.
This is why Dr. Dembski’s papers, however interesting, do not appear to be directly supportive of ID theory.
And the ID theory of Dembski and Marks conflicts with micro-evolution how? Again, I'd rather Dembski and Marks make the connection to biology specific but my assumption is that their focus is narrow and they're largely constraining their argument to specific cases relevant to (a) macro-evolution via the unguided processes of MET and (b) OOL.
This would still not be a uniform probability distribution, because the probabilities of small changes given MET mechanisms is much greater than that of large changes.
You're changing the metric again. Yes, small changes are more probable. But your point is only relevant to the localized search. What exactly about the OCCURRENCE of unguided variation makes small or large changes to any specific location of the genome unequally probable? The indirect pathway is presumed to be composed of small changes since that is what is required by MET in order to reach long-range targets. And each step is considered independently in reference to the EF. (I've directly queried Bill about this so I'm pretty certain I'm correct in saying the following unless I've misunderstood somewhere along the line.) For a specific example, the Explanatory Filter can take multiple types of inputs (which also makes it susceptible to GIGO and thus falsification). Two are (a) the encoded digital object and (b) hypothetical indirect pathways that lead to said objects. My name "Patrick" is 56 informational bits as an object. My name can be generated via an indirect pathway in a GA. An indirect pathway in a word-generating GA is likely composed of steps ranging from 8 to 24 informational bits. Let's say you take this same GA and have it tackle a word like "Pseudopseudohypoparathyroidism" which is 30 letters and 240 informational bits. It can be broken down into functional components like "pseudo" (48 informational bits) and "hypo" (32 informational bits). Start with "thyroid" (56 informational bits). For this example I'm not going to check if these are actual words, but add "ism", then "para", and then "hypo". "hypoparathyroidism" is a functional intermediate in the pathway. The next step is "pseudohypoparathyroidism", which adds 48 informational bits. Then one more duplication of "pseudo" for the target. That may be doable for this GA but what about "Pneumonoultramicroscopicsilicovolcanoconiosis" (360 informational bits) or, better yet since it's more relevant Dembski's work (UPB), the word "Lopado­temakho­selakho­galeo­kranio­leipsano­drim­hypo­trimmato­silphio­karabo­melito­katakekhy­meno­kikhl­epi­kossypho­phatto­perister­alektryon­opto­kephallio­kigklo­peleio­lag?io­siraio­baph?­tragano­pterýg?n" (1464 informational bits). I'm not going to even try and look for functional intermediates. Here's the point related to biology. I forget the informational bits for the flagellum as calculated by Dembski (isn't that in No Free Lunch...it's been years since I read it?). But if anyone submits a reasonable hypothesis for an indirect pathway composed of steps which are less than 500 informational bits EACH then Dembski's design inference for that particular biological object is rendered invalid. I challenged Khan (previously linked on this page) based upon generous assumptions to name ANY single step in the pathway when you start from the T3SS as a functional intermediate and he could not do it, nor can anyone else I know of. There might be hypothetical indirect pathway but it seems likely to consist of steps with more than 500 informational bits based upon observation.
In addition, the calculation of CSI uses the length of the genome as a factor in determining how “specified” a sequence is. That is not logical given the small changes possible with MET mechanisms.
The calculation of CSI in biology does?? Can you be more specific in what you mean (or provide a reference)? Or did you mean the "length[complexity] of the specified sequence"? Dembski's specification is very general in its application. Which leads to the problem of subjective specifications. Art is subjective and requires an intelligent agent to identify it (although I suppose it can be argued that realist landscapes can be objectively specified on a comparative basis). But biology contains objective specifications independent of the observer based upon the machine functionality. That's why people refer to FCSI in regards to biology.
Viable regions are clearly clustered or you and I wouldn’t be here. We each differ from both of our parents, and yet we live. That demonstrates that there are multiple connected points in genome space that are viable. The question, of course, is how large these regions are and how connected.
You make my point for me, which is all I meant by saying "'can' be clustered". Patrick
JayM, it's a big tent :-) tribune7
Adel DiBano @264 Thank you for the kind words. I've been trying hard not to let my personal frustration with some aspects of the ID movement affect my presentation. These ideas are important. I am starting to feel a bit like a Behe-ist in a camp of Dembski-ists, so I doubly appreciate your note. We're not big enough to schism yet! ;-) JJ JayM
JayM: This thread is winding down, but I just want to put on the record that I have enjoyed your dispassionate and well-reasoned comments. You clearly understand our opponents better than many of us. I look forward to learning more from you. (As on the Durston thread.) ADB Adel DiBagno
UpRight Biped @262
My understanding as well is that they are equally likely, but MET mechanisms do not create genomes from scratch
Uhm, yes they do.
No, they create them from existing genomes by making typically very small changes (although some mechanisms like gene duplication or frame shifts are larger).
It is impossible to get out of the “sweet spot” using MET mechanisms
Then it should occur to you that there is no way to get into the “sweet spot” with them either.
That doesn't follow. MET is not a theory about the origin of life, but about how it diversifies once it exists. MET mechanisms operate on high fidelity replicators. Those replicators are already viable, by definition. MET mechanisms just keep subsequent generations in the viable region, whatever that may be. If we want ID theory to be taken seriously, we need to address MET as it is exists. Assumptions about uniform distributions and searches from scratch don't meet that criteria and make it easy for ID opponents to dismiss our arguments. JJ JayM
JayM
My understanding as well is that they are equally likely, but MET mechanisms do not create genomes from scratch
Uhm, yes they do. The core of MET is chance and mechanical necessity. NS is a condition of the two; without them there would be no NS, but without NS, chance and law would remain. And by the way, this forms the foundation of all approved thought on the matter. - - - - - - - - You make this comment just mere inches from where I suggested that MET is "a theory of Life allowed the luxury of not having to address itself to how the conditions of the theory got started to begin with." So, the argument can go on forever, given that nothing of the core claim need be addressed - but taken for granted instead. But even so, maybe headway is being made:
It is impossible to get out of the “sweet spot” using MET mechanisms
Then it should occur to you that there is no way to get into the "sweet spot" with them either. Upright BiPed
Patrick @258
Why a uniform probability distribution? That isn’t applicable to biological systems.
Depends on the metric, does it not? If you’re talking about the information encoding an object, then in this case we’re dealing with nucleotides and as far as I’m aware all 4 options (T, A, C or G ) are equally likely within biology.
My understanding as well is that they are equally likely, but MET mechanisms do not create genomes from scratch so the assumption of a uniform probability distribution across the entire genome is invalid. MET mechanisms start from a known viable state and explore nearby regions in genome space around that known viable point. Very few changes are made and those that are made are small. Only viable offspring, by definition, survive to become part of the next generation of the population. This is about as far from a uniform probability distribution across the genome as it is possible to get. This is why Dr. Dembski's papers, however interesting, do not appear to be directly supportive of ID theory.
This can be extended to a measuring JUST the steps in a pathway, since each step in the pathway consists of changes to nucleotides and need not be in reference to the entire search space.
This would still not be a uniform probability distribution, because the probabilities of small changes given MET mechanisms is much greater than that of large changes. In addition, the calculation of CSI uses the length of the genome as a factor in determining how "specified" a sequence is. That is not logical given the small changes possible with MET mechanisms.
It’s true that viable regions in the overall search space "can" be clustered, but that objection seems only relevant to natural selection and micro-evolution (local searches within a cluster of viable regions).
Viable regions are clearly clustered or you and I wouldn't be here. We each differ from both of our parents, and yet we live. That demonstrates that there are multiple connected points in genome space that are viable. The question, of course, is how large these regions are and how connected.
As you say:
MET mechanisms demonstrably channel information from the environment (simulated or natural) to the populations that are evolving. The question isn’t whether this happens, but how far it can go.
And when you get out into this purported void of deleterious and neutral spots in-between functional regions are they not all equally likely?
It is impossible to get out of the "sweet spot" using MET mechanisms, because only offspring in that region survive to form the next generation. This is why I think this is such a rich area for ID research. It's a clear prediction of the limitations of MET mechanisms. If the islands of viability are not connected, all the way back to the origin of life 3.8 billion years ago, MET doesn't explain the evidence. JJ JayM
Patrick: Some sense at last! Thank you. Sometime, somewhere, we will have to discuss more in detail the fundamental difference between calculating a search space for the emergence of a new protein from scratch and calculating the search space for a random transition without any selection of the intermediates: I don't know why anybody in the other field seems convinced that the second thing cannot be done. And it is obvious, at least to me, that in the second case there are all the reasons to assume (and maintain) the hypothesis of a quasi-uniform distribution. But again, that's not a discussion which can be made here, at 259+. gpuccio
Patrick:
And I think a lot of people are talking past each other in this thread since (A) we know micro-evolution via unguided processes works and (B) the real problem for MET is macro-evolution (never mind OOL, etc). So some people seem to be needlessly defending A while what ID proponents are focused on is B.
A willful (and well-fed) remnant, no doubt, of a theory of Life allowed the luxury of not having to address itself to how the conditions of the theory got started to begin with. A theory entitled to intellectual dogma without ever having to square up with the growing (not diminishing) mountain of evidence against its broadest claims. Special Theory = Yes, General Theory = No. Upright BiPed
Why a uniform probability distribution? That isn’t applicable to biological systems.
Depends on the metric, does it not? If you're talking about the information encoding an object, then in this case we're dealing with nucleotides and as far as I'm aware all 4 options (T, A, C or G ) are equally likely within biology. This can be extended to a measuring JUST the steps in a pathway, since each step in the pathway consists of changes to nucleotides and need not be in reference to the entire search space. I'm no mathematician but when you're talking about a long indirect pathway where the target is inactive for natural selection should not all functional intermediates (and that's assuming they exist for ALL long-range targets) be considered as equally likely in reference to the goal? Even contained within viable cluster there are still a very large number of neutral and deleterious spots, and in regards to JUST random variation (not natural selection or foresighted variation or some limited forms of variation[mutational hotspots]) I have to ask what makes things unequally probable/non-uniform? But you're focused on the overall search space and localized/clustered viable spots. It's true that viable regions in the overall search space "can" be clustered, but that objection seems only relevant to natural selection and micro-evolution (local searches within a cluster of viable regions). And I think a lot of people are talking past each other in this thread since (A) we know micro-evolution via unguided processes works and (B) the real problem for MET is macro-evolution (never mind OOL, etc). So some people seem to be needlessly defending A while what ID proponents are focused on is B. The major point of contention is (a) whether macro-evolution involves searching for isolated clusters and (b) whether in nature there exists fitness functions for all targets (aka how do you get squirrels to die often enough that the search is funneled toward becoming a flying squirrel). As you say: "MET mechanisms demonstrably channel information from the environment (simulated or natural) to the populations that are evolving. The question isn’t whether this happens, but how far it can go." And when you get out into this purported void of deleterious and neutral spots in-between functional regions are they not all equally likely? I'm not an expert, so someone please correct me if I'm wrong in my comment. In any case, I'm hesitant to try and connect Dembski and Mark's work to overall biology. I'd like to see them make the connection specific since I get the feeling they're focused on particular scenarios which are the difficult cases and not in general (aka obviously they're not trying to argue that unguided processes do not work at all!). Patrick
R0b: "I wish you would. How else can I correct my misunderstandings?" I am really tired of this thread. Too many things have been said, and too many misunderstandings have arisen. I have continued till now not to let some points completely truncated, and out of respect for some of the participants, like Mark and yourself. I have written # 249, which is truly long even for my standards, because I thought I owed you some clarifications, in name of our past constructive discussions. And as you may know, with all the huge respect and admiration I have for Dembski, and especially because of that, I usually don't comment on direct observations about his work. I stand for my points, and take full responsibility for them, and that's all. gpuccio
gpuccio:
I think that in your post you have also seriously misunderstood Dembski, but I will not comment on that.
I wish you would. How else can I correct my misunderstandings?
Thank you for the kind appreciation. But you probably know that what originates in a way can be greatly transformed through multiple passages. That’s also the basis for very popular games, I believe.
Indeed it is. My comments came out much more culpatory than I intended, and I apologize for that. I just wanted to make sure the JayM didn't get stuck with a strawman charge. R0b
CJYman @246
For my more detailed explanation of “specification” and an actual measurement, check out “Specifications (part 1), what exactly are they?” in the top left margin on my blog.
Your blog post seems to be aligned with the standard ID definitions. Unfortunately, I still don't see where you have shown how to calculate CSI for any real system. You also didn't address the questions I asked: Why a uniform probability distribution? That isn’t applicable to biological systems. What is exactly does “able to be formulated as an independent event” mean? How could this be done for even a simple biological system? There seem to be two pervasive points of confusion in discussions of CSI. The first is that the word "specification" itself suggests that one and only one solution need be considered. The fact that a particular gene codes for a protein that is part of a flagellum, for example, doesn't make it specified a priori. In a different environment, that protein might be part of something else. In the same environment, a different protein might do an equal or better job. The second point of confusion is that the specification is computed using the length of the genome or protein, with the implicit assumption that all genomes of that length are equally probable. That's simply not the case in biological systems. This is the uniform probability distribution assumption sneaking in again. Are you aware of anyone ever computing the CSI of a real biological component and publishing a detailed, step by step description of how he or she arrived at a specific value in specific units? I've looked for something like that since first reading about CSI, with no luck. I strongly suspect, in fact, that if two different people were to undertake this task for the same biological construct, their answers would be different. Unless and until CSI can be rigorously calculated, it can't be considered as evidence for design. In a very real sense, unless it can be calculated it has no actual meaning. JJ JayM
CJYman:
Furthermore, it seems that once Dembski began to formalize the concept of active information, Wolpert fell silent.
Wolpert addressed Dembski's work only once that I know of (correct me if I'm wrong), in 2002. Dembski started to formalize active information in 2006. Why would you connect Dembski's active info work with Wolpert "falling silent"? That's a rhetorical question. You seem to be simply repeating what Dembski said in an interview. R0b
CJYman @245
The fitness functions used by Tierra and the group evolving antenna designs at MIT have no knowledge of the solution nor of what the optimum performance will be (in the case of Tierra, the concept of optimal performance is not even applicable).
They don’t know what the exact solution will look like, however, they program the search space constraints and the relevant math used to determine what the constraints of the design will be.
They specify a fitness function that permits genomes to be ranked. This is a very simple simulation of the far more complex time varying fitness function we call "the real world."
IOW, they know the parameters which need to be provided by them (the intelligent designers) and the program which will work with those parameters in order to solve a specific problem.
That's not particularly accurate in the case of antenna design and not all all accurate in the case of Tierra which is measuring replicability. The fitness function is just a simulation of natural selection. The criteria used within it may have meaning outside the simulation, in the case of antenna design, but the interesting thing about GAs is that they show that even simple simulations of MET mechanisms can transfer information from the environment (the fitness function) to the population over generations.
If they had no idea they were looking for an antenna which was optimized according to their provided criteria, would the EA produce an optimized antenna which is exactly what they are looking for?
The GA will produce populations that have successively greater average fitness, as measured by the fitness function. The fact that the fitness function provides value outside the simulation, in the case of antenna design although not of Tierra, is immaterial. This is completely analogous to a superset of the very same mechanisms producing successively more fit populations as measured by the ability to survive and reproduce in the natural world. JJ JayM
R0b: "gpuccio is the origin of what you called a straw-man variant" Thank you for the kind appreciation. But you probably know that what originates in a way can be greatly transformed through multiple passages. That's also the basis for very popular games, I believe. gpuccio
Mark: I hope that my previous post answers your # 247, at least form my point of view. I know it is a little bit long anyway, so don't feel obliged to read it... And I think you can reach CJYman'blog by clicking on his username. gpuccio
CJYman:
I was actually not responding to gpuccio re: specification not being an example of circular reasoning. I was responding to JayM comment #232.
Yes, and JayM was responding to djmullen [229] who was responding to gpuccio [226]. gpuccio is the origin of what you called a straw-man variant.
Yes, because it would take some type of intelligence to “cheat” in that way.
Yes, that's the idea, but gpuccio seems to disagree, as he's looking for a "new specification".
However, that is based on a pre-specification as opposed to a specification.
I've never actually seen any example that meets Dembski's definition of a "prespecification", but that's neither here nor there. R0b
R0b: I think there is some misunderstanding here. Is it possible to ask that affirmations be read in the context in which they are made? The paragraph about specification which you cire is taken from my response to Mark, and is intended to explain why, in my view, the ID arguments are not only "negative", but do contain a very positive element in the concept of specification. Now, first of all I was not giving an operational definition of specification for application in design detection. That was rather a reflection on the philosophical meaning of the theory. And obviously, it was my personal reflection, and did not imply anyone else, least of all Dembski. I'll try to explain my personal thought better. For me, design as we see it in human artifacts is "always" the product of a conscious, intelligent process. That process can be observed both objectively and subjectively (by the designer). I think that should be simple. Human design comes from a conscious, intelligent process. Or do we disagree even about that? At this point, I wouldn't be surprised of anything. So, when I say that "specification comes from the conscious representation of an intelligent, conscious beings" I am referring to what we observe in human design. Probably, I could have been more analytical, and should have said: "I call specification the conscious representation, meaning and purpose projected by humans into an artifact in the process of design". Let's go on. The next step is to understand how we can recognize in the artifact that specification. Here, we have to give "operational" definitions of specification, in other words we have to define which formal properties, or aspects, of an object can be considered a sign of the design process in human artifacts. That is a vast field of discussion, and I don't want to discuss it here. We have discussed this many times, and many times with you. I will just remind you that I often give a very restricted, but effective (IMO) definition of a subset of specification, which is the functionality of a digital string of information. We could discuss about the definition of functionality, but again to avoid that we can focus on specific examples, which can be more easily understood and analyzed, like human made computer software, where the function can often be easily defined in a specific context. The concept that any function has to be defined in a context remains, for me, of fundamental importance. So, to sum up: if we have a piece of software which performs a well definable function (for instance, an ordering algorithm) in a digital environment, then we can see that function as a specification of that piece of digital information. Now, if we know for certain who wrote that software, it is easy to say that the specification (the function) is the product of a consciuos intelligent process (the designer's intention, programming, implementation, and so on). So, it is easy to say: we objectively see here a specification (a function), and we know that it was produced by a consious process of design by this designer. But if we see the object, and recognize a function in it, and assume a specification and therefore a conscious process of design, without having direct evidence of it, are we right? Here comes the quantitative aspect of complexity, necessary for design detection. Remember, we are still speaking of design detection in supposed human artifacts, here. And the complexity has essentialy one role: to rule out possible configurations of the digital information we are observing which, while suggesting an intentional function and therefore a process of design, could instead be the product of chance. That's what I called "pseudo-specifications". A high complexity (improbability) of the digital, functionally specified sequence we observe is a safeguard against the possibility that the observed function may be there as a product of chance: in other words, that it may be a pseudo specification, a function which is there, but has never been designed by any conscious agent. You may like it or not, but that improbability of the observed function (or of any other kind of specification) is exactly the main tool of the empirical science of design detection in human artifacts, as Dembski has repeatedly discussed in his works. The concept of CSI is just a way to define a threshold of complexity which can give us an "operational" definition of specified complexity, not very sensitive, but specific enough to empirically avoid any false positive. The practical application of all that is the EF. All the above is about human design, and nothing else. Now, let's extend the reasoning. Once we have given some operational definition of CSI to be used as a threshold in the EF, we can observe that: 1) CSI, as defined in our context, can easily be described in human artifacts, but not in all of them. As repeatedly stated, there are many designed artifacts which are not complex at all, but are designed just the same. These are the false negatives. Paul's example of the signal by the hanging lanterns is a perfect model of that. Obviously, if the artifact is not complex enough, we can identify it as designed only if we have direct infromation about the designer and/or the design process. But just by observation of the artifact, we cannot be reasonably sure that it is designed. 2) CSI cannot be found in natural objects, with only one known exception: 3) CSI is easily described in biological information, and especially in biological digital information (genome and proteome). Up to now, there is no assumption, no inference, no theory. We have just described what we observe, and given some operative definitions which can be applied to the observation and measurement of certain observed facts. Obviously, anybody can disagree with the descriptions and the definitions, especially with the operational definition of CSI. We all know well those controversies, and this post in no way pretends to analyze them here. But now comes the true, the only important inference in ID theory. Which is more or less the following: 1) As we know that CSI is a reliable sign of human design 2) As we can observe CSI in biological digital information (genomes and proteomes) 3) As the current explanations for that CSI in biological information by non design models are completely unsatisfying and empirically inacceptable 4) We make the inference that the observed CSI, and therefore the specification, observed in biological information could be explained as the product of a process similar to the process of human design, in other words as the product of a conscious intelligent process, implying intention, programming, implementation. That inference is the basis for an explanatory theory, which we call ID, and for all possible further research or discussion about that theory. Now, it's not my intention here to defend all the steps of that reasoning from the inevitable attacks, more or less civil, which will follow here: not in this thread, which is already 239 posts long, and not with the general atmosphere which has been predominating in the last part of it. I have made the above reasoning only for two motives: a) to clarify better my thought b) to show that the final inference in 4) can be liked or disliked, accepted or ferociously refuted, but is not circular. Finally, I owe you a last clarification about the pi example. You say: "According to his logic, we should not infer design from an extraterrestial signal that contains the digits of pi, since there is “no new specification” in it." I have never implied that. My example meant only the following: a) we have a software designed to output the digits of pi. b) after some time of calculation, the program outputs the first 10^200 digits of pi. That output is CSI: it is specified (the digits of pi are a mathemathical object which cannot be found in nature); and it is complex enough. If I received a signal of that type, I would infer CSI and design. c) Now the program goes on calculating, and after some time it outputs the first 10^300 digits of pi. What has happened? If we calculate quantitatively the CSI in the second output, the complexity has greatly increased: the result is much more improbable than the previous one, and so we would say that the computer program has created new CSI. Now, my point was simply that the program has created new complexity, but under the same specification (the digits of pi). Indeed, it has only applied repetitively the same algorithm. That is evidence, IMO, that it is the specification, and not the complexity, which is the true distinctive product of conscious design. A final disclaimer, which seems to be necessary at this point: all the above is only my personal view. I think that in your post you have also seriously misunderstood Dembski, but I will not comment on that. gpuccio
CJYman For my more detailed explanation of “specification” and an actual measurement, check out “Specifications (part 1), what exactly are they?” in the top left margin on my blog. This obviously addresses my comment #247. But where is your blog? Mark Frank
CYMman #235 I am afraid I cannot see how you have avoided. In #1 you say: "we observe intelligent agents creating specifications every day" but you have not defined specification. Meanwhile Paul (227) and Gpuccio (226) above appear to define specification as "patterns which are produced by intelligence". So we need a definition of specification that does not rely on intelligence. Otherwise all you are saying is that "we see intelligent agents creating things that are created by intelligent agents" You could use Dembski's most recent definition of specification which is roughly speaking: "pattern that can be produced with a very small computer programme". But in that case crystals and snowflakes are specified. And it is no longer evidence for design. Mark Frank
JayM, For my more detailed explanation of "specification" and an actual measurement, check out "Specifications (part 1), what exactly are they?" in the top left margin on my blog. CJYman
JayM: "The fitness functions used by Tierra and the group evolving antenna designs at MIT have no knowledge of the solution nor of what the optimum performance will be (in the case of Tierra, the concept of optimal performance is not even applicable)." They don't know what the exact solution will look like, however, they program the search space constraints and the relevant math used to determine what the constraints of the design will be. IOW, they know the parameters which need to be provided by them (the intelligent designers) and the program which will work with those parameters in order to solve a specific problem. If they had no idea they were looking for an antenna which was optimized according to their provided criteria, would the EA produce an optimized antenna which is exactly what they are looking for? CJYman
Hello Rob, I was actually not responding to gpuccio re: specification not being an example of circular reasoning. I was responding to JayM comment #232. Rob: "In fact, flat-out copying is an example of design according to Dembski." Yes, because it would take some type of intelligence to "cheat" in that way. Ie: I have a deck of cards "randomly" shuffled and so do you. We both reveal our decks to each other at the same time and discover that they are identically ordered. Is chance a best explanation or did someone "purposefully cheat?" However, that is based on a pre-specification as opposed to a specification. CJYman
djmullen, Just to keep you on the same playing field as most of the rest of the ID folks, understand some basics. There is nothing in the MET or whatever the latest version of it is in terms of processes that ID objects to. We do not object to the findings of modern evolutionary biology. ID accepts everything that could be classified as micro evolution and your examples were within micro evolution. Notice I said findings. I did not say conclusions. ID does not believe the findings of evolutionary biology since Charles Darwin and especially within the last 10 years support macro evolution in the sense of the construction of novel complex capabilities. Certainly new functions and capabilities have developed but while important for the species, medicine or other areas they are trivial for evolutionary biology which must explain how the very complicated information network in a genome arose and changed over time to specify these novel complex functional capabilities. As an aside, no one has been able to explain it to us here or in any written medium that we are aware of. We have had at least four evolutionary biologists here who have been anti ID and several biologists and none of them have been able to do it either. So what are we to think? So we repeat this short message and some will engage you on longer exchanges but the result is always the same. No one makes a crack in the argument. Few of us are scientists but many of us are well read and most of the people you engage take positions not based on religion but on the results of scientific studies. So you will have to up the level a lot to be productive here. And if your objective is to provide the logic to convince us we are deluded, you will be the first. The evolutionary biologists have always given us irrelevant stuff hoping we won't notice. But when we do notice, what does that say to the strength of the anti ID argument. As I said to you on another comment. Human nature is the biggest support we have for our position. jerry
CJYman @240
I think the most fascinating implication of the NFLT is that it seems to show that any simulation of evolution requires information about characteristics of the problem to be solved to be incorporated into the behavior of the algorithm. In fact, the authors pretty much state that in their conclusion. So far, no one has shown otherwise, and no one has shown that evolutionary algorithms will generate themselves absent previous foresight of a target to be reached (actual information of the future problem being used to program and match the search procedure and search space).
I must disagree with you here. In fact, most GAs are not of the Weasel variety. The fitness functions used by Tierra and the group evolving antenna designs at MIT have no knowledge of the solution nor of what the optimum performance will be (in the case of Tierra, the concept of optimal performance is not even applicable). MET mechanisms demonstrably channel information from the environment (simulated or natural) to the populations that are evolving. The question isn't whether this happens, but how far it can go. JJ JayM
CJYman @240
As to a specification being measurable, it seems we are just going to have to disagree over that.
I disagree. ;-) We don't have to agree to disagree, actually. If specification can be quantitatively measured, that can be demonstrated to anyone interested. If it can't be demonstrated, it's not reasonable to assert that it is quantifiable. To your credit, you seize the bull by the horns in your next sentences:
I do agree that not every single act of intelligence will be measurable as a specification, but that is no problem, since many aspects are measurable. A specification merely measures the pattern against a uniform probability distribution and then multiplies this by how many patterns of same length are specified (able to be formulated as an independent event) and then multiplies this by number of trials available. Take the -log2 and you have your value.
Why a uniform probability distribution? That isn't applicable to biological systems. What is exactly does "able to be formulated as an independent event" mean? How could this be done for even a simple biological system? You've hit on exactly my frustration when reading the ID literature on CSI. I don't see how to compute it, even in principle. That doesn't mean it's not computable, but I've dived (doven?) deeper into the available material than most people would be willing to, and I haven't found a clear explanation of how to apply it. Heck, I'm not even sure what the units of CSI are, from the definitions I've seen. Let's calculate some CSI! JJ JayM
JayM: "Thanks for the detailed breakdown and clarification of the argument. Too frequently I have seen it presented as “Humans generate CSI, biological systems contain CSI, therefore biological systems are the result of intelligence.” That form is, if not circular, certainly uncompelling without further proof that CSI cannot be created by other means." Well of course, some people probably state it that way as an extremely simplified version since it does take a bit of "paper" to explain the details. But, yes, I can see to an extent how the form you mention is quite lacking. JayM: "My first comment is that there are many ways of identifying intelligence. I’m not certain that the creation of specifications is a unique identifier, or even measurable." Yes, there may be many ways to identify intelligence, but you gotta start somewhere and if there is a measurable way, then that would be the best place to start. Specifications will be a unique identifier until the ID Hypothesis is falsified and someone provides evidence of chance and law absent intelligence generating specifications. As to a specification being measurable, it seems we are just going to have to disagree over that. I do agree that not every single act of intelligence will be measurable as a specification, but that is no problem, since many aspects are measurable. A specification merely measures the pattern against a uniform probability distribution and then multiplies this by how many patterns of same length are specified (able to be formulated as an independent event) and then multiplies this by number of trials available. Take the -log2 and you have your value. Properly understood, none of those are arbitrary measurements. They can be updated upon the presentation of new data and measured to a greater precision, but that is the same with any measurement. To date, no one has shown an example of a specification being generated absent consideration for future results. JayM: "I’m also not convinced that using specification for design detection is meaningful in biological systems. As has been pointed out by Mark Frank, MET mechanisms aren’t searching for anything in particular, they just, by their nature, explore nearby genome spaces." Yah, so ... I recommend you read through my comment #121 and 122 above to see how this idea of measuring a search is perfectly applicable to biology. JayM: "Considering the current result of that exploration to be a specification is too close to drawing bullseyes around bullet holes after shooting for my taste. When there are many possible outcomes, using the outcome we observe as the specification and noting how well reality matches it is unconvincing." All you need is a specified pattern (many biological patterns are specified by functional events), an understanding of the probabilistic resources available (max. number of bit operations available to life on earth based on mutation rates), an approximation of specified patterns of same length (research in progress), and the measurement of "shannon information" (provided by Hubert Yockey and others). There is no assumption of anything to measure for a specification. A few assumptions do come in to play to tie the measurement of a specification into an hypothesis, but those assumptions are not problematic at all as I've already explained in my last comment. Furthermore, the hypothesis is fully falsifiable, so there is no major problem as far as I can see. Sure there is more research to be done but that has never posed a problem for science. JayM: "I just re-read some of the NFL literature and, as with Dr. Dembski’s two papers that started this thread, they deal with evenly distributed solution spaces (objective functions). It’s hard to see how this corresponds to what we know about biological systems." NFLT explicitly deals with the matching of correct search space to correct search procedure in order to produce better than chance performance -- better than chance performance being measured against uniform probability distribution. I have already explained the connection of "search" to biology, but I think the most fascinating implication of the NFLT is that it seems to show that any simulation of evolution requires information about characteristics of the problem to be solved to be incorporated into the behavior of the algorithm. In fact, the authors pretty much state that in their conclusion. So far, no one has shown otherwise, and no one has shown that evolutionary algorithms will generate themselves absent previous foresight of a target to be reached (actual information of the future problem being used to program and match the search procedure and search space). My question is, "if all simulations of evolution simulate the basic process of reproduction, mutation, and culling, and they require foresight of the problem to be solved (problem specific information), then why would the natural process get by without foresight if it too consists primarily of reproduction, mutation, and culling?" Of course the power of non-foresighted mechanisms is testable by creating a program which generates patterns or bit strings and a filter based on background noise and an arbitrary collection of laws (set of laws chosen with no regard for future results). As long as that produces no active information (problem specific information), evolutionary algorithm, or specification then the ID hypothesis stands. JayM: "In fact, David Wolpert, one of the authors of the NFL theorems paper, rejects the idea that NFL is applicable to MET mechanisms." Then you are going to have to tell me what MET mechanisms are, since near the beginning of the paper he states and I quote: “Given our decision to only measure distinct function evaluations even if an algorithm revisits previously searched points, our definition of an algorithm includes all common black-box optimization techniques like simulated annealing and evolutionary algorithms.” ...and... "In particular, if for some favorite algorithms a certain well behaved f results in better performance than does the random f then that well behaved f gives worse than random behavior on the set all remaining algorithms. In this sense just as there are no universally efficacious search algorithms, there are no universally benign f (optimizations problems) which can be assured of resulting in better than random performance regardless of ones algorithm.” ...and... "First, if the practitioner has knowledge of problem characteristics but does not incorporate them into the optimization algorithm, then P(f) is effectively uniform. (Recall that P(f) can be viewed as a statement concerning the practitioner’s choice of optimization algorithms.) In such a case, the NFL theorems establish that there are no formal assurances that the algorithm chosen will be at all effective. Second, while most classes of problems will certainly have some structure which, if known, might be exploitable, the simple existence of that structure does not justify choice of a particular algorithm; that structure must be known and reflected directly in the choice of algorithm to serve as such a justification. In other words, the simple existence of structure per se, absent a specification of that structure, cannot provide a basis for preferring one algorithm over another ... The simple fact that the P(f) at hand is non-uniform cannot serve to determine one’s choice of optimization algorithm.” Furthermore, it seems that once Dembski began to formalize the concept of active information, Wolpert fell silent. BTW: I have written a bit about this subject on my blog. Just click on my handle and you can peruse through the topics in the left sidebar at the top of my blog. CJYman
CJYman:
The hypothesis that specification signals previous intelligence is not circular at all if stated correctly and not in a straw-man variant.
I would call gpuccio's variant a misunderstanding rather than a strawman. He certainly wasn't intending that it be knocked down. gpuccio said: Specification comes form the conscious representation of an intelligent, conscious beings. That’s why computers can create new complexity (see my example of pi calculation), but not new specifications. But then, why do we need complexity (that is, improbability) in the concept of CSI? It’s simple. it’s because if something is simple, it could “appear” specified, even if nobody has ever thought of outputting that specification. Inother words, specification could be a pseudo-specification. So it appears that the defining characteristic of specification, as opposed to pseudo-specification, is that it originates from somebody's mind. Under this definition, to infer design from specification is certainly tautological. But obviously gpuccio's usage of the term does not match Dembski's. BTW, gpuccio's pi example is interesting. According to his logic, we should not infer design from an extraterrestial signal that contains the digits of pi, since there is "no new specification" in it. The "new specification" requirement makes no sense in Dembski's framework, which defines specification as pre-existing or easy-to-construct patterns. In fact, flat-out copying is an example of design according to Dembski. R0b
Paul, thank you for providing answers which, franky, I felt no more motivated to give, given the general tone of the discussion. I agree with everything. I must say that one of the things which I most appreciate in you is that, while being an YEC (which I am not), your religious beliefs never, and I say never, influence your serene approach to ID questions. That is really no small virtue. About Apolipoprotein A-I Milano, I would like to add that, while its protective effect in vivo is usually assumed, but not really proved, given the very small number of carriers, its usefulness as a drug, even if still in the experimental phase, is much more likely, and I hope that it will be confirmed. What is much less certain instead, as can be seen from the abstract I pasted in my previous post, and from one or two other papers I found, is that the positive effetcs of Apolipoprotein A-I Milano as a drug be necessarily greater than those of Apolipoprotein A-I wildtype, since the "natural" form too has proven effective as a therapeutic tool. I paste here the conclusions from a recent review of the subject (Shah, Indian Heart J 2006; 58: 209–215) "Infusion of HDL has been shown to reduce atherosclerosis in animal models and preliminary proof from small studies in humans supports this novel therapeutic paradigm. However, a number of questions remain unanswered, such as the consistency and durability of the athero-protective effect of HDL infusion, the optimal dose and the frequency at which such infusions would be needed. Some other concerns are the relative efficacy of different forms of HDL (wild type Apo AI vs Apo A1 Milano containing HDL; plasma-derived versus recombinant HDL), long-term safety and, most importantly, whether such therapy can reduce adverse cardiovascular events in a cost-effective manner." But even is that were the case, your observation remains perfectly valid: it remains to be seen if the positive effects are not balanced by collateral effects. And anyway, it is obvious that Apolipoprotein A-I is only one of many variants of the wildtype protein, and like in many other genetic variants of proteins, some physiological functions can be slightly "tweaked" in the variant, with effects which can be negative or neutral, and sometimes can imply some protection against some pathological state (the example you made of resistance of the sickle cell carrier to malaria is perfect). Suggesting that the natural polymorphism of most proteins in our (and other) species is a real example of darwinian evolution is beyond my understanding. Obviously, I am not making these points for you, who know all that very well. And finally, we will see if this new "trait" really expands in the world population by the powerful effect of natural selection, and becomes a basic step towards the new human race. After all, we already have about 40 people here in Italy. gpuccio
CJYman @235
The hypothesis that specification signals previous intelligence is not circular at all if stated correctly and not in a straw-man variant.
Thanks for the detailed breakdown and clarification of the argument. Too frequently I have seen it presented as "Humans generate CSI, biological systems contain CSI, therefore biological systems are the result of intelligence." That form is, if not circular, certainly uncompelling without further proof that CSI cannot be created by other means.
1. Observation: we observe intelligent agents creating specifications every day. In fact, it is these specifications that basically define something as intelligent. And this is by no means circular reasoning. Tell me, how would you know that something is intelligent if it did not produce specifications, yet was mumbling incoherently and randomly bumping into walls? This is merely an observation of what basically defines what we determine to be intelligent (system able to utilize foresight — modeling the future and generating targets).
My first comment is that there are many ways of identifying intelligence. I'm not certain that the creation of specifications is a unique identifier, or even measurable. One of the reasons I consider Dr. Behe's "edge of evolution" to be more fertile ground for ID research than the mathematical approach of Dr. Dembski is that, even with a background in mathematics and after reading the relevant literature, I can't quantitatively determine the "specification" in any artifact. Specification requires a standard against which an artifact is measured. ID theory hasn't yet reached the point where the selection of the specification criteria is non-arbitrary. I'm also not convinced that using specification for design detection is meaningful in biological systems. As has been pointed out by Mark Frank, MET mechanisms aren't searching for anything in particular, they just, by their nature, explore nearby genome spaces. Considering the current result of that exploration to be a specification is too close to drawing bullseyes around bullet holes after shooting for my taste. When there are many possible outcomes, using the outcome we observe as the specification and noting how well reality matches it is unconvincing. If I'm representing the use of specification unfairly, please excuse me. I genuinely don't see how to measure it quantitatively nor how to apply it to biological systems.
2. The math behind measuring for a specification combined with NFLT (and furthermore, active info) seems to put a no-go on what chance and law will produce absent previous intelligence.
This is another common claim that I'm not sure is justified. I just re-read some of the NFL literature and, as with Dr. Dembski's two papers that started this thread, they deal with evenly distributed solution spaces (objective functions). It's hard to see how this corresponds to what we know about biological systems. In fact, David Wolpert, one of the authors of the NFL theorems paper, rejects the idea that NFL is applicable to MET mechanisms. My view, therefore, is that specification needs to be more rigorously defined and applied, and that papers like Dr. Dembski's two newest need to be explicitly linked to biological systems before specification can be used for design detection. I suspect, although I'll be happy to be proven wrong, that investigation into the limits of MET mechanisms and the distribution of viable regions in genome space are more likely to generate strong evidence in favor of design in the near future. JJ JayM
djmullen (#228) Your comment is, as JayM stated in #231, "somewhat brutally stated". Unfortunately, it also misunderstands the subject badly and overstates the evidence enough that it is not likely to be of much persuasive value. If you spend most of your life in an echo chamber, you may not realize this. So perhaps I can give you a few pointers. First, every time you accuse ID advocates of being stealth YEC's, you turn everybody off. There are some YEC's here; I am one of them. But I am not exactly stealth, and those who disagree with me on this (and there are many, including the person who started this blog) really resent being told that they don't know their own minds. This is strawman arguing of the worst sort; building the strawman while everyone tells you it is a strawman. There is an anachronism in your presentation which betrays your inability to hear ancients on their own terms. You say, "We observe some facts, such as the great numbers of species. Then a Hebrew writer borrows a Babylonian theory, modifies it a little and ... that becomes for the moment the “best explanation”. The "great number of species" was arguably not known before Linnaeus. In fact, Linnaeus believed later in life that he had defined "species" too narrowly, and many creationists (I am now speaking of YEC's, or at least YLEC's, not ID advocates) agree with him. Creationists may be wrong, but this is not the way to go about proving it. Secondly, your paragraph,
There turns out not to have been a Garden of Eden, fossils are found of strange plants and animals that do not exist today, today’s species fall naturally into a heirarchial organization, there are severe moral and intellectual problems with the existence of the Intelligent Designer named God, a lot of the details of the designs are mind-bogglingly stupid, etc.
which is meant to be hard-hitting, misses the mark in a number of ways. First, you simply overstate your case when you say that "there turns out not to have been a Garden of Eden". It is true that we haven't found one, but to assume that the Garden of Eden, or even part of it, could survive the Flood seems to be stretching it. It would either have to have been removed by a miracle, in which case there is no particular reason for it being on the earth, or it would have gotten buried in the Flood. Now, if you had argued that "no physical evidence for the Garden of Eden has been found," you would not have been overstating the evidence, and you might have left some people with questions to mull over. But when you make statements that you clearly cannot prove, you tempt people to dismiss you as a crank. As far as strange fossils go, who says that all life forms, especially all species, that have ever existed must exist now? That is a rather gross mischaracterization of creationist theory. Today's species falling naturally into a hierarchical organization has been recently disputed here, and while it may be convincing to your buddies, its strength as an argument is much less here. And stupid design is only stupid if someone can do better, and if you don't know how to do better, you might want to be a little more careful about accusing the original designer of incompetence. Perhaps your best argument is that there are moral and intellectual problems with God. But be careful even here. For if your best arguments are here, and the scientific evidence for design is strong, you risk making this a science versus religion argument with you on the side using religious arguments. You talk about Darwinian evolution as being "both naturalistic and simple." The problem some of us see is that it is too simple, and naturalism is only an advantage if one starts with the presupposition that naturalism is good and even modified supernaturalism is bad. That sounds suspiciously like a religious commitment. In #229 you say, referring to Apolipoprotein A-I milano,
No consciousness or intelligence whatsover is required to make this complex, specified protein. A single base pair mutates and the specification is followed automatically and mechanically.
but this misses the point. If one is given wild-type protein, one can get to milano protein with one mutation. But very few people here would insist that single mutations cannot arise and be helpful in certain environments. Sickle cell trait is a case in point. Now if you can create the wild-type protein with only a series of beneficial single mutations from some other protein, we will be quite impressed. And you have not established even that milano is in fact superior. For all we know, too low cholesterol may predispose to cancer or spontaneous abortion or otherwise be injurious to health. The wild-type protein may very well be a better protein to have if one is not sedentary and eating the standard western diet. You need to be more tentative about your conclusions. Your comments about teosinte (#230) are addressing a problem that, AFAICT, does not exist. The vast majority of ID advocates, myself included, have no problem with single mutations, or progressively improving series of single mutations, being a pathway from teosinte to modern corn. I know that PT is bad at stereotyping ID, but I didn't think it was that bad. Besides, from a creationist perspective, maybe corn was the original and teosinte is the degenerate form, in which case we are just recovering the original information. Use some imagination. Or are you wishing to use an argument from personal incredulity without even trying to understand the other side? Periodically someone will burst upon the scene here spouting platitudes from Panda's Thumb as if they are Gospel, seeming to think, like some street preacher, that we just haven't been given the straight truth and that if you just hit us between the eyes, our belief systems will all crumble and we will be converted. I hate to tell you this, but it will be a whole lot more work than that. On the other hand, maybe that's the effect you want. You can claim that we "just won't listen to reason" and go back to feeling smugly that you have witnessed to the incorrigible, who can now be safely damned, or at least fired in good conscience. But if you want to actually engage, you have a lot of work to do. Paul Giem
JayM (#232), The hypothesis that specification signals previous intelligence is not circular at all if stated correctly and not in a straw-man variant. 1. Observation: we observe intelligent agents creating specifications every day. In fact, it is these specifications that basically define something as intelligent. And this is by no means circular reasoning. Tell me, how would you know that something is intelligent if it did not produce specifications, yet was mumbling incoherently and randomly bumping into walls? This is merely an observation of what basically defines what we determine to be intelligent (system able to utilize foresight -- modeling the future and generating targets). 2. The math behind measuring for a specification combined with NFLT (and furthermore, active info) seems to put a no-go on what chance and law will produce absent previous intelligence. It seems that specifications most probably (as a best explanation) will not be produced by chance and law absent intelligence. And again, there is no circular reasoning here. This merely extends the math into a type of no-go theorem, which can be stated informally as "unless there is an infinite regress of active information, chance and law absent intelligence will not produce specifications." 3. Combine #1. (observation) with #2. (no go theorem) and you have created an hypothesis that a specification will reliably signal previous intelligence. 4. We flow from a definition of intelligence (foresight) and an observation of what foresight produces, to a no-go theorem, to an hypothesis. The only assumptions that need to be made is that there is most likely not an infinite regress of active information (an infinite bias toward our universe, life, and the evolution of intelligence) and that other humans do indeed possess foresight so that we can conclude that the specifications which they produce are indeed associated with their foresight since I know as a fact that my foresight is essential in me producing specifications such as this conversation. As an aside, the first assumption can even be tested somewhat to see what chance (background noise) and law absent intelligence (a set of laws arbitrarily chosen without consideration for future results) will produce. If this random configuration does not produce active information, then I see no reasons why chaotic and possibly random fluctuations on the quantum scale would have a bias (active information caused from the correct search procedure being matched with the correct non-uniform probability distribution) toward our universe, life, and the evolution of intelligence. If you still believe that there is circular reasoning involved as opposed to arguing from observation and math to hypothesis, then please point out the circular argument. CJYman
"But the people who believe in the “God in six days” theory won’t let go. 150 years later, they’re still clinging to their old, superseded theory. They have no genuine new evidence to support their theory and mostly attack evolution. Indeed, they often don’t seem to even understand the new theory they are criticizing, but they insist on being taken seriously. It’s rather frustrating for all involved." What utter nonsense. This and some of your succeeding posts indicate your ignorance on the debate. You make a lot of unsupported assertions and act like they are true and then reveal your naiveté by bringing up irrelevant examples. Have at it my friend because you are making us look good. The specification argument is not circular. The only two places specification is found are in human intelligence and in life. It is found no where else. No natural process has ever been seen creating specification while it is seen in human activity and possibly some animal activity. The whole design argument is based on probability of cause not a logical certainty. If one sort of activity can cause something and another sort of activity has never been seen creating this result then it strengthens one over the other as the likely answer. Talk about utter faith. Anyone who believes that natural processes create specification is basing it on nothing and then we get the lecture on how this faith based epistemology now rules the day. As I said what utter nonsense. It shows more about the people who believe in it then it does about truth. jerry
Cardiovasc Diabetol. 2007; 6: 15. "Gene transfer of wild-type apoA-I and apoA-I Milano reduce atherosclerosis to a similar extent" Corinna Lebherz,corresponding author1,3 Julio Sanmiguel,1 James M Wilson,1 and Daniel J Rader2 Background The atheroprotective effects of systemic delivery of either apolipoprotein A-I (wtApoA-I) or the naturally occurring mutant ApoA-I Milano (ApoA-IM) have been established in animal and human trials, but direct comparison studies evaluating the phenotype of ApoA-I or ApoAI-Milano knock-in mice or bone marrow transplantated animals with selectively ApoA-I or ApoAI-Milano transduced macrophages give conflicting results regarding the superior performance of either one. We therefore sought to compare the two forms of apoA-I using liver-directed somatic gene transfer in hypercholesterinemic mice – a model which is most adequately mimicking the clinical setting. Methods and results Vectors based on AAV serotype 8 (AAV2.8) encoding wtApoA-I, ApoA-IM or green fluorescent protein (GFP) as control were constructed. LDL receptor deficient mice were fed a Western Diet. After 8 weeks the AAV vectors were injected, and 6 weeks later atherosclerotic lesion size was determined by aortic en face analysis. Expression of wtApoA-I reduced progression of atherosclerosis by 32% compared with control (p = 0.02) and of ApoA-IM by 24% (p = 0.04). There was no significant difference between the two forms of ApoA-I in inhibiting atherosclerosis progression. Conclusion Liver-directed AAV2.8-mediated gene transfer of wtApoA-I and ApoA-IM each significantly reduced atherosclerosis progression to a similar extent. gpuccio
djmullen @229
gpuccio [226]
Specification comes form the conscious representation of an intelligent, conscious beings.
That puts your result, ‘There is an intelligent and therefore conscious designer,’ into your premises, creating a circular argument.
Well put. The whole notion of "specification" is one of the reasons I responded to Dr. Fuller's post in another thread. Design detection, by its very nature, must make certain assumptions about the nature of the designer and must generate additional knowledge of the designer. When ID proponents ignore this inherent connection and refuse to discuss the nature of the designer, it opens the ID movement up to charges of being disingenuous. It also results in the type of circular reasoning you identified. Specification is asserted to only originate from intelligence, but that is the very thing we're supposed to be proving.
It’s also an argument that is easily shown to be wrong. Look again at the example of Apolipoprotein A-IMilano I gave in message 205. This protein is wonderfully specified: it cleanses plaque from the arteries and prevents heart attacks. Yet it’s caused by a simple point mutation that changes an arginine to a cysteine at position 173. No consciousness or intelligence whatsover is required to make this complex, specified protein. A single base pair mutates and the specification is followed automatically and mechanically.
Thank you for the example. This is why I think the fertile ground for ID research lies in determining where the edges of evolution actually lie. Certainly some point mutations can have remarkable consequences. The question is, how far can the MET mechanisms get? A related question is, how large is the connected region of viability in genome space? These questions have objective, reproducible, and therefore scientific answers, which have the potential to provide direct support for ID theory. That is something that, sadly, I still don't see Dr. Dembski's two papers, however interesting, capable of. JJ JayM
djmullen @228
But the people who believe in the “God in six days” theory won’t let go. 150 years later, they’re still clinging to their old, superseded theory. They have no genuine new evidence to support their theory and mostly attack evolution. Indeed, they often don’t seem to even understand the new theory they are criticizing, but they insist on being taken seriously.
Your overall message, while somewhat brutally stated, isn't a bad summary of why modern evolutionary theory is accepted by the overwhelming majority of scientists. Unfortunately, I am forced to agree that many ID proponents have a very limited understanding of MET specifically and science in general. The amount of painstaking, detailed, and aggressively reviewed work done by scientists every day is truly phenomenal. The fact that some ID proponents dismiss it out of hand reflects badly on the ID movement. On the other hand, your insinuation in this one paragraph is as overbroad as those I've complained about by people on "my side." Not all ID proponents are young earth creationists. Most that I have interacted with personally are not (as far as I can tell). I'm rather surprised that YECs are willing to enter the "big tent" of ID at all, given that their worldview is fully faith-based and does not require science for validation.
It’s rather frustrating for all involved.
It's just as frustrating for those of us on the ID side to see ID proponents discrediting the movement because of a lack of understanding of MET. We need to be even more rigorous than our detractors if we are to be taken seriously. JJ JayM
Several people have claimed that mutations cannot create new traits or functions in organisms. Take a look at this site: http://www.koshland-science-museum.org/exhibitdna/crops02.jsp It shows three of the mutations that converted teosinte into modern corn: "For example, a gene on chromosome #1 causes the ears of corn to be big and to grow on a few short branches. In contrast, the ears of teosinte are scattered over many small branches. A gene on the second chromosome causes more rows of kernels to grow, yielding more food per corn plant. A gene on the fourth chromosome causes corn kernels to have small, soft casings. Teosinte kernels have much larger, harder kernel casings that make them hard to eat." The article includes picturs of the small, hard kernels of teosinte and modern corn. Scroll down from here: http://www.matrifocus.com/LAM06/corn.htm and see a picture of a teosinte bush and compare it to a modern corn plant. Random changes can very definitely create new traits and functions. djmullen
gpuccio [226] "Specification comes form the conscious representation of an intelligent, conscious beings." That puts your result, 'There is an intelligent and therefore conscious designer,' into your premises, creating a circular argument. It's also an argument that is easily shown to be wrong. Look again at the example of Apolipoprotein A-IMilano I gave in message 205. This protein is wonderfully specified: it cleanses plaque from the arteries and prevents heart attacks. Yet it's caused by a simple point mutation that changes an arginine to a cysteine at position 173. No consciousness or intelligence whatsover is required to make this complex, specified protein. A single base pair mutates and the specification is followed automatically and mechanically. djmullen
gpuccio [222] "For once, I will make an analogy: we onserve some facts, like objects falling, etc. Then one Newton comes, and he creates a mathemathical theory which can well explain that. For a moment, let’s forget that it is a theory based on necessity, and let’s concentrate only on the epistemological aspect: it is an explanatory theory of some kind. Well, in absence of any other explanatory theory, or even of any other “credible” explanatory theory, that becomes for the moment the “best explanation”. But if another good explanatory theory is suggested, we have to choose between the two." An excellent point. We observe some facts, such as the great numbers of species. Then a Hebrew writer borrows a Babylonian theory, modifies it a little and in absence of any other explanatory theory, or even of any other “credible” explanatory theory, that becomes for the moment the “best explanation”: An Intelligent Designer named God created every species in six days about 4000 BC. This theory explains all of the evidence and it reigns for over 2000 years. Then new evidence is found that doesn't fit the old theory. There turns out not to have been a Garden of Eden, fossils are found of strange plants and animals that do not exist today, today's species fall naturally into a heirarchial organization, there are severe moral and intellectual problems with the existence of the Intelligent Designer named God, a lot of the details of the designs are mind-bogglingly stupid, etc. Then, about 150 years ago a new theory is proposed by Darwin. His theory accounts for the heirarchial organization of species, both modern and ancient, doesn't need a Garden of Eden or an Intelligent Designer (who is incredibly more unlikely than all of the proteins in all of the species combined), explains naturally the existence of diseases and hyper-nasty plants and animals, etc. and does it all with a theory that is both naturalistic and simple. This is a new and credible theory and we have to choose between it and the old one. The new theory is gradually adopted by just about every scientist in the world, easily encompasses new discoveries such as how genes work and 150 years later reigns triumphant. But the people who believe in the "God in six days" theory won't let go. 150 years later, they're still clinging to their old, superseded theory. They have no genuine new evidence to support their theory and mostly attack evolution. Indeed, they often don't seem to even understand the new theory they are criticizing, but they insist on being taken seriously. It's rather frustrating for all involved. djmullen
Mark Frank (#224) gpuccio has answered you, but let me give it another try. Your claim that "this information is defined in terms of the improbability of the outcome given chance" is probably true, if you want an operational definition. But it actually misses the point of the information itself. Let me give you an example of what I mean. At a certain point in American history, the need was felt to communicate the actions of the British army in Boston. An agreed upon signal was that if one lantern was hung in the bell tower of a church (apparently the Old North Church), the British were coming by land, and if two lanterns were hung, they were coming by sea. The number of lanterns thus gave a binary message (or perhaps trinary--if no lanterns were hung, they were not coming). In any event, it was necessary for the sender of the message to be in control of how many lanterns were hung; otherwise the message would not have been reliable. It did not matter, for purposes of communication, whether no lanterns were customarily hung, whether one lantern was customarily hung, whether two lanterns were customarily hung, whether there was some kind of regular rotation (two lanterns on Sundays, for example), or whether the number of lanterns was random or semi-random. All that was necessary was that the sender be able to control the number of (lighted) lanterns hung. Now, to conceal the design from the British, it would help if there were a random or semi-random number of lanterns hanging usually, as the hanging of one or two lanterns then would have aroused less suspicion. But even if normally there were no lanterns hanging, without some kind of reason to connect the number of lanterns with British troop movements, it would have been hard to make that connection. This is a nearly perfect example of undetectable intelligent design. But notice that what makes it intelligent design is not its improbability. It is its controlled conformation to a specific code which specifies a message. Without that code, you haven't got a clue as to what the message is, or even if there is a message. And without control, you have no way of being confident that the message received is the message that the sender wanted to give. This example illustrates that the essence of intelligent design is not complexity (how much more simple can you get than control over one bit of information?). It is specification. Detectable design is a whole different matter. If semi-randomly one or two lights was hung in the tower, and the person hanging the lights this time was the usual person to hang lights, the finding of one or two lights in the tower, with the attendant message getting out, would have given the British no reason to suspect that their intentions were being broadcast. On the other hand, if a messenger were caught with a message stating, "The British Army is coming by sea" (or by land), the army officers would probably have made the appropriate design inference and the bearer of the message would have found himself on the end of a noose, with the message not reaching its intended recipients. Our problem is not, by and large, that of the colonists trying to send a message. It is rather that of the British army trying to detect a message. Towards that end, we have to be careful of being too suspicious, as we can then easily become paranoid. But if we find a message written in English, saying "The British are coming by sea", we do well to make the appropriate design inference. Even if we find one we cannot understand, we do well to see if it is in English code, or in French or German. If it is in Iroquois, we might still miss the significance of the message. Getting back to the issue of design in biology, the problem is entirely reversed. The appearance of design is obvious even to those who would wish to deny design. Dawkins comes to mind, as do George Gaylord Simpson and Frances Crick. So that criterion of design is already fulfilled. The only remaining question is, can some other process explain the facts equally well, or even remotely equally well? If the answer is no, then the design hypothesis is by far the best-supported one. Saying that design is simply the negation of law-guided and chance-produced outcomes is only true of detectable design, not design itself. And it does not capture the essence of design. Design itself has to do with intending one event to happen and not another, coordinated with some other aspect of reality (broadly defined to include mathematics and aesthetics). That's why the "functionally" is important in functionally specified complex information. That's why, if you were able to demonstrate a pathway from other proteins to the flagellum where each step was in some real sense, in some real environment, an improvement on the last one, you might still fail to convince everyone that there really was no design. Many would abandon ID (if you could do that for the OOL I would), but many of those would go the theistic evolution route, contending that the universe was set up so that life would evolve, and that life was planned after all. But right now mechanistic evolutionists do not have even that. All they have is promises that such a pathway will someday be found. That's a lot of faith, brother. Forgive me if I do not have such faith. Paul Giem
Mark: I hope Jerry can ghange his mind! I think there is still a basic misunserstanding about design, which I tried to suggest in one on my posts in this thread (I can't remember which). In the concept of CSI, the true mark of design is specification, not complexity. I will be more clear: designed things are always specified, but not always complex. Specification comes form the conscious representation of an intelligent, conscious beings. That's why computers can create new complexity (see my example of pi calculation), but not new specifications. But then, why do we need complexity (that is, improbability) in the concept of CSI? It's simple. it's because if something is simple, it could "appear" specified, even if nobody has ever thought of outputting that specification. Inother words, specification could be a pseudo-specification. Seeing forms in clouds is just an example. So, complexity (improbability) is added to the filter to avoid those false positives. And complexity acts as a negative filter. But the specification, that is a very positive connotation. That is the sign of consciousness, of intelligence, of meaning. The fact that pseudospecifications may happen, when complexity does not act as a safeguard against them, does in no way change the fact that true specification (a true meaning inputted by a conscious agent, and recognizable, in the right context, by other conscious agents) is the positive sign of design. gpuccio
Re #221 Gpuccio I fear Jerry does not share your opinion of me! Mark Mark Frank
Gpuccio Actually my point is more empirical than you may realise. I am actually talking about observable evidence. The argument for design seems to hinge entirely on a particular definition of information. But this information is defined in terms of the improbability of the outcome given chance. So if that improbability is not true - then there seems to be no other evidence for design. What other observation can you make to support design that does not rely on the improbability of a chance explanation? Mark Frank
Mark: I forgot: you say: "A designer of undefined power and motives is always the perfect explanation for everything and might well have designed things to look like Darwinism." You know that that has never been my point, and never will be. I declare officially that I will never use that argument, and that if others in ID use it, I will openly criticize them! What can I do more? :-) gpuccio
Mark: you are never satisfied: we have just given you a way to falsify ID, and you still complain! :-) As often, I don't agree with your last point. The absence of a consisten non design explanation of biological information is one of the strongest points of ID, but that does not mean that ID is only a negative theory. Let's put it this way: 1) A design scenario can explain well biological information but 2) If a non design scenario can explain it equally well, the affirmation that design is the "best explanation" will very much lose strength, unless it can be based on other kinds of evidence. A constant difference between us is that you tend more to think in purely logical terms, while I prefer to think in empirical terms. Empirical science is not a branch of pure logic. For once, I will make an analogy: we onserve some facts, like objects falling, etc. Then one Newton comes, and he creates a mathemathical theory which can well explain that. For a moment, let's forget that it is a theory based on necessity, and let's concentrate only on the epistemological aspect: it is an explanatory theory of some kind. Well, in absence of any other explanatory theory, or even of any other "credible" explanatory theory, that becomes for the moment the "best explanation". But if another good explanatory theory is suggested, we have to choose between the two. And there never will be complete consensus about which one is the best. Scientific consensus is only the temporary opinion of the majority (or, sometimes, of those who have morfe power). But would you say that a theory of gravity, like Newton's theory, is purely negative only because it needs to be the best explanation? Any eory can meet a sudden crisis if it encounters a better theory as its competitor. The same is true for darwinian evolution and design, which at present are practically the only two competing theories to explain the causal origin of biological information. One must choose between them. And the choice is personal, and will never be a question of majority consensus, of authority, or of final evidence. I really don't believe in final truth in science. Even our scientific approach, in the end, is a question of free choice. gpuccio
Jerry: I think it's a problem of mnmotivation. Some come here to discuss, others to disturb. We can accept both, but there are limits to the exchange you can have with someone who only wants to disturb. I am never unhappy with others thinking differently from me. We have witnessed many serious and creative interlocutors here (Mark Frank and Sal Gal are perhaps the most recent which I would cite). They have different opinions, and are very sincere and precise in expressing them. I never want to convince anybody, but I really enjoy intellectual confrontation, even if passionate, especially if passionate. But intellectual confrontation it has to be. gpuccio
Mark --I don’t think the evidence for ID would be changed one jot. A designer of undefined power Mark, you are correct in that God can always be invoked but ID does not invoke God nor does it attempt to describe the nature of the designer. What you are describing is a means of falsifying a method of ascertaining design (IC), which is an aspect of ID. tribune7
So I am interested to know what evidence for ID, if any, would remain if Darwinism was established as plausible. If you take away the evidence against Darwinism - what’s left?
First off, the major qualification is whether a particular process operates uniformly for all problems. There may be a subset of FCSI, or just plain 'ol CSI, that is within reach of unguided processes. If so, ID theory would have to adjust to that. But if it's not operating uniformly, and this can be shown, then, no, ID theory in relation to MET would not be falsified in its entirety. Second, I'd make the distinction between "Darwinism" and MET. Darwinism in its common, improper usage usually denotes a dedication to the concept that life in its entirety does not rely on any intelligent interaction in any form. So in that case, the grand claims of MET would be vindicated but ID would still stand, with the remaining ID-compatible hypothesis being intelligent evolution. What justification remains? OOL and chemical evolution, which Darwinism in a general sense overlaps. In this hypothetical scenario MET only becomes probable based upon the rules and conventions of the information-based replicator. And if anything this case for ID is easier to make. But since this conversation has already showcased just how weak the case for Darwinism and certain claims of MET (but not micro-evolution!!!) really is I don't see how considering this scenario would help. Patrick
gpuccio, I welcome the nonsense we get here even if it is frustrating. It shows two things First, the level of information that the anti ID people think they can bring against ID. It is pretty pathetic. Second, the character of the individuals who do it. To me this is most interesting thing about the debate. Why do they do it? IQ wise they are not dumb but they sure can act the part really well when they try to debate here. The real audience are the lurkers who come here to get information so trying to convince those who support the Darwinian paradigm for everything is useless. They are not going to change since this commitment is based on ideology which must be defended at all cost. As opposed to many of us who would accept the Darwinian paradigm if there were evidence for it and thus, our position is not based on ideology. We have the advantage since if someone shows us some facts and good reasoning we will agree with them. And they paint themselves into absurd corners as they defend something their ideology requires. jerry
"It odd how powerful metaphors can be. “Hurdles” imply barriers on a route to somewhere. Evolution does not have to find or get over hurdles. It just exploits opportunities to go to somewhere completely unplanned." The metaphor is a teaching mechanism. The hurdles metaphor indicates that there are some places naturalistic processes just can not get to. You seem to quibble over the obvious. Are you trying to understand anything? jerry
gpuccio @214
Isn’t anybody else tired of reading “inconsistent and intellectually frustrating” posts from a couple of people? I am.
I am as well. The kind of unsupported, overbroad assertions made by some long time participants here do the ID cause no favors. JJ JayM
Re #211 Gpuccio You can probably see where I am going. As you know, it is my belief that ID is entirely negative. It only comprises evidence against Darwinism. So I am interested to know what evidence for ID, if any, would remain if Darwinism was established as plausible. If you take away the evidence against Darwinism - what's left? Re #212 Tribune - my post was a question not a statement. As it happens I don't think the evidence for ID would be changed one jot. A designer of undefined power and motives is always the perfect explanation for everything and might well have designed things to look like Darwinism. Mark Frank
Isn't anybody else tired of reading "inconsistent and intellectually frustrating" posts from a couple of people? I am. And yes, this is a totally unsupported claim. Interested people will have to read the posts themselves to judge. gpuccio
Jerry If you give the MET a very low bar to hurdle it will succeed but if it has to find some medium or high hurdles it has a tough time getting over them. It odd how powerful metaphors can be. "Hurdles" imply barriers on a route to somewhere. Evolution does not have to find or get over hurdles. It just exploits opportunities to go to somewhere completely unplanned. Mark Frank
Mark Frank -- Suppose that biologists eventually produce such a model and were consistently able to produce such models for favourite ID examples such as the cliched flagellum. Mark, you have made the point that ID is falsifiable and hence legitimate science. Well done :-) tribune7
Mark (#204): That would certainly be a very hard blow for ID, and would make the darwinian theory much more credible. But not necessarily true. There would probably be still much space for discussion. But it would certainly change things very much. gpuccio
Mark Frank @202
All this talk of islands of function in the sea of possible proteins (or DNA) made me think that it would be more appropriate to talk of evolution exploring the functional space rather than searching it. Evolution is not looking for something. It is stumbling across things that are useful or just different (genetic drift). I don’t think it makes the maths any more or less relevant but it is a more appropriate metaphor.
This is an excellent point. Many ID proponents think naturally in along telic lines. We need to remind ourselves that the only goals within the MET framework are immediate reproductive success. MET mechanisms do, in fact, explore very limited regions of viability without concern for an end goal. Unfortunately, this makes it more difficult to tie Dr. Dembski's papers (the original source of this thread, if I remember correctly) to the ID controversy. The search that he discusses isn't obviously analogous to the exploration performed by MET mechanisms. JJ JayM
gpuccio @203
His point, as I understand it, is that the “search strategy” provided by MET mechanisms explores only very small regions around the current, known viable, point. Since viable points in genome space are not uniformly distributed, this search doesn’t need to look at all possible sequences. It works because it is so limited.
Nonsense. No search needs to look at all possible sequences. In a random search, it is just a question of reasonable probabilities. And a search limited to the surroundings of what already exixts will never find anything distant.
Exactly! This is why Dr. Dembski's two papers don't appear to be "pro ID." The MET mechanisms are not a random search on a uniformly distributed genome space. Far from being nonsense, djmullen explained what is observed in real organisms.
He is explaining exactly what we observe.
I was not aware that we have been observing natural history.
The scientists studying MET mechanisms at the cellular level are observing exactly what djmullen described. That is the evidence we have to explain as well or better than MET for ID to be a real scientific theory. Those scientists are also extrapolating backward through history. If we want to challenge them on the validity of those extrapolations, we need to do so based on the evidence, not on baseless claims.
As far as I know, I can assert what I like. . . . I can make all the claims I want . . . I don’t think you are in the best position to decide what claims ID supporters should make. ID supporters can decide for themselves.
You can indeed make all the claims and assertions you like. So long as they are unsupported, though, you are giving ID opponents more ammunition. They can point at ID proponents who make baseless claims and use them to discredit or ignore the whole theory.
There’s no need for name calling, this is supposed to be a collegial environment.
It was not meant as name calling.
Ah, clearly it's my misinterpretation. "Stupid" can't be in any way considered offensive.
You decide your strategy for yourself. And obviously, research is always welcome.
The research must come before the claims or we will be rightfully ignored by the scientific community. There is much that ID can validly claim, with support. Your overbroad claims in this thread are not in that set. JJ JayM
"what evidence leads you to believe that evolution cant get over a mdedium or high hurdle?" Human nature. The fact that there is no example where it has ever been shown to happen. Other wise humans being humans would not be letting us forget about the one let alone myriad of examples that exist. I suggest you read Silver Blaise and the famous incident of the "Dog barking in the Night." But the dog didn't bark. Precisely. And so do the evolutionist fail to bark. jerry
Jerry, what evidence leads you to believe that evolution cant get over a mdedium or high hurdle? djmullen
"Evolution is not looking for something. It is stumbling across things that are useful or just different (genetic drift)." Genetic drift is not exploring or searching or different but cuddling up in a cave and never coming out again. The MET says that evolution is stumbling but the question is how often is what is found useful, or really "how" useful. It finds useful things but can it find combinations that are novel complex and functional. If you give the MET a very low bar to hurdle it will succeed but if it has to find some medium or high hurdles it has a tough time getting over them. That is the issue, has it ever gotten over a medium hurdle let alone a high hurdle. So far the evidence say no. jerry
gpuccio [118]: "Now you state the most incredible thing: that when a bacterium duplicates its DNA, there is no search necessary. That is incredible, not because it is not true, but because it is so trivial that I really can’t understand why you have to state it." I have to state it because Dembsk and Marks seem to have missed it altogether. Remember, their papers describe searching for this protein amongst the uncountable gazillions of possible proteins of the same length when the truth is that organisms just copy what they already have. gpuccio [149] "There are single point mutations which are incompatible with life. Much more difficult would be to find a single point mutation which gives a reproductive advantage to a complex organism." Easy: http://www.circ.ahajournals.org/cgi/content/full/104/20/2498 "Apolipoprotein A-IMilano is the result of a point mutation, with an arginine to cysteine substitution at position173" Humans are complex organisms. This is a point mutation which gives a reproductive advantage to a complex organism. A few decades ago, a family in Milano, Italy was found to almost never have heart attacks. The cause of their good fortune was tracked down to a single point mutation in the DNA that codes for Apolipoprotein A-I. The new molecule, Apolipoprotein A-Imilano removes cholesteral and atherosclerotic plaque. Better yet, scientists have discovered that injecting it directly into the blood stream also works its wonders in people who were born without that single point mutation too! I can't wait for that drug to hit the market! gpuccio [150] "Let’s take the example of myoglobin. At some point in natural history, this protein appears. It is about 154 aminoacids long. Where did it come from?" http://www.hawaii.edu/microbiology/Alam/downloads/Evolutionary%20implications%20myoglobin-like%20proteins%20found.pdf "UH researchers find myoglobin-like proteins in ancient microorganisms A University of Hawaii research team has discovered a new class of myoglobin-like proteins in ancient microorganisms. In animals, myoglobin (found in muscle) and its close relative hemoglobin (the main protein in the red blood cell) play an essential role in oxygen transport and storage. The newly identified proteins may be the evolutionary ancestors of proteins involved in oxygen sensing as well as transport and storage." "...djmullen’s statement that DNA replication was not a search was suspiciously trivial. Maybe he (and you) have not well considered that the search happens before, when you have to find the information, and not when you simply copy or transmit it?" Searching in the immediate vicinity of a known good genome is what lowers the odds enough to make it possible. gpuccio [167] "There is nothing in the homologies which shows incremental change in the sense of a gradual modification of the function." Ahhh... you're searching through proteins in current organisms, right? So if the proteins diverged hundreds of millions of years ago and the intermediate forms have long since gone extinct, what would you expect to find in current proteins? gpuccio [171] "First of all, my expertise in proteins is not your business. This is a blog, and not a scientific journal. And I, and all the other sincere people who come here, have never spoken “from authority”, or requested any academic title or position from others." Since all you give us is your opinions on proteins and BLAST, your argument is entirely an argument from authority except that you haven't established your authority. "Second, I feel honored of being compared to Salvador Cordova." Well, you're going a long way towards emulating him. Salvador's greatest blunder was to take Avida, a program he knew nothing about, set the single point mutation rate to about 99% and triumphantly announce that the organisms were still growing and that Avida was bunk. Unfortunately for him, one of the readers showed his comment to one of the authors of Avida who dug into it and discovered that every single organism was being blasted to pieces in the first generation and that the poor program was counting fragments of the defunct digital organisms as living, giving the false count. Since typical mutation levels in real life are far under one percent and no real researcher had ever been foolish enough to set it to "puree", Salvador basically just proved that Avida wasn't fool proof. You, on the other hand, are taking a program, BLAST, which was invented to find relationships between proteins and is used for that purpose every day. (Wikipedia says the article introducing BLAST was the most widely citedpaper published in the 90s.) Take my advice, don't let people who understand and use BLAST every day hear that you've used it to prove that proteins can't possibly be related to each other. Or, if you do, ignore whatever they say and take a page from Salvador's book: He keeps telling people how he found a serious problem in Avida, one that even the authors didn't know was there. Works for him, might work for you too. djmullen
Gpuccio While I think of it - I have a queston - arising from what you said in post #167. But there is no model which can explain the step by step derivation of myoglobin from some other simpler protein with a different function. And remember, such a model should explain not only the variations in aa sequence, but also the variations in function which allow the selecrion and expansion of each single intermediate. Again, that’s why I am in ID, and not in the darwinian field. Suppose that biologists eventually produce such a model and were consistently able to produce such models for favourite ID examples such as the cliched flagellum. You imply that you would then move to the Darwinist camp. But would there still be any evidence for design or would this wrap it up? Mark Frank
JayM (#194): In brief: "His point, as I understand it, is that the “search strategy” provided by MET mechanisms explores only very small regions around the current, known viable, point. Since viable points in genome space are not uniformly distributed, this search doesn’t need to look at all possible sequences. It works because it is so limited." Nonsense. No search needs to look at all possible sequences. In a random search, it is just a question of reasonable probabilities. And a search limited to the surroundings of what already exixts will never find anything distant. "He is explaining exactly what we observe." I was not aware that we have been observing natural history. "That’s the point under discussion. Personally, I am interested in finding where the limits of MET mechanisms lie. I think that is the most fertile ground for ID research. However, no one currently knows where those limits are, so you can’t simply assert that they are insufficient to produce the structures you note. We have to prove that limits exist and determine where they are." As far as I know, I can assert what I like. We are debating scientific theories here. The fact that we don't exactly know something is exactly the reason why we build theories for the best explanation. And scientific theories do not prove anything (they are not mathemathical theorems). They suggest a possible best explanation. Please, check your scientific epistemology. "As long as a possible path from one genome to another can be hypothesized (and the literature is full of examples where researchers looked at families of proteins coded for by those genomes and found homologs) and as long as that path doesn’t violate any known natural laws, the argument that MET mechanisms are in principle unable to take that path is refuted. We need hard evidence before making the claims that you do above." I can make all the claims I want, and suggest that they are the best explanation fot known data. Anybody can judge how good an explanation they are. The fact that other explanations, possible "in principle", but empirically unsupported and unrealistic, are suggested by others means only that we are in a free world, and thyat different scientific theories are competing for explanation of what we know. I am perfectly fine with that. And no one can simply affirm that his antagonists need "hard evidence", while he can stick to explanations which are possible "in principle". "Again, that isn’t his claim. He simply noted that observation shows that populations evolve in very small steps. Since viable genomes are clustered rather than uniformly distributed, this results in finding new viable genomes much, much more efficiently than random search." Again, nonsense. The fact that viable genomes are clustered (true) does not mean that they are all in the same spot of the space. It just means that they are clustered, and that the clusters are interspersed in the space. So, the kind of search that you describe is efficient only to remain in the same cluster. "There’s no need for name calling, this is supposed to be a collegial environment." It was not meant as name calling. Please notice that I explicitly refer to "what you say", and was preceded by a long and explicit "if", which gave space to the possibility that I was misunderstanding your statements. Anyway, if you just object to the terms as too rude a comment for an expressed idea, then I apologize. Sometimes I am really primitive. Let's say that "then what you say is both inconsistent and intellectually fristrating". "You have shown, at best, that there are a wide variety of proteins in existing organisms." That's exactly what I wanted to show. "You have not shown that the islands of viable genomes are too dispersed to permit all that diversity to have arisen from a common ancestor via MET mechanisms." That was not my point, nor was it the subject of djmullen's affirmations. "That would be a phenomenal achievement and a huge amount of work." And we are working at it. "Until that work is done, though, ID supporters shouldn’t make overarching claims about the existence of those limits. This is especially true given the extensive research being done to trace proteins back in history." I don't think you are in the best position to decide what claims ID supporters should make. ID supporters can decide for themselves. You decide your strategy for yourself. And obviously, research is always welcome. gpuccio
An aside. All this talk of islands of function in the sea of possible proteins (or DNA) made me think that it would be more appropriate to talk of evolution exploring the functional space rather than searching it. Evolution is not looking for something. It is stumbling across things that are useful or just different (genetic drift). I don't think it makes the maths any more or less relevant but it is a more appropriate metaphor. Mark Frank
"Personally, I am interested in finding where the limits of MET mechanisms lie." There are no limits according to the MET. I assume that is what this thread is about. To show that the MET has limits. Let me provide what I believe the MET says. When a population has offspring and the environment is stable then the variation of the offspring will be limited by this stability or I guess as one person put it here, it stays within the sweet spot. I am not sure that is a good term but you can use it. This doesn't say that some offspring cannot be born that are far away from this sweet spot. Probabilites say that such a thing could happen and the organism could be viable. It just says that given the current environment most of these will not likely be able to move the average genome far away from the original population average. Unless there is an extremely useful variation in these offspring, gene flow will bring future offspring back into the fold. It is possible that some characteristics that push an offspring outside the so called "sweet spot" will take hold and the new average for the population will move slightly away. And that is the MET in a nutshell. If a sub population separates from the main population and encounters a new environment, then the sweet spot may not be so sweet anymore and offspring outside the sweet spot may become dominant because of natural selection in the new environment. Now if one wants to say that the sweet spot includes all these new populations then one has to deal with the potential that the average genome may keep moving and soon a new population cannot inter breed with the old population and so is now on its own moving slowly away from the original. So I am not sure sweet spot is a meaningful concept for the MET though the idea explains why species do not vary much over time. But it fails to take into account outliers and their ability to change a population or a sub population. Now these outliers may result from variation that is due to recombination, or a mutation or they could just be due to a unique shuffling of the genes that are currently in the gene pool that now exist but have never appeared in any offspring before. The later case can be seen in artificial selection or breeding to get to some combination that is possible but would not show up in any natural way and has to be intelligently guided even though there is always a probability that it could happen naturally. After all Chihuahuas and Great Danes are both dogs. Over time MET predicts these isolated populations will diverge even more. By the way a sub population could be within spitting distance of its parent population and not interact. Some insects that are just one generation away from its parents may find a certain new plant to live on and because of this new plant may never go back home and thus form a new species within just a few feet of it original population. Mammals need a much greater distance to form a sub population. Anyway that is the theory but the changes that can happen has limits as Behe has shown in the Edge of Evolution and discussed in a great book by Ray Bohlin and Lane P. Lester called he Natural Limits to Biological Change. I assume all these searches discussed in this thread is about the same thing or the possible movement of a genome away from an original to another which has some element that provides a new functional capability that is not just trivial but complex in nature. The future will show the way as new mappings of genomes and more understanding of the function of non coding elements will tell one how feathers are formed and bats have radar and why mammals do not lay eggs etc. Then one will be better able to see what must have taken place in a Darwinian sense to get to these new complex and functional capabilities and whether the creeping of a genome over time could be the answer or rather whether massive relocations of the genome had to have taken place. Based on the logic of what is presently know, neither will be a viable option and then a new escape will have to be planned for the Darwinist to move on. One of the things that recent threads has revealed is that these discussions are not about data and logic and a striving to learn but overt attempts to preserve one's current ideological turf through whatever methods one thinks will work. Different people choose different tactics including civility but all this has been seen before and what is interesting is how they play out in different people. One of the ways it nearly always plays out is in ad hominems. jerry
JayM, Apologies. I copied my response to Khan, along with additional comments, so parts of it didn't connect to what you were saying. My main point is that it's much more practical and feasible to provide a single reasonable hypothesis than it is to wade through the entire data set for biology to prove that none exists. We should be looking at what we know now. What you are talking about is gathering the knowledge of all genomes and completing the project we call biology.
While I think that ID theory has considerable promise, I have yet to see how it could be falsified.
Personally I consider theistic evolution to be ID-compatible, although people like Ken Miller would disagree. And while many ID-compatible ideas are very much open to falsification I'd agree that there are some ID-compatible ideas that are not.
We need to identify the real edges of evolution, both in terms of limits of the MET mechanisms and disconnected islands of viability in the genome space.
Long term, yes. I just disagree that the evidence we do have now is not enough to make certain claims. Patrick
Patrick @196
We are merely asking that the grand claims be verified in ANY manner. Essentially, you’re asking us to accept as an article of faith that pathways of hundreds and thousands of informational bits leading to an irreducible core are as easily achieved as pathways consisting of less than 50 informational bits (and IC is often not an issue at all)
I'm not asking any such thing. My point in that section of my post was to note that overbroad claims do not do the ID movement any favors. When someone claims that "ID means that X could never happen." and an ID opponent says "Actually, it could have happened in any of these ways....", we all lose credibility.
Until that work is done, though, ID supporters shouldn’t make overarching claims about the existence of those limits.
It’s long been stated the design inference is open to falsification
That's a separate, but also important issue. Stating something is easy, supporting it is more difficult. While I think that ID theory has considerable promise, I have yet to see how it could be falsified. That's one reason I hopped into Dr. Fuller's thread related to learning about the designer. If the designer is not bound by any constraints, voluntary or otherwise, I don't see how to falsify the design inference. If literally anything is possible, any observations can be claimed to support the theory, which makes it unfalsifiable.
I’ve elaborated at greater length about this, but I’m open to the potential that the only instantiation of Design is OOL and the configuration of the system (code/language conventions, limited foresighted mechanisms, etc.) allows unguided processes to largely work from there, but for the reasons gpuccio has already given this seems unlikely.
I agree, although not for gpuccio's reasons. We need to identify the real edges of evolution, both in terms of limits of the MET mechanisms and disconnected islands of viability in the genome space. Until we have real evidence, we need to be circumspect in our claims if we want to be taken seriously. JJ JayM
CJYman:
Simply put, the foundational randomness and chaos from which everything arises would be uniform and would not be biased toward any specific outcome.
I misinterpreted this statement in the post above and falsely accused it of being self-contradictory. My sincere apologies. But it raises the question: If other distributions arise from that foundational randomness, shouldn't they be uniformly distributed? In other words, we would expect the vast majority to be non-uniform and biased toward their intrinsic targets.
The implications of this would be that there has always been an infinite regress of bias in the foundational search space to produce our universe, life, evolution, and intelligence. However, that infinite regress of active info is only one of the options which I have discussed in my comment #122
According to the principle of indifference, each of your n options has a 1/n chance of being correct. ;-) And how do you know that "life, evolution, and intelligence" is the target? Maybe the Designer was trying to create a sterile universe, but some replicators arose accidentally and infected a planet with life. In that case, where's the active info? I'm being tongue-in-cheek, but the point is that we're dealing with ultimate questions that can be answered by neither science nor math. BTW, I find it strange that talk of design regress is verboten on this board, but it's okay to talk about probability hierarchies that regress to before the universe existed.
and if there is indeed “uniform randomness,” then the ID position becomes more of an obvious and accurate choice.
There is not uniform randomness, regardless of whether the ID position is true or not.
One way to begin to answer if the preceding assumptions are justified is to “test randomness” and see if it has any tendency to match non-uniform search spaces with search algorithms to generate active information.
If this "randomness" is over M(omega) as Marks and Dembski define it, then obviously it will have no such tendency, although it doesn't tend very far in the other direction, as Prof. Olofsson has pointed out. How do you propose we apply this mathematical fact to reality? R0b
ROb[195], CJYman claims a "non-arbitrary uniform probability distribution" but as ROb points out, it is determined by the choice of metric, thus not unique. Prof_P.Olofsson
As long as a possible path from one genome to another can be hypothesized
Does such a hypothetical non-trivial pathway even exist?
(and the literature is full of examples where researchers looked at families of proteins coded for by those genomes and found homologs)
These basic comparisons do not meet the same standards for a hypothesis for what you then write:
and as long as that path doesn’t violate any known natural laws, the argument that MET mechanisms are in principle unable to take that path is refuted.
There's a reason I kept referring to previous discussions:
You have not shown that the islands of viable genomes are too dispersed to permit all that diversity to have arisen from a common ancestor via MET mechanisms. That would be a phenomenal achievement and a huge amount of work.
This discussion has reached the same point as my discussion with Khan. Yes, let's MERELY calculate every single potential indirect stepwise pathway and set that as the milestone that the ID community must meet in order to qualify itself. Yes, that's sarcasm. For the Darwinist "[a] single plausible pathway will do, which does not require analyzing ALL possibilities. And if this pathway happens to consist of processes that act uniformly in regards to other biological objects, then a large chunk of most design hypotheses would be falsified. But if you plan on verifying whether a plausible pathway does not exist at all, then indeed we’d need to calculate every single potential pathway. We are merely asking that the grand claims be verified in ANY manner. Essentially, you’re asking us to accept as an article of faith that pathways of hundreds and thousands of informational bits leading to an irreducible core are as easily achieved as pathways consisting of less than 50 informational bits (and IC is often not an issue at all)" which are all within local genome space.
Until that work is done, though, ID supporters shouldn’t make overarching claims about the existence of those limits.
It's long been stated the design inference is open to falsification. I've elaborated at greater length about this, but I'm open to the potential that the only instantiation of Design is OOL and the configuration of the system (code/language conventions, limited foresighted mechanisms, etc.) allows unguided processes to largely work from there, but for the reasons gpuccio has already given this seems unlikely. Patrick
CJYman:
First, those assumptions could be a part of an hypothesis which incorporates the math within these papers. As such, until the assumptions are falsified or the math is shown to be incorrect, we have a standing scientific hypothesis.
An active info analysis requires, according to Dembski, "a search space omega, a metric D that induces a compact topology on omega, a target T included in omega, a uniform probability measure U on omega induced by D, and an arbitrary probability measure phi on omega." In order to come up with those parameters, we need to make assumptions about how the underlying phenomenon should be modeled. We may also have to make assumptions about the phenomenon itself. Only the latter type of assumptions can be falsified.
You seem to be saying that these assumptions have to do with how we measure active information by individuating the possibilities, yet I don’t see this as a problem at all, since the probability of a pattern is measured against a non-arbitrary uniform probability distribution.
But how do we define the space over which this uniform distribution sits? It depends on how we individuate the possibilities. To borrow an example from Dembski, what is the endogenous information in a roll of a die? That depends on whether we define the die landing on the floor as a different outcome than the die landing on the table.
Breaking up the pattern and convoluting it into a couple of different searches would be the arbitrary action, and I’m not sure that would even return a different measurement when these separate searches are each measured against a uniform probability distribution.
Actually, we're not talking about breaking anything into different searches. The point is that the active info measure depends on how we individuate the possibilities, contra Dembski's earlier position on how information should be measured. And I should qualify my usage of arbitrary. It's arbitrary according to the mathematical framework, which doesn't dictate how to model things. But if we want to use the framework as a tool and draw conclusions from it, then our modeling choices may be determined by what we're trying to find out. So in that sense it could be non-arbitrary (although I'd love to see an example with a modicum of rigor). But if all we're trying to find out is the amount of active info in something other than a pre-defined mathematical model of a search, then it's arbitrary.
As far as I understand, Active info = probability associated with bit operations taken to find the pattern – probability of pattern measured against a uniform distribution. Neither the number of bit operations nor the uniform probability distribution are arbitrary figures, so neither is their difference (active info) arbitrary.
They aren't arbitrary given omega, D, and T, but how do we non-arbitrarily come up with those parameters? That's one of the questions we'll need to answer if we want the framework to tell us anything meaningful about reality. For instance, how do you non-arbitrarily define T when you model biological evolution as a search?
Actually no, that would not be a non-uniform distribution of distributions, if *everything* is characterized by a uniform distribution. If you actually mean *everything*, then that would include the distribution of distributions.
Exactly. It's logically impossible for everything to be characterized by a uniform distribution. Which renders your subsequent statement self-contradictory: "Simply put, the foundational randomness and chaos from which [b]everything[/b] arises would be uniform and would not be biased toward any specific outcome." (emphasis mine) This post is too big. More later... R0b
gpuccio @192
Let’s not play games, please. If djmullen is saying that, as long as living beings just reproduce themselves with very minor changes, they remain the same and are in the same “sweet spot” of the search space, and so no search is necessary
That's not what he is saying. Here is an exact quote:
In practice, they do this by making only small changes to the DNA they hand down to their offspring. This keeps the results of their so-called “search” either inside the sweet spot or very close to it. No other “search strategy” is necessary.
His point, as I understand it, is that the "search strategy" provided by MET mechanisms explores only very small regions around the current, known viable, point. Since viable points in genome space are not uniformly distributed, this search doesn't need to look at all possible sequences. It works because it is so limited.
he is saying something literally true, but completely trivial, and which does not correspond to what we onserve in natural history
He is explaining exactly what we observe.
and does not explain in any way the appearance of new proteins, new functions, new body plans, new species, and so on.
That's the point under discussion. Personally, I am interested in finding where the limits of MET mechanisms lie. I think that is the most fertile ground for ID research. However, no one currently knows where those limits are, so you can't simply assert that they are insufficient to produce the structures you note. We have to prove that limits exist and determine where they are. As long as a possible path from one genome to another can be hypothesized (and the literature is full of examples where researchers looked at families of proteins coded for by those genomes and found homologs) and as long as that path doesn't violate any known natural laws, the argument that MET mechanisms are in principle unable to take that path is refuted. We need hard evidence before making the claims that you do above.
But if, as I believe, djmullen (and probably you) are affirming that all the proteins and functions and body plans and information we find in the living world were generated starting from a simpler ancestor without the need for any search
Again, that isn't his claim. He simply noted that observation shows that populations evolve in very small steps. Since viable genomes are clustered rather than uniformly distributed, this results in finding new viable genomes much, much more efficiently than random search.
because all the transitions happened in the same sweet spot of the search space, then what you say is both stupid and irritating. I have shown clearly that proteins, as we know them, are interspersed in the ocean of possible sequences.
There's no need for name calling, this is supposed to be a collegial environment. You have shown, at best, that there are a wide variety of proteins in existing organisms. You have not shown that the islands of viable genomes are too dispersed to permit all that diversity to have arisen from a common ancestor via MET mechanisms. That would be a phenomenal achievement and a huge amount of work. Until that work is done, though, ID supporters shouldn't make overarching claims about the existence of those limits. This is especially true given the extensive research being done to trace proteins back in history. JJ JayM
Re #188 First - I am sorry - my comment in #187 was more suitable for a private e-mail than a public forum - sometimes my frustrations spill over. The issue of uniform probability distributions has been done to death here and elsewhere. For example, the calculation that goes: (Number of functional outcomes(/(number of trials) = (probability of functional trials) relies on a uniform probability distribution. Mark Frank
JayM: Let's not play games, please. If djmullen is saying that, as long as living beings just reproduce themselves with very minor changes, they remain the same and are in the same "sweet spot" of the search space, and so no search is necessary, he is saying something literally true, but completely trivial, and which does not correspond to what we onserve in natural history, and does not explain in any way the appearance of new proteins, new functions, new body plans, new species, and so on. So, if djmullen (and you) want to spend your time accumulating tautologies which do not mean anything, you're welcome, but I will not follow you any more. But if, as I believe, djmullen (and probably you) are affirming that all the proteins and functions and body plans and information we find in the living world were generated starting from a simpler ancestor without the need for any search, because all the transitions happened in the same sweet spot of the search space, then what you say is both stupid and irritating. I have shown clearly that proteins, as we know them, are interspersed in the ocean of possible sequences. That is simply true. See also my answer to rna. So, one of two things: a) a lot of search was necessary, and the ocean of possible sequences has been traversed a lot of times (which is the opposite of what you seem to say) or b) no search was necessary, because all the proteis we know were already present in LUCA, as we know them or in a very similar form (which is simply wrong and foolish) And that's enough, for me. Unless you say something new and sensible, I don't think I will answer any more to that kind of nonsense. gpuccio
gpuccio @167
I appreciate your efforts, but I am afraid I can only repeat for you what I have already said about djmullen: your characterization is inaccurate.
In the interests of brevity, I'm going to focus just on djmullen's original statement:
But every living organism that exists is the product of organisms that are already in that tiny proportion of the possible genomes and if they successfully reproduce then their DNA is already in it too. In practice, they do this by making only small changes to the DNA they hand down to their offspring. This keeps the results of their so-called "search" either inside the sweet spot or very close to it. No other "search strategy" is necessary.
This is an entirely accurate representation of the MET position. Let's take it step by step so you can point out where you disagree: 1. Every living organism is in a viable point in genome space. This is true by definition, else the organism wouldn't be living. 2. Changes in DNA between generations are small. This is observed. 3. Living offspring are in viable points in genome space. This is true by definition, like 1. 4. The points in genome space occupied by offspring are very close to those occupied by their parents. This is a result of 2 and is also observed. From these four points, we can see that: 5. The viable regions in genome space are not uniformly distributed. If they were, the odds of getting viable offspring would be vanishingly low. And so we come to djmullen's conclusion:
Evolution says that every living organism since the first self-reproducing molecule is already in the genomic sweet spot and all they have to do is keep their offspring in it too.
That follows logically from the earlier points. Now, you and I may disagree with djmullen about the probability that all life, from the very first common ancestor 3.8 billion years ago to today, is in a connected network of viability within genome space, but you can't say that his characterization of MET is "inaccurate." What is your precise objection to what he said? JJ JayM
CJYman @161
That’s great, but it still seems to cede the field of biology to modern evolutionary theory.
It cedes biology to evolution and it takes evolution for itself.
That's an interesting and refreshing view, that seems to stretch even the ID community's Big Tent. While I personally think that ID research would be most profitable in identifying exactly where the "edge of evolution" lies and why, I share your respect for the work done by scientists. In order for ID theory to replace MET, it needs to explain everything that MET does, and then prove itself to have greater predictive power. We're not going to get to that point by ignoring the knowledge already gained by evolutionary biologists. So, although I don't agree that we need to cede biology to ID opponents, I find your position thoughtful and rational.
An aside due to my inner geek: Is there a formal proof that CSI cannot be generated from evolution simulations?
The only formal proof presented is that CSI won’t generate itself through only chance and law and this is implicit within these two papers.
I was hoping for something more explicit. CSI isn't even mentioned in either paper.
Since CSI is a measurement of finding a specified or pre-specified pattern at better than chance performance, problem specific information is required to find it (according to NFLT). Problem specific information is measured as active information within these two papers.
The reason why I still don't see how either of these papers can be claimed to be "pro ID" is that active information can be transferred to the genome from the environment via MET mechanisms. This is why GAs work so much better than random search. From this, the papers could be said to be "pro MET."
I don’t believe that "all simulations of evolution require foresight." We’ve discussed several different genetic algorithms in recent threads here and Dawkin’s Weasel is the exception rather than the rule.
My apologies for not being clear. If we are discussing any simulation relevant to biology, then evolution is the discovery and increasing of CSI. Can you provide any simulation of biological evolution which discovered CSI from initial conditions that were not programmed with any consideration for future results?
I'm not sure I understand the constraints you're specifying. There are a number of GAs documented on the web that simulate a small subset of MET mechanisms and generate complex, previously unknown solutions to the problems they are posed. One of my favorites is the evolution of antenna designs done at MIT. Another interesting one is Thomas Ray's Tierra. While we're on the topic, I would be very interested in modifying one of those GAs to address your questions. How would you go about measuring the CSI for the results of one of those, then finding the source of it? JJ JayM
Mark Frank [187], The last claim in the video is indeed classic confusion of conditional probabilities: "It was about 10^80000 times more probable that ID was required..." Sigh. You're probably right, better not waste too much time. I'll read the article when I have time though because tibrune is such as nice person! Prof_P.Olofsson
the usual assumptions about uniform probability distributions How so, pray tell? tribune7
Re #185 I suspect you have better things to do with your time. The speaker is described as: Biophysics PhD candidate, University of Guelph If he gets his Ph D then I will never trust anything from the University of Guelph again. E.g the usual assumptions about uniform probability distributions confusing relative likelihood with relatively probability quoting papers such as Axe which have been thoroughly debunked in the literature without mentioning they even contentious and so on Basically there is nothing new in it. Mark Frank
And my motto is never trust a patrician :-) Just want to see if it will change your mind just a tad :-) tribune7
Hey tribune, protector of the people! No, I haven't. I'll take a look when time permits. Thank you. Prof_P.Olofsson
Professor O, have you seen UD's most recent post? https://uncommondesc.wpengine.com/intelligent-design/mathematically-defining-functional-information-in-biology/ tribune7
CJY, Also see my post [155]. You can beat chance by chance without mysterious explanations. Prof_P.Olofsson
CJY[180],
From what I understand, NFLT has already proven that in order to increase the probability of search, the search procedure needs to be matched to the correct search space.
That seems to be part of the ID folklore but it is not what the NFLT (Wolpert & Macready 1997) says. It states that all search algorithms are equally good/bad if the fitness function is chosen uniformly. So, if the conclusion of NFLT is false, all we can conclude is that the uniformity assumption is not satisfied. In evolutionary biology, it is quite obvious that the conclusion does not hold (for example, the darwininan search clearly beats random search in cases we all can agree upon such as chloroquine resistance, per Michael Behe) and that the assumptions are not met (for example, the fitness function that makes all genotypes equally fit is not very likely); hence, the NFLT simply does not apply. The ID folklore now seems to be that (a) either NFLT applies and Darwin loses (we all agree about that) or (b) the NFLT does not apply and Darwin loses which is completely unsubstantiated.
Thus, when Dembski assumes uniformity, he takes the perfect non-teleological non-biased starting point and begins from there.
More ID folklore. Assuming uniformity means that you make an assumption. The only way to possibly make such an assumption valid is if you state it as a prior distribution in the sense of Bayesian inference, and update it in the presence of data. Prof_P.Olofsson
Rob: “ Informal claims such as “it is as difficult to find a search as it is to find a target” are not supported by the paper unless you make a lot of arbitrary assumptions. One of the arbitrary assumptions involved in measuring active information even violates Dembski’s own repeated warning. He has told us several times that “how we measure information needs to be independent of whatever procedure we use to individuate the possibilities under consideration.” And yet, the active information measure depends very much on how we individuate the possibilities.” First, those assumptions could be a part of an hypothesis which incorporates the math within these papers. As such, until the assumptions are falsified or the math is shown to be incorrect, we have a standing scientific hypothesis. Second, what are those arbitrary assumptions? You seem to be saying that these assumptions have to do with how we measure active information by individuating the possibilities, yet I don’t see this as a problem at all, since the probability of a pattern is measured against a non-arbitrary uniform probability distribution. Breaking up the pattern and convoluting it into a couple of different searches would be the arbitrary action, and I’m not sure that would even return a different measurement when these separate searches are each measured against a uniform probability distribution. As far as I understand, Active info = probability associated with bit operations taken to find the pattern – probability of pattern measured against a uniform distribution. Neither the number of bit operations nor the uniform probability distribution are arbitrary figures, so neither is their difference (active info) arbitrary. Rob: “ The recurring use of the uniform distribution in many attempts to apply math to biology is a model assumption must be argued just like any other model assumption. Indeed, Olle Haggstrom’s response to the active info approach is titled “Uniform distribution is a model assumption”. The question asked by Haggstrom and others is, why should we always expect uniform randomness? In fact, it’s an impossible expectation. If everything were characterized by a uniform distribution, then that would be a non-uniform distribution of distributions." Actually no, that would not be a non-uniform distribution of distributions, if *everything* is characterized by a uniform distribution. If you actually mean *everything*, then that would include the distribution of distributions. Simply put, the foundational randomness and chaos from which everything arises would be uniform and would not be biased toward any specific outcome. Thus, when Dembski assumes uniformity, he takes the perfect non-teleological non-biased starting point and begins from there. However, there is no need to expect “uniform randomness” in order to make sense of these papers. Olle Hagstrom seems to be missing the point in that one option of the papers is that there can be an infinite regress of active information and thus no true uniform search space exists “outside” of our universe. The implications of this would be that there has always been an infinite regress of bias in the foundational search space to produce our universe, life, evolution, and intelligence. However, that infinite regress of active info is only one of the options which I have discussed in my comment #122, and if there is indeed “uniform randomness,” then the ID position becomes more of an obvious and accurate choice. One way to begin to answer if the preceding assumptions are justified is to “test randomness” and see if it has any tendency to match non-uniform search spaces with search algorithms to generate active information. As long as no active information is generated, we can conclude that a random source (background noise and set of laws put together with no consideration for future results) will remain uniform with respect to randomly generated search procedures. Any other assumption is merely that – an arbitrary, un-necessary, and unfounded assumption. Excellent ID research, eh? CJYman: The only formal proof presented is that CSI won’t generate itself through only chance and law and this is implicit within these two papers. Rob: “Just for clarification, are you saying that this formal proof is implicit in these two papers? (An implicit formal proof sounds like an oxymoron to me.) If not, where is this formal proof to be found? Thanks." Sorry for not being more clear, but it is the fact that CSI (as a measurement of better than chance performance) necessarily requires active info which is implicit in these papers. The formal proof that chance and law (barring previous active info) won’t generate active info is what these papers seem to produce. Thus, the formal proof of active info not being generated via chance and law also formally proves that CSI won’t generate from merely chance and law. CJYman
Prof. Olofsson: “Let me clarify: I am questioning the claims that the “search for a search” paper is pro-ID as was claimed in the introduction. In order to be considered pro-ID, it would have to have some implications for biology.” It only shows that random search won’t match search algorithm to search space in order to increase the probability of finding a given target. From what I understand, NFLT has already proven that in order to increase the probability of search, the search procedure needs to be matched to the correct search space. The NFLT could be presented with Dembski and Mark's paper as an hypothesis with ID connections as I’ve previously explained in my comments #121 and 122, and could be falsified as shown in my comment #162. Simply, if biological evolution can be modeled as a search and if our set of natural laws can be modeled as a search (which many physicists seem to do in discussing the values of our laws in relation to all possible mathematical universes) then the implications of the paper apply to all aspects of nature including biology. Prof. Olofsson: “I ask, for example, (1) what is the rationale behind claiming that search algorithms are chosen according to probability distributions, in particular the K-W distribution [154]? and (2) if they are, why do these distributions have to be uniform [159]? and point out (3) even if they are uniform, they are still quite likely to beat random search [155].” 1) It seems that the only assumption necessary is that the laws generating the search algorithm and search space are one out of many possible mathematical configurations of those laws and that whatever produces the search algorithm and search space matches those laws in a fashion which is blind to future results. Thus, the matching of (values for) laws = a search through a probability of all possible (values for) laws. I don’t see how this assumption could be controversial unless the universe and its laws where eternal (without beginning) and the only possibility. 2) They don’t have to be. Don’t the papers merely show that the degree to which a search space is non-uniform is the same degree to which the higher order search for that search space is non-uniform? As I have explained previously (in #121 and/or 122), this allows us the option of an infinite regress of active information. 3) Are you saying that any type of search through a uniform search space will return better than chance performance? Doesn’t that directly contradict the NFLT? Are you saying that searching for an evolutionary algorithm through unguided processes (matching of search procedure to search space with no consideration for future results) will stand a better chance of being found than attempting to locate the original pattern that the search algorithm is to find at better than chance results? It seems that you are saying that it is easier for non-foresighted processes to find the matching of algorithm and search space responsible for finding an optimized antenna than it is to find that optimized antenna by random search. Can you provide any evidence for this? If that were true, then I ask again: … if what you are stating is true and actually has practical effect (ie: actually happens in the real world), why doesn’t anyone just show that background noise (chance) and an arbitrary set of laws (set of laws collected without any consideration for future results — absent foresight) will produce systems which process signs/units and evolve into greater and greater specified complexity, and just falsify ID theory and be done with it. If evolution is so powerful absent previous foresight, why do programmers bother programming boundary conditions for future known targets into the evolutionary algorithms used to solve problems (ie: max efficiency antenna shape)? CJYman
Gpuccio That was very clear and informative thanks. There were a number of things I did not understand, both about antibody maturation and what you believe. I realise now that you proposed antibody maturation only as an analogy of the supposed design process and not as an example. Mark Mark Frank
ROb[165], Good point! :) Prof_P.Olofsson
rna: Thank you for your very correct post, with which I agree completely. What you say if true, but I am afraid you have misunderstood my point (I must not be in good form in these days). In the above discussions, I was not arguing on how rich the general space of proteins is of functional islands, or on how big those islands of functionalities are. I have often discussed that problem elsewhere, and, although I do believe that the islands of functionality are really distant and interspersed, I am perfectly aware that the problem is not at all solved, and that it is very important. All the facts that you cite are correct, even if there are other aspects, which I will not detail here. Mark can witness how, even on his blog, I have often pointed to the measurement of the size of the functional space as to one of the fundamental problems, and like you I expect important clarifications from ongoing research in that field. But in my posts I was just responding to the statements made by others that all functional proteins would lie in a "sweet spot" of the search space, and that therefore no search was really necessary. That is simply false. Now, to be more clear, let's take all the known proteome frome any database of functional proteins. Let's forget, for a moment, the relationship with function, and let's look simply at their primary structures. What I was saying is that those primary structures are absolutely interspersed in the space of possible sequences (obviously, with similar proteins sharing various levels of homology). That is absolutely true, and requires a search to get to the different islands of functional structures. We may discuss on how easy it could be to get to some island in a search (in other words, we could discuss if the search space is more like the coast of Maine or like the Pacific Ocean), but it is absolutely true that it is an ocean with a lot of islands. The things you say confirm my point: it is true that two proteins may have similar 3d structure and function, and completely different primary structure (although that's more an exception than the rule). But you see, random variation (the search) is supposed to work on the primary structure, knowing nothing of the function. It is only the selection part which is interested in the function, and with all the restraints which I have already discussed in this thread. So, those two proteins are distant islands for the search, because of their different primary structures. Therefore, while you say "Thus, the assumption you made of functional proteins being remote islands in sequence space is maybe a little premature.", I have to counter that I made no assumption: the functional proteins we know "are" remote islands in sequence space. You may argue that they could be very big islands, and not so distant as I believe. That can be discussed. But there is no doubt that they are islands, and that they are interspersed in the ocean. The "sweet space" is only a sweet invention. gpuccio
Mark: Now, that's where you have misunderstood my point: "You are positing that an intelligence intervenes every time an antigen enters an individual vertebrate to direct the search for an antibody. The evidence for this intervention is the improbability of all current non-design solutions." No, I was not saying that. I don't believe that a conscious intelligence (the designer) is directly active in the process of the immune response. What I believe is that the process of immune response, in all its parts, is a very intelligent algorithmic procedure embedded in the immune system, probably in that part of the genome which controls procedures, and which we don't understand very much. For the rest, it works like any other software: a complex algorithm which produces intelligent results mechanically, because it has been programmed to do so. So, the designer does not intervene directly, but he does intervene indirectly, through the programmed procedure. Now you will have noticed that the procedure is very complex, and that it requires the careful interaction of many different cell types. So, it is certainly a very good example of complex reality which suggests design. But that was not the reason why I have cited it in our discussion. As you remember, our discussion was about the modalities by which specific proteins could be obtained in a design perspective. I quote from my post: "Let’s take an example. In the case of antibody maturation, which I have cited, the intelligent procedure embedded in the immune system realizes a very significant increase of the affinity of the primary antibody response in just a few months. As I have said, that is attained through targeted hypermutation and intelligent selection through direct measurement of the affinity of the new clones to the antigen. And still, that is a rather indirect method, because the system does not know in advance the sequence to be found, but only the function to be measured." So, my reasoning is as such. We don't know the modalities by which the designer implements the information "directly", when he designs the genome or its variations, but we have at least this interesting model where the designer "indirectly" designs proteins, through an embedded procedure. And this model is interesting because it uses partial random search, coupled to function measurement for selection. That is also the main method used by protein engineers today. So, as we know that the designer uses that kind of algorithm in the model of the immune system, it could be reasonable to hypothesize that he could use it also in building the genome information, but this time "directly", for instance through targeted, guided random variation (think to some possible hypermutation process on a duplicated gene), followed by function measurement and direct selection (which need not pass through the long process of genome expansion though reproductive advantage; the "good" results could just be kept, and the bad results passed again through hypermutation). In alternative, I have offered also the possibility that the designer may work knowing already the solution: in that case, he could still use targeted hypermutation, but just "keep" the correct nucleotides, without any preliminary function measurement. The third alternative is direct, intelligent mutation. That would obviously be the easiest way. That was simply my argument. I hope it is clear now. gpuccio
gpuccio: I would be really careful in any statements regarding the possible distribution of functional proteins in sequence space. This is an area under intense experimental investigation and I don't think a consensus view is emerging yet. Let me give some examples: The first comes actually from DNA or RNA in vitro selections. This is of course an artificial and simplified example but it yields an idea about the distribution of functional solutions to a chemical problem (enzyme activity, binding of a ligand, regulation) in sequence space. In vitro selection experiments typically start with a pool of random RNA sequence of 40 to 70 nucleotides in length e. g. with a sequence space of 4 to the power of 40 or ~ 10 the power of 24. Due to technical limitations normally only 10 to the power of 14 molecules can be synthetized and investigated (more do not fit in a test tube :-)). Thus only a very tiny fraction of the possible sequence space can be explored in these experiments. Yet, in these experiments normally not only one but multiple solutions to a given problem are found and normally on must restrict oneself to look only for the most abundant solutions. This seems to indicate that the sequence space is rather rich in possible solutions = functional sequences and moreover that for any given problem multiple structurally and sequentially unrelated solutions can be found. The functional RNAs found in such experiments do not only trivial things but there have been molecules e. g. working as RNA-polymerases, catalyzing peptide-bond formation or catalyzing chemical reactions that are not even possible with naturally occuring protein enzymes. Second, many examples are known where proteins with totally different folds and unrelated in sequence catalyze exactly the same reaction using the same chemistry such as the proteases subtilisin and trypsin but there are many more examples. Third, the same reaction can be catalyzed using enzymes not only unrelated in structure and sequence but also using very different chemistry such as the metalloproteases and the cysteinproteases but examples for many other enzyme classes can be found. If anything this argues for a sequence space rich in functional proteins. Fourth, there are examples where the exchange of a single amino acid or very few amino acids leads to a novel stable three-dimensional fold of a protein with a change in fold being a prerequisit for a possible novel function. On the other hand each functional protein sequence seems to be surrounded by similar sequences with the same fold and function in sequence space such as the functional myoglobins from different species. Thus, the assumption you made of functional proteins being remote islands in sequence space is maybe a little premature. rna
Mark: Well, we have seen how the primary antibody response takes place. I would like to remark again that the primary repertoire is built without any information from outside antigens, and that its production can be already considered a highly engineered procedure, using random variation in a strictly controlled scenario, and attaining the best blind defense for any (or most) possible antigens. But once an antigen really enter the body, information is inputted from the environment in the form of the epitopes, and is processed and stored in the antigen presenting cells (APCs). It's that information which allows the specific selection of active antibodies from the pool of the repertoire, and their amplification through proliferation of the appropriate B cells. Let's remember also that that process is strictly controlled by the T compartment. Then, in the following months (3-6), ensues the very interesting process of antibody maturation. That is less well understood, but we know many things about it. The process implies two different mechanisms: a) Somatic hypermutation, probably due do specific mutating enzymes, strictly targeted to the small region of DNA coding for the active site recognizing the antigen (about 2000 bp) b) Selection of the resulting mutated clones according to their affinity to the antigen: the clones with higher affinity are stimulated, and those with lower affinity are suppressed. The process is cyclically repeated in the Germinal Center of the lymph node, until high affinity is achieved. Future, secondary responses to the same antigen are base on those high affinity antibodies, and are much more efficient. A few comments. The hypermutation is the random part of the process, but again it is highly engineered: it takes place only in a very short segment of DNA (the appropriate one). It is probably well controlled as to its rate and modality, and is actively accomplished through appropriate enzymes. And it takes place in a very precise window of the reproductive cycle of the lymphocyte. Now, let's go to the selection. It is almost certainly based upon exposure of each mutated clone to the antigen stored on APCs, and the following inhibition or stimulation of the clones is probably effected by the usual T regulation lymphocytes, and certainly in a very precise manner, very strictly dependent on the measured affinity. So, in this second process we see again the application of targeted and controlled random variation, like in the building of the primary repertoire, but this time it: a) starts with a pool of antibodies which are already a functional island of the space (they have been selected from the primary repertoire by the exposition to the antigen). b) is followed by measurement of the resulting affinity after the variation, and the consequent stimulation or inhibition of the clone. In other words, here the information inherent in the antigen plays a fundamental role, and "guides" the process towards its goal. In hope that's clear enough. If you have any doubts, please ask. Now, in the following post, my comments and my point. gpuccio
Mark: My compliments, you have understood much about the immune system, but you are still missing some key points (not your fault). And I am afarid you are also missing my point in citing that system as an example (again, not your fault: I wrote very quickly, giving too many things for granted). So I will proceed this way: in this post I will try to make a general review of how the immune system works, staying as simple as possible (not an easy task). And in my next post I will elaborate better on my point about that system in our discussion. First of all, it should be clear that for simplicity we will palk only of the B cell response (antibody mediated). So, B cells are specialized lymphocytes present in the body wherever there is lymphatic tissue, and in the blood. They derive from hematopoietic stem cells, which become committed to maturation as B lymphocytes. During the first part of life (fetal life and the first few years) the immune system undergoes a process of ontogenesis and maturation which brings it tu full functionality. During that process, mature B cells are formed which we will call here "virgin" B cells (because they have never met the antigen). Now, a very important thing happens during the maturation of virgin B cells. In the single clones of B cells, a recombination happens at the level of DNA, in a series of genes which are implied in antibody production. As a consequence of that DNA recombination, each final clone of B cells has a different DNA (in that tiny part), and produces a different antibody. Each of those clones is represented by a number of lymphocytes with the same DNA recombination, and is distributed in the body. So, that is what is called the primary antibody repertoire. It is through it that we can have the primary immune response to almost any antigen in nature. Now, we have to notice some important points: a) The target of antibodies are mainly small peptide segments, usually about ten aminoacids long, called epitopes. While antigens are big molecules, usually proteins, epitopes are the real unit which is recognized by antibodies. So, the combinatorial space of possible epitopes is big, but not huge (about 20^10). b) The primary repertoire is achieve through a process of recombination which includes many random components, in a very controlled scenario. That's what allow to achieve a repertoire which is blind, in the sense that it covers in an interspersed, random way, the combinatorial space of possible epitopes. No information from outer antigens is used in that process. In the ens, primary antibodies are in enough number (nobody knows how many) and sufficiently interspersed in the space, that almost any possible antigen will meet one, and usually many, antibodies which can bind to it. Indeed, the primary immune response is policlonal. But for the same reasons the affinity of the primary antibodies to one specific antigen is almost always low. c) B (and T) lymphocites are practically the only cells in the body which undergo a somatic modification of their DNA. Theur DNA is therefore slightly different from the DNA of all other cells. Now, let's go to the primary response. When some antigen enters the organisms, it is processed by specialized cell, the antigen presenting cells, whose role is to expose the epitopes on their surface, correctly associated to other molecules, so that virgin B cells may be exposed to them, until the clones which have an antibody which can bind the epitope are found and their proliferation begins, stimulated and controlled by other lymphocytes, the regulatory T cells. The competent clones proliferate and start producing their antibodies, giving so a first defense to the organism, after just a few days from the exposure to the antigen. That is the primary response. It is specific, but at low levels of affinity. Well, I think I will stop here and continue in the next post. gpuccio
Gpuccio Re #169 Thanks to you I am learning a lot of biochemistry. Debate is very powerful learning process if properly facilitated – in this case we are doing well at facilitating ourselves. The example of antibody maturation is rather interesting. If I understand it correctly, it really shows up the difference between non-design approaches and design approaches to solving problems. In particular, because it is a restricted problem, it makes clear the negative nature of the evidence for the design – at least in this case. As I am not a biochemist I will check my understanding of some basics: 1) When an antigen enters a vertebrate system then after an initial generic response the vertebrate immune system creates an antibody that is specific to that antigen (it locks on to it and marks it for destruction by T cells) 2) This specific antigen is created by vastly increasing the usual rate of mutation in some specific genes and areas of genes (but not in the germ cells so the mutation is not inherited) 3) Once a successful antibody is created presumably there is then some feedback to the genes to increase production of the successful antibody I will also assume that 4) The target area (the combinations that will successfully lock onto the antigen) is so small and the domain so large that it is effectively impossible to reach the target by simply spinning the wheels and hoping to come down with the right combination. 5) There no value in “partially fitting” the antigen – a mutation either fits or it doesn’t Now we have an interesting scientific problem. A non-design approach is to explore how that search might work. Options include: * The antigen somehow causes mutations which are closer to the target * The target area of all possible antigens is not randomly distributed around the domain but is clustered – so if the starting position(s) for mutations was in the cluster it would greatly increases the chances of hitting the target * Mutations are not random (random in the sense that any position in the relevant genes is equally likely to mutate and is equally likely to chance to one of the other three base pairs). And this lack of randomness increases the chances of hitting a target. There may be other options to explore but these exist and I understand they are currently being actively researched. Now what about the design approach? You are positing that an intelligence intervenes every time an antigen enters an individual vertebrate to direct the search for an antibody. The evidence for this intervention is the improbability of all current non-design solutions. Now here is the crunch. What evidence is there for this design explanation other than the failure of known non-design explanations? When talking of bacterial flagella etc this point gets obscured. It is possible to argue for irreducible complexity, and talk about codes and symbols etc, as positive evidence for design, because you are talking about the mystery of the whole transcription system etc. But in this case things are a simpler. We are taking most of biology and the transcription system for granted. As far as this issue is concerned all the rest might be designed or not. We are only asking how do mutations end up creating the appropriate antibodies? You can see the negative nature of the design explanation by imagining that someone comes up with a plausible non-design explanation. What evidence remains for a design explanation? None. By finding a plausible non-design explanation you have removed all evidence for design. Ergo the only evidence for design was the lack of a plausible non-design explanation. Mark Frank
djmullen: How boring. Typical behavior once again. When the discussion is no more rewarding, just: a)Make extreme and false affirmations about what the other has said: "Let’s see, all I have to do to convince you is to basically document every single evolutionary step from the first reproducing chemical to you, me and every other protein in every other organism on the planet?" I would have been happy with a single example for one protein, just as a start... 2)Shoot a series of irrelevant and rather silly questions as a form of personal attack: "Fine. Let’s see you do the same thing. You say that some proteins could not possibly have been produced by a series of small evolutionary steps. OK. Which proteins?" Practically all of them. "If they weren’t produced by evolution, then what steps were taken to produce them?" Please see my detailed asnwers to Mark at #134 and #169. In brief, possibly the same mutations and selections hypothesized by darwinists, but intelligently guided, and not random. "Was some mega-mutation necessary to produce some particular protein, changing hundreds or thousands of DNA bases at once?" Not necessarily. But it's always a possibility. "Which bases were changed?" Those which, according to the design, needed to be changed. "What were they before and after the change?" Before the change they were those of the ancestor, after the change those of the target protein. "When did this happen?" In the course of natural history, as new species appeared. But some of my IDist friends would prefer the theory of frontloading, which I personally don't like very much. You choose. "How was it done?" I don't know. That's open to research. But again, if you read my answers to Mark, you will find some suggestions. "How was the correct pattern found?" By design. Again, in my answers to Mark you will find two different possibilities: a) the designer knew the solution; b) the designer knew the searched for function, and could measure it. "Please document your claims. Saying “The Designer” is not documenting anything unless you can tell me about The Designer." Then how is it that you are here, on an ID blog? Let me understand, you are listening to people who have always stated clearly that they have, at present, no scientific knowledge about the designer, and you go on discussing in detail many serious issues of the theory with them, and now suddenly you come out with such fundamental disbelief in its basic premises? Tell me, what's happened? "Who is he, where is he, how did he get this info?" My personal (non scientific) opinion? He is a God, He is transcendent, but also very much present in His creation, He can easily get all the info He needs. But I would not discuss these non scientific points here. "Do you have any independent evidence of his existence other than your protein sequences which you claim can’t be produced by evolution?" A lot. But it's not scientific evidence, at least not yet. But, at the scientific level, I am very happy with my protein sequences which can’t be produced by evolution. Very happy indeed. 3) Bring the personal attack at another level, challenging the other's competence and doubting his "authority", possibly tying him to some other supposedly uncomfortable people: "While we’re on this subject, what is your expertise in proteins? I’ve been reading your messages about BLASTing protein sequences and I’m uncomfortably reminded of Salvador Cordova’s Avida fiasco. Have you had any training on proteins or the BLAST software?" First of all, my expertise in proteins is not your business. This is a blog, and not a scientific journal. And I, and all the other sincere people who come here, have never spoken "from authority", or requested any academic title or position from others. Second, I feel honored of being compared to Salvador Cordova. And third, as I have told that many times here and it's not a secret, I am a Medical Doctor. And finally, I am not really looking forward to your response. gpuccio
gpuccio: Let's see, all I have to do to convince you is to basically document every single evolutionary step from the first reproducing chemical to you, me and every other protein in every other organism on the planet? Fine. Let's see you do the same thing. You say that some proteins could not possibly have been produced by a series of small evolutionary steps. OK. Which proteins? If they weren't produced by evolution, then what steps were taken to produce them? Was some mega-mutation necessary to produce some particular protein, changing hundreds or thousands of DNA bases at once? Which bases were changed? What were they before and after the change? When did this happen? How was it done? How was the correct pattern found? Please document your claims. Saying "The Designer" is not documenting anything unless you can tell me about The Designer. Who is he, where is he, how did he get this info? Do you have any independent evidence of his existence other than your protein sequences which you claim can't be produced by evolution? While we're on this subject, what is your expertise in proteins? I've been reading your messages about BLASTing protein sequences and I'm uncomfortably reminded of Salvador Cordova's Avida fiasco. Have you had any training on proteins or the BLAST software? djmullen
Mark (#156): I come back to you with real pleasure. Yous posts are always intelligent, sincere and creative. I agree with you that frameshift mutations as a source of functional variation should not ba a classical darwinian concept. The fact that they are pursuing it so earnestly, apparently even after the debunking of nylonase, is just a sign, for me, of how desperate they are. Apparently, not all darwinist biologists share JayM's faith in an unfragmented proteome. "So if we come across a protein which is radically different from others and there is no evidence that frameshift was involved, then there is a problem knowing how that protein was generated." You bet! "But it is a problem for all theories that assume DNA is created by modification of parent DNA. Unless you believe God inserts complete DNA strings into cells from time to time then, this equally a problem for a non-Darwinian theory." Yes and no. The implementation of information remains a problem, but if you have not to look for that information, or at least if you can look for it with intelligent means, the main obstacle is overcome. Let's take an example. In the case of antibody maturation, which I have cited, the intelligent procedure embedde in the immune system realizes a very significant increase of the affinity of the primary antibody response in just a few months. As I have said, that is attained through targeted hypermutation and intelligent selection through direct measurement of the affinity of the new clones to the antigen. And still, that is a rather indirect method, because the system does not know in advance the sequence to be found, but only the function to be measured. But let's imagine that an intelligent force, which knows the sequence to be obtained, like the weasel algorithm, could just "fix", by some biochemical, or quantum, method which we presently don't know, any random variation which is correct. Or, better still, just induce the right variation directly, "guiding" the apparently random events at the molecular level. Wouldn't that make the implementation of the information extremely easier? You see, the real problem is the search. It's the search which is completely out of any possibility, without any intelligence and any information about the result to be attained. But, as GAs show, if you have a good oracle, you are in all another situation. And, in a sense, God could well insert complete DNA strings into cells from time to time. After all, that's what plasmids and transposons, and even ERVs, seem to do. Why couldn't God do the same, directly or indirectly? Frankly, frameshift mutations are so problematic that I don't believe that even the Designer would use them. But I think there is some literature about genes which can be read in two different ways, both functional, through a different (frameshift) start of transcription (but I can't remember the details now). Well, that would be tremendously smart design! Like some brilliant enigmas, or some paintings by Escher. gpuccio
CJYman:
Since CSI is a measurement of finding a specified or pre-specified pattern at better than chance performance, problem specific information is required to find it (according to NFLT). Problem specific information is measured as active information within these two papers. Do you see the connection?
One problem with the connection that you're drawing is equivocation on the word "chance". Under the active info framework, chance means a uniform distribution. Under the old CSI framework, it meant any distribution, including the distribution conferred by RM+NS. Dembski in NFL: Chance as I characterize it thus includes necessity, chance (as it is ordinarily used), and their combination. Dembski in his Specification paper: Moreover, H, here, is the relevant chance hypothesis that takes into account Darwinian and other material mechanisms. So products of evolution have no CSI, by definition, in spite of the fact that evolutionary processes have boatloads of active information. R0b
JayM (# 157): I appreciate your efforts, but I am afraid I can only repeat for you what I have already said about djmullen: your characterization is inaccurate. "The underlying partial homologies are evidence that supports MET mechanisms of incremental change." Absolutely not. There is nothing in the homologies which shows incremental change in the sense of a gradual modification of the function. The conservation of domains has a functional basis, but sometimes the same domains and 3d strictures are obtained, in different species, with very different primary sequences. And there are a lot of different domains, and different foldings, whose biochemical nature is completely different. If you were familiar with the biology of proteins, you would never say the things you say. It is true that darwinists affirm that proteins have formed by incremental change in function, but that is only one of the many just so stories. There is no empirical confirmation for that, and a lot of theoretical impossibilities. Why do you think that ID exists at all? "You can’t consider just the full protein as though it came into being all at once." I am not saying that it came into being at once, I am saying that it was designed. But you seem to ignore that parts of one protein are not functional. You need a whole protein, with its correct folding and domain, to get a function. Very big proteins can be multidomain, and can be deconstructed in different functional portions, but proteins like myoglobin are very compact globular proteins, and the whole sequence must be there. That's why myoglobin has approximately the same length in most species, with only minor variations. That is true of many important proteins. "An evolutionary biologist would probably say it came from a protein that was about 153 amino acids long. Or from a very slightly different protein that was about 154 amino acids long. Unless ID researchers can show this is a mathematical impossibility, the MET mechanisms remain scientifically credible." We all know what evolutionary biologists would say, and that's why we are here in ID. A 153 aminoacids myoglobin would still be a myoglobin, with the same 3D structure and the same function. Not all myoglobins are "exactly" 154 aa (that is the length of the human form, and of most examples). Some are of 146, 147, 149 and so on. But those are minor difference, in part probably due to neutral evolution. But the protein is essentially the same. But there is no model which can explain the step by step derivation of myoglobin from some other simpler protein with a different function. And remember, such a model should explain not only the variations in aa sequence, but also the variations in function which allow the selecrion and expansion of each single intermediate. Again, that's why I am in ID, and not in the darwinian field. You can stay where you like, or just go on observing and posting. "Could you point to some literature cites that support this claim? I have looked for such information for quite some time, because it seems that this is a very promising area for ID research. I haven’t found any clear results that show that genome space is fragmented in this way." There is no need to support my claim. The data are there, for anybody to look at them. I have just showed you a few examples. Go to SCOP, do some blast homework, and you will see for yourself. Everybody who works with proteins knows what I am saying. "While I appreciate your effort, that isn’t a particularly rigorous process for tracking homologies. You might identify a new homology with that approach, but failure to find one after a handful of searches doesn’t invalidate the entire corpus of evolutionary molecular biology." Blastp is one of the most currently used tools to look for homologies. I have also used ClustalX, with the same results. Do you really believe that there is a way to find significant homologies between two proteins like dinG and clpA? Not even darwinists are so smart. And the entire corpus of evolutionary molecular biology is based on false assumptions, and that's what invalidates it. "Taking humans as an example, if MET mechanisms are sufficient you should see similar molecular constructs, for a certain level of complexity, in chimpanzees." Again you miss the point. There are proteins which are almost identical in extremely different species. That is easily explained as functional conservation, but tells us nothing about the emergence of the function. Neutral evolution, and especially synonymous mutations, are interesting tools to try to understand variations due to time and random errors, but tell us nothing of the mechanism of function generation. "MET does not claim that the whole protein came into being in one fell swoop. We have to address what MET actually says, not what we wish it said." I am still waiting for them to say how that happened. Can you help? "I will be delighted if you have really identified the disconnected islands that disprove MET, and if you have you should definitely publish your results, but I don’t see support for that in what you’ve presented so far." If you can't see what is under the eyes of all, I really can't help you. "ID opponents make some strained arguments, but one that is valid is that anyone who discovered what you claim here would be world famous. It stretches credibility to think that everyone with access to the same tools you used could be Expelled and prevented from publishing." I would be very pleased if I could become famous so easily. Unfortunately, one does not become famous for saying what is known to all, and is easily retrieved in public databases. I am just discussing the interpretation of what is known to all. The only reason why I spent all this time giving you some real example, is that you (and djmullen) were saying completely wrong things with a certainty which, I hope, could derive only from your ignorance of the matter. gpuccio
CJYman:
The only formal proof presented is that CSI won’t generate itself through only chance and law and this is implicit within these two papers.
Just for clarification, are you saying that this formal proof is implicit in these two papers? (An implicit formal proof sounds like an oxymoron to me.) If not, where is this formal proof to be found? Thanks. R0b
Prof_P.Olofsson:
The recurring use of the uniform distribution in many attempts to apply math to biology is a model assumption must be argued just like any other model assumption.
Indeed, Olle Haggstrom's response to the active info approach is titled "Uniform distribution is a model assumption". The question asked by Haggstrom and others is, why should we always expect uniform randomness? In fact, it's an impossible expectation. If everything were characterized by a uniform distribution, then that would be a non-uniform distribution of distributions. R0b
Prof_P.Olofsson:
Informal claims such as “it is as difficult to find a search as it is to find a target” are not supported by the paper unless you make a lot of arbitrary assumptions.
One of the arbitrary assumptions involved in measuring active information even violates Dembski's own repeated warning. He has told us several times that "how we measure information needs to be independent of whatever procedure we use to individuate the possibilities under consideration." And yet, the active information measure depends very much on how we individuate the possibilities. R0b
CJYman[162], Let me clarify: I am questioning the claims that the "search for a search" paper is pro-ID as was claimed in the introduction. In order to be considered pro-ID, it would have to have some implications for biology. I ask, for example, (1) what is the rationale behind claiming that search algorithms are chosen according to probability distributions, in particular the K-W distribution [154]? and (2) if they are, why do these distributions have to be uniform [159]? and point out (3) even if they are uniform, they are still quite likely to beat random search [155]. Prof_P.Olofsson
Hello Prof. Olofsson, I'm not sure that I follow you, however if what you are stating is true and actually has practical effect (ie: actually happens in the real world), why doesn't anyone just show that background noise (chance) and an arbitrary set of laws (set of laws collected without any consideration for future results -- absent foresight) will produce systems which process signs/units and evolve into greater and greater specified complexity, and just falsify ID theory and be done with it. If evolution is so powerful absent previous foresight, why do programmers bother programming boundary conditions for future known targets into the evolutionary algorithms used to solve problems (ie: max efficiency antenna shape)? CJYman
JayM: "That’s great, but it still seems to cede the field of biology to modern evolutionary theory." It cedes biology to evolution and it takes evolution for itself. JayM: "An aside due to my inner geek: Is there a formal proof that CSI cannot be generated from evolution simulations?" The only formal proof presented is that CSI won't generate itself through only chance and law and this is implicit within these two papers. Since CSI is a measurement of finding a specified or pre-specified pattern at better than chance performance, problem specific information is required to find it (according to NFLT). Problem specific information is measured as active information within these two papers. Do you see the connection? CSI can be transferred through evolution, but will never be generated from scratch by evolution. Why? Because these two papers show that active information continually regresses to higher and higher levels of search. So, if active information is necessary for the transferring of CSI, then evolution can't generate CSI. It can only transform active information into CSI. JayM: "I don’t believe that “all simulations of evolution require foresight.” We’ve discussed several different genetic algorithms in recent threads here and Dawkin’s Weasel is the exception rather than the rule." My apologies for not being clear. If we are discussing any simulation relevant to biology, then evolution is the discovery and increasing of CSI. Can you provide any simulation of biological evolution which discovered CSI from initial conditions that were not programmed with any consideration for future results? JayM: "I’m not sure how this is relevant to the two papers, though." It's more relevant to the NFLT, where it is important that problem specific information be incorporated into the behavior of the algorithm (almost an exact quote), thus searching for a given pattern becomes as difficult as searching for the problem specific information necessary to discover that pattern at better than chance performance. These two papers merely provide a metric for measuring problem specific information and showing that it regresses to higher and higher level searches. The relevance to ID is that we seem to have only three choices which I've already discussed ... an infinite regress of active info., a falsification of the paper by showing active info generating through background noise and arbitrary collection of laws, or use a foresighted system to program the initial conditions -- actually incorporating problem specific information through knowledge of the future problem to be solved. CJYman
And, finally, I know that my comments [154,155,159], although directly related to the Dembski & Marks paper, are aside of the main topics of the thread. If somebody is interested, we may take it off the air as well. Prof_P.Olofsson
And, to [154] and [155], I'd like to add a third question: Even if we adopt the model that searches are being searched for via a probability distribution, how can we argue that the distribution must be uniform? Even if you don't believe the darwininan search can be successful, you have to admit that it exists and is reasonably well understood: DNA is replicated, mutations cause imperfections which leads to a new genotypes (vastly simplified). How would you argue that this search is no more likely that one that is uniform over the entire sequence space ("blind search") which cannot even be given a reasonable biological or phyiscal meaning? The recurring use of the uniform distribution in many attempts to apply math to biology is a model assumption must be argued just like any other model assumption. Prof_P.Olofsson
CJYman @151
Thanks for the detailed reply. It appears that I do understand you correctly. In order for Dr. Dembski to present these two papers as support for ID, though, he (or someone) needs to make the links you hint at explicit.
I see where you are coming from and I think that the reason he doesn’t make the connection explicit is because it seems that these papers are more of a response to inflated claims about how evolutionary algorithms can account for biodiversity. These papers were merely showing that the extent to which an evolving system increases probabilities is proportional to the improbability of finding that system (evolutionary algorithm) in the first place.
I still don't see this from the papers themselves. When I saw the title about peer-reviewed papers supporting ID, I was excited. After reading them, I can see it will be all too easy for ID opponents to dismiss them as having no biological relevance. I would love to see a third paper that ties this all together.
The papers [along with the NFLT] explicitly detail the fact that in order to increase probability of generating a given pattern, a set of laws matching search space structure and search procedure, which is even more improbable, must first be generated. Thus, evolution is not, as some have claimed, a free source of information — information being an increase of probability. This increase of probability, labeled active information in the paper, is merely pushed back to higher and higher levels.
Again, this argument seems to be that the physical laws of the universe are designed, but that, given them as they are, evolutionary mechanisms are sufficient to explain biological diversity. I suspect this is not Dr. Dembski's position. ;-)
What the paper does show is that this increase in probability seen in the operation of evolution still needs an explanation. Evolution is not an explanation of its own success at increasing probability anymore than an EA can create itself absent foresight.
That's great, but it still seems to cede the field of biology to modern evolutionary theory.
If a computer program based on only chance and law (absent previous foresight) won’t produce CSI, why would an infinite amount of such programs be a better explanation rather than foresight, which is indeed characterized by its generation of CSI?
An aside due to my inner geek: Is there a formal proof that CSI cannot be generated from evolution simulations?
IOW, all simulations of evolution require foresight so, barring this universe being the only option, on what grounds could someone say that the "real deal" requires no foresight as a scientific fact or that there is "no evidence for teleology within life" or that "ID is unscientific?"
I don't believe that "all simulations of evolution require foresight." We've discussed several different genetic algorithms in recent threads here and Dawkin's Weasel is the exception rather than the rule. I'm not sure how this is relevant to the two papers, though. JJ JayM
gpuccio @150
This overstates the case somewhat. While some full proteins are species specific, most do have underlying partial homologies. It is those that are used as evidence for common descent.
And so?
The underlying partial homologies are evidence that supports MET mechanisms of incremental change. You can't consider just the full protein as though it came into being all at once. Because even those species specific proteins have significant overlap with proteins in other species, evolutionary biologists can and do consider them consistent with MET.
Even if you group the proteins which have strong homologies, you still have myriads of separated islands.
Could you point to some literature cites that support this claim? I have looked for such information for quite some time, because it seems that this is a very promising area for ID research. I haven't found any clear results that show that genome space is fragmented in this way.
Let’s take the example of myoglobin. At some point in natural history, this protein appears. It is about 154 aminoacids long. Where did it come from?
An evolutionary biologist would probably say it came from a protein that was about 153 amino acids long. Or from a very slightly different protein that was about 154 amino acids long. Unless ID researchers can show this is a mathematical impossibility, the MET mechanisms remain scientifically credible.
I’m afraid that “extremely varied” isn’t supported by the empirical evidence.
As an answer, I quote again from my previous posts: from my post # 52: "The proteins we do know (and we know a lot of them) are really interspersed in the search space, in myriads of different and distant "islands" of functionality. You don’t have to take my word for that. It’s not an abstract and mathematical argument. We know protein sequences. Just look at them. Go, for example, to the SCOP site, and just look at the hyerarchical classification o protein structures: classes (7), folds (1086), superfamilies (1777), families (3464). Then, spend a little time, as I have done, taking a couple of random different proteins from two different classes, or even from the same superfamily, and go to the BLAST site and try to blast them one against the other, and see how much "similarity" you find: you will probably find none.
While I appreciate your effort, that isn't a particularly rigorous process for tracking homologies. You might identify a new homology with that approach, but failure to find one after a handful of searches doesn't invalidate the entire corpus of evolutionary molecular biology. Identifying proteins that are a result of a genome in an unreachable area of genome space will require extensive, painstaking work to identify candidates and trace them back to their progenitors in a recent common ancestor with another species. Taking humans as an example, if MET mechanisms are sufficient you should see similar molecular constructs, for a certain level of complexity, in chimpanzees. Other great apes should show less similarity, and other primates less still. If you find a relatively complex construct (protein, gene, etc.) without a corresponding similarity, you'll have a solid candidate for further research. I would love to see ID researchers taking this tack.
In the genome of E. coli, a well known bacterium, we can find at least 2018 proteins described, grouped in 525 groups of similar proteins. The biggest group includes 52 proteins, while the smallest groups include only two proteins each, and there are 291 of them. Do you still think that "extremely varied" is not supported by the empirical evidence? What empirical evidence are you suggesting?
There are certainly a lot of different proteins, but again you can't assert that they can't be arrived at through MET mechanisms without looking at the probably precursors and overlapping sequences in other organisms. MET does not claim that the whole protein came into being in one fell swoop. We have to address what MET actually says, not what we wish it said.
Again, we in the ID camp need to be very careful to address the real positions held by ID opponents
That’s exactly what I am trying to be.
rather than strawmen that not only can be easily knocked down but that can also be used to distract endlessly from our core message.
I don’t think I am using any strawmen. You can see that I am very precise and explicit in my arguments and in my answers, trying to address exactly what others, including you, are saying. If sometimes I misunderstand what the other is saying, I am ready to apologize.
Your work with BLAST does not, unfortunately, support your claims. Instead of looking at entire proteins or genes, you must consider subsequences shared with likely precursors and other descendents from those precursors. This is exactly what molecular biologists and bioinformaticists do on a daily basis. Numerous homologies have been identified and the literature contains more every day. I will be delighted if you have really identified the disconnected islands that disprove MET, and if you have you should definitely publish your results, but I don't see support for that in what you've presented so far. ID opponents make some strained arguments, but one that is valid is that anyone who discovered what you claim here would be world famous. It stretches credibility to think that everyone with access to the same tools you used could be Expelled and prevented from publishing. JJ JayM
Gpuccio This business of frameshifts is rather interesting. I wish I had time to really understand the papers. But there is a broader thing that concerns me. I can't see that frameshift is a Darwinian concept. It is just a way generating a DNA string from a parent that gives a radically different protein string. All other known methods, it appears, do not generate radically different proteins. So if we come across a protein which is radically different from others and there is no evidence that frameshift was involved, then there is a problem knowing how that protein was generated. But it is a problem for all theories that assume DNA is created by modification of parent DNA. Unless you believe God inserts complete DNA strings into cells from time to time then, this equally a problem for a non-Darwinian theory. In fact frameshifts could be interpretated as evidence for ID using classic ID arguments. If, as you say, it is highly improbable that a frameshift will generate something useful, then the prescence of over two hundred proteins created by frameshift is extraordinarily improbable - unless big G was controlling the shifts. On the other hand if the proteins were generated by some, as yet not understood method of moderating parents, then it might be a much more plausible mechanism. Mark Frank
CJYman and others, Even if we do accept all the assumptions about "searching for a search," shouldn't we compute probabilities rather than talk about averages? I'll give an example following Dembski&Marks's construction on page 2 in the "search for a search." Let Omega={0,1} and search for 0. If we choose at random, the probability is 1/2 to find it. If we first choose a probability distribution on Omega at random, the probability to find 0 is still 1/2 (compute the expected value of p where p is uniform on [0,1]). That's the "conservation of uniformity" in this example. Now instead ask "What is the probability to do better than random search?" This is the probability to find a p that is greater than 1/2 and as p is uniform on [0,1], there is a 50% chance to beat random search. If we "search for a search" twice, we have a 75% chance to beat random search at least once. And so on. In other words, it is easy to beat random search by repeatedly randomly searching for searches. Now assume that Omega is huge but still finite, Omega={x1,...,xn} where we search for x1. [Note that x1 may represent an entire search string so finding x1 in one step may mean that we have found some target in a finite number of steps.] The probability to beat random search turns out to be about 37% [approximately = exp(-1)], regardless of n, and searching twice gives us about 60% chance to succeed. Conclusion: If we "randomly search for a search" at least twice, we are more likely than not to beat random search. Prof_P.Olofsson
CJYman and others, The only logic I can see in how the "search for a search" paper supports ID would be to claim: "Either (a) the darwinian search algorithm was chosen according to the Kantorovich-Wasserstein distribution or (b) it was desiged." Note that the K-W distribution on M(Omega) depends on what metric you choose on Omega so there is no unique way to interpret "randomly choosing a search." Informal claims such as "it is as difficult to find a search as it is to find a target" are not supported by the paper unless you make a lot of arbitrary assumptions. Prof_P.Olofsson
CJYman,
These papers were merely showing that the extent to which an evolving system increases probabilities is proportional to the improbability of finding that system (evolutionary algorithm) in the first place. The papers [along with the NFLT] explicitly detail the fact that in order to increase probability of generating a given pattern, a set of laws matching search space structure and search procedure, which is even more improbable, must first be generated.
If I may layman-ize: Darwinian processes are generally limited to searching for short range targets in local genome space. In order to search for long range targets a set of specific conditions relevant to the long term ("a set of laws matching search space structure and search procedure") must first be found as well. And a Darwinian search in nature is limited in scope to that which increases overall survivability, not functionality for merely functionalities sake as with engineering projects using GAs. Does that summarize correctly? Patrick
sparc, You should hang out with Americans more often. You will pick up common sense and some humor. I know you have referred to us as dummkopfs but you read things too literally. My use of the term SFCSI was meant to show the absurdity of Mark Frank's objections and my guess you too if you deny FCSI. Do you deny the reasoning of FCSI or that DNA is information or a code? If so then you can join us as dummkopfs. For those who do not know, sparc is a biologist who is unable to provide any critical analysis on the evolution debate. And as I said earlier on this thread we need more like sparc here who only can contribute inane remarks when they disparage ID. Go Sparc. jerry
JayM: "Thanks for the detailed reply. It appears that I do understand you correctly. In order for Dr. Dembski to present these two papers as support for ID, though, he (or someone) needs to make the links you hint at explicit." I see where you are coming from and I think that the reason he doesn't make the connection explicit is because it seems that these papers are more of a response to inflated claims about how evolutionary algorithms can account for biodiversity. These papers were merely showing that the extent to which an evolving system increases probabilities is proportional to the improbability of finding that system (evolutionary algorithm) in the first place. The papers [along with the NFLT] explicitly detail the fact that in order to increase probability of generating a given pattern, a set of laws matching search space structure and search procedure, which is even more improbable, must first be generated. Thus, evolution is not, as some have claimed, a free source of information -- information being an increase of probability. This increase of probability, labeled active information in the paper, is merely pushed back to higher and higher levels. What the paper does show is that this increase in probability seen in the operation of evolution still needs an explanation. Evolution is not an explanation of its own success at increasing probability anymore than an EA can create itself absent foresight. From reading the paper, we can see that we have a few options ... infinite regress of active info. which is akin to "turtles all the way down" and never truly explains increases in information (probability). That "explanation" merely explains away by handwaving. The other two options are merely chance and law, or foresighted systems. As I have already explained, no one has shown background noise (chance) and an arbitrary set of laws (set of laws with no regard for future results) to be up to the task of generating active information. If someone showed such an experiment, the math within these papers would be falsified, since they forbid such increases in probability. However, we do know through our own experience and through the creation of AI systems that foresight using systems (systems which model the future and generate targets) do exist and do increase probabilities of generating specified or pre-specified patterns. I have explicitly, yet briefly, outlined the ID connection in my post #122. JayM: "Unless one subscribes to the multiverse theory, it isn’t clear that the physical laws of this universe are the result of any kind of search (or search for search) process." The multiverse theory provides no more of an explanation for life and evolution than it provides an "explanation" for this conversation we are having. It is merely a chance of the gaps "explanation" and there is no criteria for when we can invoke it and when we can't. Can we invoke infinite chance to explain the results of all scientific tests? If so, then we would never discover anything since the proper explanation is that "the multiverse did it." Sure, using an infinite resource of chance comes in handy when trying to explain everything and anything, but is it the best explanation? This is where other ID argument come in, such as CSI. If a computer program based on only chance and law (absent previous foresight) won't produce CSI, why would an infinite amount of such programs be a better explanation rather than foresight, which is indeed characterized by its generation of CSI? As to the universe being the result of search, as long as it is one of any number of possibilities in "quantum chaos," then it is a search as defined by these two papers and as I previously explained in my post #121. At least our universe is confined to an extremely small set of possible mathematically described universes which even allow life and evolution. If, however, our universe is the only option, then no there is no search involved but this would raise other extremely intriguing questions. However, the papers would still show that any simulation of evolution requires the previous raising of probabilities before the evolutionary algorithm even starts. As we all know, intelligent humans raise the probabilities by introducing problem specific information into the behavior of the interaction between search algorithm and search space, by programming law and chance in such an improbable way as to solve a given problem in the future. That basic concept has been proven in the NFLT and extended in these two papers. Moreover, human intelligence uses its foresight to add problem specific information. IOW, all simulations of evolution require foresight so, barring this universe being the only option, on what grounds could someone say that the "real deal" requires no foresight as a scientific fact or that there is "no evidence for teleology within life" or that "ID is unscientific?" CJYman
JayM: # 147 "This overstates the case somewhat. While some full proteins are species specific, most do have underlying partial homologies. It is those that are used as evidence for common descent." And so? Even if you group the proteins which have strong homologies, you still have myriads of separated islands. Who has said that an island must be made of only one protein? Even if myoglobin is present in many living beings, it is still completely different form insulin, or from c-myc. And who has ever criticized the concept of common descent? I accept it. But what has common descent to do with traversing the ocean of possibilities? Let's take the example of myoglobin. At some point in natural history, this protein appears. It is about 154 aminoacids long. Where did it come from? The fact that, after its appearance, it is maintained in most species means only that the function has been exploited in most species. But the traversing is necessary to get to that function, the first time. That's why I said that djmullen's statement that DNA replication was not a search was suspiciously trivial. Maybe he (and you) have not well considered that the search happens before, when you have to find the information, and not when you simply copy or transmit it? "I’m afraid that “extremely varied” isn’t supported by the empirical evidence." As an answer, I quote again from my previous posts: from my post # 52: “The proteins we do know (and we know a lot of them) are really interspersed in the search space, in myriads of different and distant “islands” of functionality. You don’t have to take my word for that. It’s not an abstract and mathematical argument. We know protein sequences. Just look at them. Go, for example, to the SCOP site, and just look at the hyerarchical classification o protein structures: classes (7), folds (1086), superfamilies (1777), families (3464). Then, spend a little time, as I have done, taking a couple of random different proteins from two different classes, or even from the same superfamily, and go to the BLAST site and try to blast them one against the other, and see how much “similarity” you find: you will probably find none. And if you BLAST a single protein against all those known, you will probably find similarities only with proteins of the same kind, if not with the same protein in different species. Sometimes, partial similarities are due to common domains for common functions, but even that leaves anyway enormous differences in term of aminoacid sequence.” from my post # 118: "In the genome of E. coli, a well known bacterium, we can find at least 2018 proteins described, grouped in 525 groups of similar proteins. The biggest group includes 52 proteins, while the smallest groups include only two proteins each, and there are 291 of them." Do you still think that "extremely varied" is not supported by the empirical evidence? What empirical evidence are you suggesting? "but the underlying amino acid sequences are not all over the map as you imply." That is simply wrong. They are. Go back to the two proteins I took at random from the E. coli genome. They are more than 700 aminoacids long, each one of them. In the best possible alignments, four different ones, you can find only 40 identities (not consecutive). Theose two proteins are "all over the map". And the same can be said if you take any other pair of proteins from different groups in the E. coli genome, or from different families or superfamilies in the SCOP classification. And even proteins with homologies are very different one from the other. Homologies are partial. Some are very partial. Choosing a conserved part of the c-myc sequence (a very important transcritpion factor), i have performed a blastp serach with a sequence of only 7 consecutively conserved aminoacids. 7 aminoacids is not much, they correspond to a search space of "only" 10^9. The search was done against the whole database of known proteins, and guess what? Only c-myc molecules had that exact sequence in their primary structure. You had to drop at least one aminoacid to find that sequence in other kinds of proteins. Can you see how terribly specific even a small sequence of aminoacids can be? Just consider that the specific immunological response in higher animals and humans is targeted against very small aminoacid sequences, the epitopes, whose length is usually less than 10 aminoacids, and yet those small sequences are so specific that they are the basis for our defenses. So, you can see that protein sequences are at the same time: a) extremely varied and interspersed on the whole map of possible sequences b) extremely specific, with very short sequences representing often a definite signature for one protein. "Again, we in the ID camp need to be very careful to address the real positions held by ID opponents" r That's exactly what I am trying to be. "rather than strawmen that not only can be easily knocked down but that can also be used to distract endlessly from our core message." I don't think I am using any strawmen. You can see that I am very precise and explicit in my arguments and in my answers, trying to address exactly what others, including you, are saying. If sometimes I misunderstand what the other is saying, I am ready to apologize. gpuccio
JayM: # 146 "The fact that most evolutionary mechanisms result in tiny changes between generations is not just something that djmullen is “affirming”, it is a repeated empirical observation." What do you mean? Obviously a sinle point mutation changes amnly one nucleotide, but if it is a frameshft or a stop mutation a whole protein can change. Deletions and inversions can cause great changes in one signle step. The problem is that all those steps are random, and that's why a frameshift mutation , which is the equivalent of a blind long leap into the ocean, has practically no hope to find an "island". I have the impression that you and djmullen are maybe suggesting that a change can be "small" in regard to the whole genome. But we have not to reason about the whole genome. That is senseless. Our unit of reasonong must remain the single protein coding gene. There are single point mutations which are incompatible with life. Much more difficult would be to find a single point mutation which gives a reproductive advantage to a complex organism. Empirical observation of "useful" mutations is limited to microevolutionary events, like antibiotic resistance and even nylonase. In those observed events, the edge of evolution seems to be, at present, about two coordinated useful aminoacid substitution. That is very little to explain how you can get to a new protein of 700 aminoacids. Therefore, there is no empirical observation of mechanisms which can eliminate the need for a blind search in the ocean of sequence possibilities. "but even hinting that djmullen’s characterization is inaccurate suggests a lack of understanding that can be exploited by ID opponents. We need to be careful in our public statements." djmullen's characterization is inaccurate. I take full responsibility for this statement. And I think I am very careful in my public statements (which, obviously, does not mean that I cannot be wrong, like anybody else). "The obvious objection to this suggestion is that we’re being asked to prove a negative." We have nothing to prove. What darwinists are assuming is obviously incredible. It's them that have to provide some example or model to make it at least debatable. Otherwise, they have no theorym they have no model, and especially they have no empirical support. Which is exactly the case. "This would be a true vindication of The Edge of Evolution." The edge of evolution is already vindicated. But, obviously, we can do even better, and we will. gpuccio
gpuccio: Obviously I'm not an expert in biology like you, and I'm assuming that all the claims you make about these islands of functional proteins are based on published research but it is still hard to see how two research papers about the mathematics of search algorithms, which don't explain why they are relevant or applicable to biology, are actually applicable to biology, in fact as a comp scientist I found it hard to see how the first paper could be applied to any real world situation given the way the authors seem to misunderstand some of these algorithms. What is needed is for an expert in the field like you to take all this evidence on the awkwardness of these protein search spaces and the evidence presented by Dembski and Marks, write it up as a paper and submit it to a biology journal. I doubt it would get published because of the conspiracy but you would still be able to post it on the web as evidence of the way these darwinians are suppressing the truth. Laminar
gpuccio @144
2) All the proteins we can observe are extremely varied both as primary sequence and as function. Many of them are even species specific.
This overstates the case somewhat. While some full proteins are species specific, most do have underlying partial homologies. It is those that are used as evidence for common descent. I'm afraid that "extremely varied" isn't supported by the empirical evidence. There is certainly enough difference to make one question how non-intelligent mechanisms could be the cause, but the underlying amino acid sequences are not all over the map as you imply. Again, we in the ID camp need to be very careful to address the real positions held by ID opponents rather than strawmen that not only can be easily knocked down but that can also be used to distract endlessly from our core message. JJ JayM
gpuccio @144
If you go on affirming that evolution “mutates a very very tiny percentage of the parental DNA”, then you have to explain why proteins are so different and interspersed in the ocean of sequences.
The fact that most evolutionary mechanisms result in tiny changes between generations is not just something that djmullen is "affirming", it is a repeated empirical observation. Even "large" changes like duplication or frameshifts don't change the genome significantly in the next generation, although they can provide more variety in subsequent generations. There is much that ID proponents can legitimately argue against MET, but even hinting that djmullen's characterization is inaccurate suggests a lack of understanding that can be exploited by ID opponents. We need to be careful in our public statements.
Or show that in all cases bridges exist bewteen the islands of function where a selectable and growing function is maintained in all the intermediates, and the jumps between one intermediate and the others is always tiny enough that no important search has to be performed. I really doubt that anyone has ever done that, or will ever do that, even for one single important case (I mean, for one case which is not simply a microevolutionary, insulated case of one-two aminoacids substitution with immediate and definitive selection). Nobody can build such a model for a very simple reason: it does not exist.
I would turn this around and suggest that it is an excellent avenue for ID research. If it can be shown that no path exists, that would destroy MET. The obvious objection to this suggestion is that we're being asked to prove a negative. The problem is not, however, as intractable as that. There are a limited number of mechanisms in MET. There are a limited number of base pairs and amino acids. There is at least some historical record of what paths were actually taken. It would be difficult, but by no means impossible to show that a particular protein could not have arisen by MET mechanisms. This would be a true vindication of The Edge of Evolution. JJ JayM
CJYman @138
Please correct me if I’m misunderstanding you, but it seems that you’re saying that Dr. Dembski’s papers provide support for the idea that the universe is designed, but have nothing to say about the likelihood that the mechanisms of MET can account for all the biological diversity we see.”
They provide evidence that life and evolution are designed. They show that in order for a system (including evolution) to increase the probabilities of discovering a function/pattern or solve a problem, the higher order system which causes that evolution must be at least as improbable.
Thanks for the detailed reply. It appears that I do understand you correctly. In order for Dr. Dembski to present these two papers as support for ID, though, he (or someone) needs to make the links you hint at explicit. Unless one subscribes to the multiverse theory, it isn't clear that the physical laws of this universe are the result of any kind of search (or search for search) process.
So, as a foundational explanation, we need to look for something which can increase probabilities and foresighted systems are capable of that, however background noise and arbitrary collections of laws (set of laws created without any consideration for future results) are not up to that challenge.
That's what we have to prove rather than assert, though. MET mechanisms demonstrably have some ability to allow biological organisms to adapt within some limits, so we can't make a blanket statement that foresight is absolutely required. We need to find out what the limits are and prove them both mathematically and experimentally before we can claim to have positive evidence for ID theory. I hope that Dr. Dembski's papers are eventually seen as one small step in that direction. JJ JayM
djmullen (143): if you have followed all the previous discussions about prtoeins, you should already have the answers. To sum up: 1) Proteins are the essential material on which evolution has to work. 2) All the proteins we can observe are extremely varied both as primary sequence and as function. Many of them are even species specific. 3) The ocean of possible proetin sequences is, indeed, an ocean, and a very big one. Functional proteins are located in myriads of small separated islands in that ocean, as can be easily verified comparing their primary sequences, which are well known in great detail for a lot of them (just think of my example of two real proteins, completely different one from the other, with two different functions and 3D structure, and a conbinatorial space of more than 20^700). 4) If you believe that all the existing proteome is derived from some original ancestor, with a much smaller proteome and simpler functions, and by unguided darwinian means, that implies that the ocean of possible sequences has been traversed myriad of times in the course of natural history, and even more amazingly at OOL, in some mysterious prebiotic setting. 5) That's why Dembski and Mark's work about random searches is absolutely relevant, if you want to defend darwinian theory. I am not going into the details of their work, but relevant to biology it certainly is. 6) If you go on affirming that evolution "mutates a very very tiny percentage of the parental DNA", then you have to explain why proteins are so different and interspersed in the ocean of sequences. Or show that in all cases bridges exist bewteen the islands of function where a selectable and growing function is maintained in all the intermediates, and the jumps between one intermediate and the others is always tiny enough that no important search has to be performed. I really doubt that anyone has ever done that, or will ever do that, even for one single important case (I mean, for one case which is not simply a microevolutionary, insulated case of one-two aminoacids substitution with immediate and definitive selection). Nobody can build such a model for a very simple reason: it does not exist. And, to save darwinian theory form the paradise of inconsistent fantasies, that should be necessarily done not for one, but for all known cases of different proteins with different functions. I hope that answers more explicitly your questions. gpuccio
gpuccio: Although proteins are interesting, my main concern is with the subject of this thread, the Dembski and Marks papers and how they relate to evolution. As you say, when copying DNA, "...there is no search necessary." You even say this fact is so trivial, I don't even have to state it. So I ask again, what do Dembski and Marks' papers on searching through the vast search space of a particular genome have to do with evolution? So far as I know, evolution never does anything like that. It either does a direct copy of the parental DNA, which means no search at all, or it mutates a very very tiny percentage of the parental DNA which means it "searches" an area of the genome very very close to the known-good genome of the parents. I don't see any relevance of the Dembski-Marks papers to evolution at all. djmullen
Mark (# 136): I quoted that paper just to show you an example of how many proteins exist in the proteome which are specific to one single species. That means that they show no homology to proteins in other species. Even in the category of PDFs, they were citing 17,554 species specific proteins! That seems really a lot to me. That does not mean, obviously, that all the remaining proteins are conserved in all the species. It just means that they show some homology (certainly of various level and significance) to "some" other proteins in the tested genomes. In the interesting article linked in Paul Nelson's recent thread about the tree of life, we find the following interesting statement: "The battle came to a head in 2006. In an ambitious study, a team led by Peer Bork of the European Molecular Biology Laboratory in Heidelberg, Germany, examined 191 sequenced genomes from all three domains of life - bacteria, archaea and eukaryotes (complex organisms with their genetic material packaged in a nucleus) - and identified 31 genes that all the species possessed and which showed no signs of ever having been horizontally transferred." 31 genes. That's remarkable, isn't it? And about homology, please remember that the assumption that any level of homology in protein sequences is evidence of unguided derivation is only another myth of darwinism. First of all, homology is tested against the null hypothesis of randomness. Therefore, it is perfectly obvious that we will find some homologies bewteen proteins which have a similar function, for the same reason that all motorcars have wheels. That's certainly statistically improbable in a random assemblage of pieces, but it is certainly expected in a collection of machines which are designed to move. And finally, even if homology can show derivation, in no way it proves "unguided" derivation. So, to sum up, I think that the origin of "de novo" proteins does remain a completely unsolved mystery for darwinian theory. gpuccio
Mark: You are, as usual, a careful reader. # 135: I had obviously read with great attention the other papers mentioned in the Wikipedia page. I did not mention them explicitly in my post pusposefully, but I was referring exactly to them when I wrote: "The difficulty of obtaining even a single new protein, however, is evidently troubling darwinists ever more. That can explain the recent attempts to find an important role to frameshift mutations, which are the only hope for traditional darwinism to obtain radically new sequences in a “simple” way. So, although frameshift mutations are totaly unrealistic as a way to obtain functional information, darwinists are trying just the same to exploit them in their “models”, driven probably by sheer desperation." I hoped that someone mentioned them, so that we could deepen the discussion. Those papers are really the product of wishful thinking, or if we want of desperation. The only interesting thing in them is that the authors frankly aknowledge that, without the frameshift mutation echanisms, it would really be difficult for the model of darwinian evolution to explain the observed diversity of proteins. But the papers themselves are absolutely generic and speculative. They are of the kind: let's take all possible frameshifts in the human proteome, and just blast them against the existing genome, and see if we find some partial homologies somewhere more frequently than we would expect in a completely random system. In other words, they are only abstractly playing with thousands of sequences to give some support to an unbelievable assumption. The only real example of an observed frameshift mutation which would give a functional protein cited by them as a basis for their assumptions is, obviously, nylonase. And, obviously, nylonase is bogus. It is one of the most beautiful examples of a bogus assumption emphasized by darwinists for years as "truth" only because it seemed to give support to them against IDists. It is a very good example of what dogmatism can do to scientific reasoning. Very briefly, the fairy tale of nylonase's frameshift origin begins (and ends) with a paper from Susumu Ohno on PNAS in April 1984. It was a paper exactly of the kind of the more recent ones cited in wikipedia: highly abstract, and unsupported. Ohno has just observed that there was a "possible" ORF in the genome of the plasmid containing the gene for nylonase which "could" have been, in the past, a protein gene for a £never observed" ancestral protein which "could" have given origin of the observed gene for nylonase by a frameshift mutation. It is interesting that Ohno was suggesting that an observed protein (nylonase) "could" have originated from a hypothetical, and never observed, protein by an hypothetical, and statistically almost impossible, mechanism. That could have remained just an interesting but bizarre paper about an interesting, but almost certainly false, theory, if darwinists had not decided to make of it a favoured piece of propaganda, much like they have done with another useful fairy tale, Matzke's theory about the "evolution" of the flagellum. So, the frameshift origin of nylonase became "truth", and frameshift mutations became the tool of darwinian evolution to quickly and efficiently traverse, by statistical magic, that ocean of improbability which, at the same time, darwinists were swearing there was no need to traverse. And why not? They had empirical evidence for that! We had to wait more than twenty years to get rid of those myths, by means of the serious work of those serious researchers who, thanks God, still exist. And please, take notice that the supporting myth that nylonase was a "de novo" protein, emerged in a few decades by the wonderful mechanisms of darwinian evolution, was based on a quick and superficial statement, in the paper about its discovery, that it presented no obvious homology to a bunch of known proteins, and no other apparent function than digesting nylon. Both these statements were, obviously, false. Finally, I must thank Zachriel for having prompted me to review the current literature about nylonase by citing it on your blog. It took a little time, but it was really rewarding. Strangely, the important news about that very important acquisitions had not been boasted around by darwinists, and even we IDists had probably not noticed it. But I must give credit to Wikipedia for correctly citing it, even if with the understandable, but very awkward, attempt at covering the facts by the prompt citation of those other absolutely irrelevant papers. gpuccio
systematic, functional complex, specified information or SFCSI
Do you think that it will really help to introduce yet another term? Currently we have IC (Behe) and CSI (Dembski) which seem well established in the ID community. In addition one occasionally finds FCSI (KairosFocus, Jerry, gpuccio) and FSIC (Gordon Mullings). And now Jerry just introduced SFCSI. Will a six letter abbrevation (e.g. IFSCSI) be the next step in the evolution of CSI? sparc
"Other people have demonstrated that MET is most likely not up to the task of generating all the biological diversity and function which we see (wasn’t that the point of the Altenburg 16). However, that only means that we don’t know everything about how evolution works yet. Big Deal … there’s much to learn about the mechanisms." The MET is an evolving synthesis of processes that account for changes in a population over time. As such it is far different from what it was in the early 1940's when it was initially finalized. It changed dramatically with the discovery of the structure of DNA and all the multitude of processes surrounding micro biology in the next 40 years after Crick and Watson discovered the DNA structure. It is constantly changing today as things such as epigenetics are added. The two constants in all the various syntheses have been natural selection and naturalistic mechanisms for providing variation. Gradualism is even being thrown to the wolves but never natural selection and never naturalistic mechanisms of variation. Changes to the MET are not thought of as supporting ID. Read Hunter's new website. They now say that organism have the capability to evolve built in. Is this support for ID according to the evolutionary biologists? No, because this capability was selected after it developed naturally. Evolutionary biologists have more outs from a sticky situation than Jack Bauer is capable of executing in 24 hours. jerry
JayM (#133): "This does not, as you touch on below, reflect the mechanisms that are part of modern evolutionary theory (MET). The search space for viable organisms is not uniform and the mechanisms are not random (although some do include a random component)." That's the point. Since evolution is not random, what allows it? The answer: a matching of search space to search procedure which is just as improbable as the effects which it produces. ie: if we wouldn't expect the chemical constituents which make up a human brain to randomly coalesce in someone's backyard pool, then we shouldn't expect background noise and an arbitrary set of laws to produce life and evolution. JayM: "If we’re going to argue that MET is an insufficient explanation for what we observe in the natural world, we need to address what evolutionary biologists actually say." But neither myself nor these two papers say that MET (+ all other evolutionary mechanisms yet to be discovered) is insufficient to explain the biology which we see around us. These papers take evolution as granted. These papers merely get to the foundational point of the debate -- "what causes life and evolution" and "is life and evolution possible absent previous foresight" and "will background noise and an arbitrary set of laws cause life and evolution." The rest of my post which you responded to outlines the significance of these questions and how these two papers plus ID Theory help to provide answers. JayM: "I just read the NFLT papers I could find on the web. I’m not sure I’d summarize quite the same way, but I think I get your point." I have also read through them and they stress the significance of matching search procedure to algorithm in order to achieve any significant results. They even conclude (from what I can remember) by discussing the importance of incorporating problem specific information into the behavior of the algorithm in order to solve the problem. JayM: "This seems to be a significant change to ID theory. Instead of attempting to demonstrate that the mechanisms of MET are insufficient, we’re now reduced to arguing that the physical laws that result in those mechanisms are unlikely." Actually, that isn't strictly an ID argument. Poking holes in evolution is not equal to an ID argument. Many non-IDers, such as James Shapiro, point to the insufficiency of MET to explain the diversity of life which we observe. I believe he is on the right track researching cellular non-random genetic engineering in which living organisms control their own evolution to an extent. ID Theory, by definition, is a search for patterns which signify previous intelligence. Once I actually began to understand ID Theory, I noticed that it is not an anti-evolution argument. In fact, these two papers begin to provide evidence that evolution is one of those patterns which signifies previous intelligence. CSI and IC are other patterns which also signify previous intelligence. JayM: "The cosmological argument is interesting, but it cedes all of biology to ID opponents." Not so. Life and evolution itself is evidence for a previous intelligence. Think about it a bit. There would be no cosmological argument if life didn't exist. There would be no patterns to explain and thus no cosmological ID. Biological ID merely recognizes that certain features of life (including the system of life itself and its evolving process) require previous intelligence and Cosmological ID provides evidence that this intelligence is fundamental to the laws of nature. JayM: "MET doesn’t have anything to say about how the mechanisms arose, it merely shows how they explain what we see." That's pretty much the point. So when you see your friendly prof for the public understanding of science foaming at the mouth attempting to say that evolution proves that there is no teleology within life or nature, then you know that he is way out of his element. ID Theory is interested in quantifying the patterns produced by intelligence and what causes life and evolution. JayM: "Please correct me if I’m misunderstanding you, but it seems that you’re saying that Dr. Dembski’s papers provide support for the idea that the universe is designed, but have nothing to say about the likelihood that the mechanisms of MET can account for all the biological diversity we see." They provide evidence that life and evolution are designed. They show that in order for a system (including evolution) to increase the probabilities of discovering a function/pattern or solve a problem, the higher order system which causes that evolution must be at least as improbable. So, as a foundational explanation, we need to look for something which can increase probabilities and foresighted systems are capable of that, however background noise and arbitrary collections of laws (set of laws created without any consideration for future results) are not up to that challenge. Other people have demonstrated that MET is most likely not up to the task of generating all the biological diversity and function which we see (wasn't that the point of the Altenburg 16). However, that only means that we don't know everything about how evolution works yet. Big Deal ... there's much to learn about the mechanisms. CJYman
Mark Frank, Use data as a synonym for the word information. The molecule has a name so you can use that or better yet since we know we are talking about nucleotides, so use A, T, C, G. Thus the information in the DNA or some segment of the DNA is the molecule A, T, C or G or a string of these letters. It is simple as that. For a rock you can use the molecule at some coordinate and then proceed to list every other molecule at each coordinate point. I cannot imagine why anyone would want to do this but it theoretically could be done and when through you would have a very complex set of information with each data point an individual piece of information. Now each triplet of letters in the DNA or certain parts of the DNA is another piece of information. What is so hard about this. If you want to pursue some philosophically obscure path to analyze this example be my guess but the rest of us will proceed on. Now certain segments of this DNA can be thought of as a unit and that unit is another level of information. Think of letter, word and sentence. And these sentences have meaning and that meaning is a string of polymers of amino acids which people call proteins. It is all quite simple and straightforward. So when we say that DNA contains information and complex information it becomes very obvious what is meant but so far it may not be any different than the rock. But the DNA data goes further and relates or specifies something else while the data in the rock has just hit a stone wall and is just still a rock. And we can go further and look at look at the thing specified and we see that it has a function so the nucleotides in DNA is information, complex, specifies something else and this something else is functional. We might even take this to a new level above this and say that the functional element or protein or RNA then becomes part of a coordinated system and maybe we can call the data or information in DNA something like systematic, functional complex, specified information or SFCSI. We have now left the rock in a distant galaxy as we accelerated to light speed to get to where the data is leading us to. And who knows that if we can determine new levels we will have a longer abbreviation that we can put in our Mickey Mouse decoder ring to get the meaning of life. If you want to quibble over this, be my guess but I do not see the point of it. I assume that these papers of Dembski and Marks are looking at the likelihood a sequence of mutations can lead a particular DNA string from one functionally complex specified point to another that is not related to it. If that is not what it is about then maybe someone could enlighten us peons. jerry
Gpuccio I then went on to read the paper you reference in #129. It seemed surprisingly easy to understand. Of course I may have got it wrong, but it appears to conclude that there is a high (92.5%) conservation of proteins among the PDF proteins - those with clearly defined function. This among organisms as diverse as yeast and mammals with a last common ancestor hundreds of millions of years ago. There is much lower conservation among POF proteins. But these are the protein for which the function is not understood. And a lot of them have quite distinctive characteristics - so they cannot be treated as like PDF proteins for which the function has yet to be discovered. Many of these POF proteins might do nothing at all. Those that do something we have no idea whether it is the whole string that is relevant or just a part of the protein that is active or how wide a range of other proteins could perform the same function. If these things are true then this would greatly lighten selective pressure and under Darwinian assumptions they would diversify very rapidly - because a wide range of changes would be viable. Am I failing to understand something? Mark Frank
Gpuccio You wrote: So, although frameshift mutations are totaly unrealistic as a way to obtain functional information, darwinists are trying just the same to exploit them in their “models”, driven probably by sheer desperation. What a pity that the only empirical support for those theories comes from the old model of nylonase, and that that model has been recently shown to be false (if you don’t believe me, just go to the Wikipedia page about nylon eating bacteria, and read carefully the linked paper by Negoro et al, to see how the famous nylonase, for years boasted by darwinists as an example of new function created by a single frameshift mutation, is instead a classical example of microevolution, where the mutation of one or two aminoacids allows a new substrate affinity in an existing esterase). About the fourth paragraph in Wikipedia reads: A series of recent studies by a team led by Seiji Negoro of the University of Hyogo, Japan, suggest that in fact no frameshift mutation was involved in the evolution of the 6-aminohexanoic acid hydrolase.[1] However, many other genes have been discovered which did evolve by gene duplication followed by a frameshift mutation affecting at least part of the gene. A 2006 study found 470 examples in humans alone.[2] with suitable references. What gives? Mark Frank
Mark: excuse me if I have not answered your previous post (I will do that as soon as possible), but as you see the discussion has shifted to other issues. So I answer first you 131. Yes, you are right: great changes take place at molecular level in natural history, and we don't know why ot how. We in ID are certain that those changes are guided, for the reasons you know, but for the rest we are as ignorant as the darwinists. It is perfectly possible that a new protein comes out of an initial gene duplication, as darwinists believe. That's what programmers often work when they want to modify an existing code. And then in some way the new code is superimposed. I don't believe too much in the importance of working with big strings, because that has many restraints. In many cases you would need to change individual nucleotides at the right place, and point mutation remains the most powerful way to do that. I suggest here, just for discussion, some possible mechanisms of guided programming of a new protein. They are purely hypothetical, but I will mention where there could be some corresponding model in biology: 1) Guided variation: a variation which is not random, but pseudorandom, and where special events are intelligently favoured (for instance, at quantum level). That could include both point mutation and other mechanisms, like deletion, inversion, and so on. Guided vairation could also be mediated through intermediate tools, like transposons, ERVs, and others. 2) Random targeted hypermutation: we have an example of that in antibody maturation, but that is realized by an algorithm embedded in the immune system, and is not destined to genetic transmission. However that mechanism works, but it has to be coupled to: 3) Intelligent selection. In other words, the results which are in the sense of the needed change must be kept by a guided, intelligent intervention, even if they are not yet truly functional. For instance, the change in function could be measured, even if minimal, and preserved, as it happens in antibody maturation when the affinity for the original antigen is in some way measured (probably by the antigen presenting cells) and determines the suppression or retention of the new clone. That is similar to what happens in GAs, and it could take place in the context of the individual biological being, and not through the much slower (and imaginary) process of NS. Or, if the designer already knows the solution, he could act as a direct oracle, maintaining only the variations which are in accord with the information to be implemented (a la weasel). In all cases, intelligent selection, as we know from the example of appropriate GAs, can realize very quickly what NS can never achieve. But, while guided variation, targeted random variation and intelligent selection are certainly possible mechanisms by which a designer can implement intelligent information in the genome, huge mysteries do remain. First of all, is the implementation of the information really gradual, like traditional darwinism supposes? In other words, does the designer act in a very slow and gradual way? It is possible, but there are a number of issues against that: first of all, obviously, natural history, which is almost certainly discontinuous, especially at three major points: OOL, and the ediacaran and cambrian explosions. And the whole fossil record, in general, does not favour continuity, as Gould had well understood. And finally, as I have said often, we cannot really think that change can come hrough one new gene, however powerful. Real change in a complex system requires a redefinition of multiple genes and parameters, of the regulation, of the procedures, and of many other things. It is not a case that even very similar species are sometimes very different at the genomic level. Moreover, we still don't know how and where regulations and procedures are really implemented in the genome (or in some other place). The difficulty of obtaining even a single new protein, however, is evidently troubling darwinists ever more. That can explain the recent attempts to find an important role to frameshift mutations, which are the only hope for traditional darwinism to obtain radically new sequences in a "simple" way. So, although frameshift mutations are totaly unrealistic as a way to obtain functional information, darwinists are trying just the same to exploit them in their "models", driven probably by sheer desperation. What a pity that the only empirical support for those theories comes from the old model of nylonase, and that that model has been recently shown to be false (if you don't believe me, just go to the Wikipedia page about nylon eating bacteria, and read carefully the linked paper by Negoro et al, to see how the famous nylonase, for years boasted by darwinists as an example of new function created by a single frameshift mutation, is instead a classical example of microevolution, where the mutation of one or two aminoacids allows a new substrate affinity in an existing esterase). gpuccio
CJYman @121
What is the probability that a given pattern (measured in bits) will be generated by unguided (random) bit operations? That is simply calculated as the pattern’s probability — assuming a uniform search space.
This does not, as you touch on below, reflect the mechanisms that are part of modern evolutionary theory (MET). The search space for viable organisms is not uniform and the mechanisms are not random (although some do include a random component). If we're going to argue that MET is an insufficient explanation for what we observe in the natural world, we need to address what evolutionary biologists actually say.
Now, the point of these papers is to ask [and answer] the question: “how can the probability of generating a pattern be increased *and what is the probability of finding a way to increase the probability of finding that given pattern*?” Little bit of a tongue twister, but read it a couple times and it’ll make sense. Well, basically, there can exist a search procedure (bit flipping operation) which rejects some bit operations/flips and keeps others. However, if the search space is uniform, this does not increase the probability for finding a given pattern. What needs to happen is that the search procedure needs to be matched with the proper search space that will allow the filter to actually improve the probability for finding a given pattern. This has already been proven in the NFLT.
I just read the NFLT papers I could find on the web. I'm not sure I'd summarize quite the same way, but I think I get your point.
Dembski and Mark’s paper picks up from there. What is the probability of matching a search space to a search procedure in order to increase the probability of finding a given pattern? They have merely shown that the probability of finding that match can be no less (and apparently increases exponentially with every higher level search) than the probability of finding the given pattern within a uniform search space in the first place. So appealing to a non-uniform search space and an evolutionary filter does not provide a solution to increasing the probability, since the probability (information) is merely moved back a step to the probability of finding the set of laws and initial conditions which will provide a non-uniform search space and ratcheting filter to increase the probability of finding the given pattern.
This seems to be a significant change to ID theory. Instead of attempting to demonstrate that the mechanisms of MET are insufficient, we're now reduced to arguing that the physical laws that result in those mechanisms are unlikely. The cosmological argument is interesting, but it cedes all of biology to ID opponents.
Thus, according to the math provided by Dembski and Marks, it is just as improbable to find a given pattern (measured against a uniform search space) as it is to find the search procedure (and landscape) to increase the probability of finding that given pattern.
MET doesn't have anything to say about how the mechanisms arose, it merely shows how they explain what we see. Please correct me if I'm misunderstanding you, but it seems that you're saying that Dr. Dembski's papers provide support for the idea that the universe is designed, but have nothing to say about the likelihood that the mechanisms of MET can account for all the biological diversity we see. JJ JayM
gpuccio @129
New molecular functions arise in new species. New proteins, new protein cascades, new regulations. New organs, new tissues, new cells, There is a lot of novelty, wherever you look at, in biology.
Certainly. There is also a lot of commonality. The ratio between those two is one of the pieces of evidence for common descent. Organisms with a more recent common ancestor share more features than those who diverged earlier.
djmullen’s reasonings of “sweet places” and “no searchs” are pure imagination, and bear no relationship to biological reality. And you accept them for the same reason: you don’t know how things are.
With all due respect, my knowledge of biology is more than sufficient to recognize that djmullen's comments are spot on. It's very simple to understand just by looking at yourself and your parents. You do not share a genome with either. In fact, you have a number of mutations so that your genome is not even a proper subset of the combined genomes of your parents. Nonetheless, you live. You are a viable organism. That demonstrates beyond a shadow of a doubt that there is a "sweet spot" in genome space within which the evolutionary mechanisms that operate between one generation and the next work. This is why I have great difficulty seeing how Dr. Dembski's two papers are at all applicable to the ID issue. I could see how they might be applied to the origin of life problem, but evolutionary biology starts from a viable point and explores nearby points, most of which are also viable. Evolutionary mechanisms do not search the whole possible multidimensional genome space.
Again, that’s not your fault, or djmullens’ fault. It is really strange how many people believe that the difference between species are easily explained by some reshuffling of existing information.
I've been enjoying our conversation, but your patronizing tone in this post is unappealing. You seem to be conflating multiple topics. In the context of observed evolutionary mechanisms, Dr. Dembski's papers do not seem to be obviously supportive of ID. The fact that small changes to a viable genome result in other viable genomes, as shown by the differences between you and your parents, indicates that the viable regions in genome space are not uniformly distributed. That means that mathematical discussions of searching the whole space, however elegant they may be, do not apply in this domain. The broader question of just how far these mechanisms can go is, as you point out, quite interesting. The fact that there are limits at one scale does not invalidate djmullen's explanation at the scale he was discussing. JJ JayM
Gpuccio I am thinking about your post #129. I wish there were a biochemist watching this discussion but I have a logical query. If: a lot of proteins arise practically “de novo” at sone time in the course of natural history, even if they can persist after in many species. what is the mechanism that causes that to happen (whether it be guided or not)? It is presumably a rearrangement of DNA in the genome and we know the mechanisms involved: * point mutation - clearly not sufficent to produce a radically new protein * crossing over during meiosis - but you seem to be saying new proteins arise in asexual species as well as sexual It seems that what we are really talking about is: * insertions, deletions, and duplications. These operations could instantly create new proteins that appear radically different from their predecessors. But they would be working with strings of DNA that had proven effective before. The unit being rearranged is much larger than a single base pair or even a single amino acid. I don't have enough biochemistry to know if this makes sense but it seems worth pursuing. Mark Frank
Sal Gal: I can only quote myself (#104): "I am saying only two things: a) The genetic code is a symbolic code, in the sense that the information stored in DNA as protein coding genes can only be retrieved by means of a complex system, the translation system, where exactly the same symbolic correspondences are embedded. And that correspondence is not in any way connected to biochemical laws, but only to a semantic connection between the stored information in DNA and the translating system in tRNAs. b) The information in protein coding genes is functional information, because it is perfectly apt to guide the synthesis of a perfectly functional protein. Please notice that the function is in no way present in the DNA sequence (DNA can never act as an enzyme), but arises only in the final protein, as a consequence of the information of the DNA. The possibility of errors in the process does not change anything of that." To be even more clear, I am saying: 1) UCU is translated to serine in practically all protein coding genes. 2) That happens for no biochemical property which connects the tri-nucleotide UCU to the aminoacid Serine, but only because in all those beings there is another molecule, a specific tRNA, which has an anticodon in a key position which can connect (through biochemical laws) to the UCU codon in mRNA, and that same molecule (the specific tRNA) has another, separated site, in its structure, which links the aminoacid Serine and attachs it to the growing protein sequence. 3) The results of all that is that Serine strangely appears at the right functional place in all existing functional proteins. 4) The same can be said for all the other 19 aminoacids, and for all the other codons in the genetic code (except the stop codons), where each aminoacid is "mounted" by the appropriate tRNA, recognizing the appropriate codon on mRNA. 5) I am curious about that, and would like to know why it is that way. I don't accept that as a law of nature. And I don't accept the explanations of darwinists (provided that they have ever offered one). 6) If you are not curious about that, or if deeper problems prevent you from asking yourself such trivial questions, I really don't know what else to say. In case you want to go on with the discussion, could you please explain, as clearly as possible, what you don't agree with in the above, very simple, points? And what in them is "reification"? Thank you. And again, I certainly admire the complexity and creativity in nature, and often I am really overwhelmed at its depth and beauty. I feel many different emotions and intuitions when I look at nature, both in a garden or in a biology text (although the garden is usually better). But what I really "wonder" at is the darwinist folly, and how many intelligent people can share such a theory and such arguments without any apparent doubt or second thought. That is a real wonderment for me. That I would bet against, in my best Bayesians moments, as absolutely improbable, if I did not see it happen every day. gpuccio
JayM (#120): I am afraid that you are wrong in what you say, and again that derives form little familiarity with biology (which, obviously, is not your fault). The problem is that darwinist must have convinced people, more or less indirectly, of a lot of things that are absolutely false. The point is that what I say does not refer "only" to the problem of OOL. It is perfectly true also for the problem of evolution. Indeed, I could well have made my example with two proteins from different species. Let's be clear: some proteins can be found in meany species, at different levels of natural history, and remain very similar, or chance in a more or less limited range. But a lot of proteins arise practically "de novo" at sone time in the course of natural history, even if they can persist after in many species. Are you suggesting that practically all the essential proteins, and 3d structures of proteins, were already present in the mythical LUCA? Well, that's not the case. Almost any species has proteins which are exclusively found in that species, and have no homology with other known proteins. In a recent paper "What makes species unique? The contribution of proteins with obscure features". available online on Genome Biology, a comparison of "predicted proteomes derived from 10 different sequenced genomes, including budding and fission yeast, worm, fly, mosquito, Arabidopsis, rice, mouse, rat, and human" showed that "7.5% of the proteins with defined features (PDFs) were species specific (17,554 in total)" and "60% of the Proteins with obscure features (POFs) identified in these 10 proteomes (44,236 in total) were species specific". And "Approximately one-quarter of eukaryotic proteins are POFs". The simplest bacteria have approximately 500 proteins. E. Coli has more than 2000. Eukaryotes are much more complex. Do you believe that all the proteins in eukaryotes can be traced to those in simple bacteria, with only minor evolutional "play". No. That's not true. C. elegans, matbe the simplest multicellular being, a small worm made of only 1000 cells, has a proteome of about 20000 proteins, almost the same as humans. Do you believe they are the same? Or that all human proteins are very similar to those in the proteome of C. elegans? New molecular functions arise in new species. New proteins, new protein cascades, new regulations. New organs, new tissues, new cells, There is a lot of novelty, wherever you look at, in biology. djmullen's reasonings of "sweet places" and "no searchs" are pure imagination, and bear no relationship to biological reality. And you accept them for the same reason: you don't know how things are. Again, that's not your fault, or djmullens' fault. It is really strange how many people believe that the difference between species are easily explained by some reshuffling of existing information. Perhaps they read too much Dawkins (certainly not a good idea), and their best model is dog breeding. But sometimes I wonder how darwinists have succeeded in spreading so much disinformation about biology, and in hypnotizing people making them believe things that have no scientific basis. gpuccio
But here you have slipped up and treated the abstract code as physical, and have gone from there to say that if scientists model genes as encoding proteins, then something must have done the encoding. ... In mainstream evolutionary theory, there is no encoder. Merely changing the bases in a chromosome is not encoding. Exchange of DNA among organisms, perhaps mediated by viruses, is not encoding.
Huh?? Your objection just confuses the issues. Unguided variation may not be an "encoder" (which I presume you're saying is inherently an intelligent agent) but changes in information must meet the coding scheme in order to be valid for transcription. Error correction just enforces this requirement of being encoded properly. And I'd like to see how you'd get from a basic chemical replicator to an information-based replicator with rules defining a coding scheme. In any case, this particular discussion already took place not too long ago in A Simple Gene Origination Calculation. Patrick
Leaving aside the points related to OOL and the initial genetic toolkit, this is the main issue:
Once you have a population of imperfect replicators, djmullen is absolutely correct when he notes that there is no need to search a large genome space. ... That’s equivalent to just looking around the (very) local region of genome space for equivalent or better fitness. Over many generations, the genomic makeup of the population may change, even significantly. At no point, however, is there a need to search prohibitively large regions when starting from a viable point.
The unqualified assumption is that for every potential long-range target there is a series of fitness-improving functional (or at least non-deleterious) intermediates in the "(very) local region of genome space". ID proponents think that for most starting "viable points" in genome space that the search landscape could be likened to a valley surrounded by sheer cliffs. The valley consists of other viable points in local genome space, but is limited in scope. Some ID proponents believe these sheer cliffs can be traversed by intelligent mechanisms. Darwinists who think that "many micro-evolutionary events leads to macro-evolution" think there exists very narrow paths burrowing to other valleys, but even these are very difficult to traverse, and we're lacking evidence for their existence. Other Darwinists think there are mechanisms that allow the traversal of the cliffs to produce macro-evolution. And when it comes to long-range targets this leads to standing challenges like the flagellum. Although I'd say that gpuccio at #118 does a better job of describing the problem than I do in those previous discussions. Patrick
CYJman: The problem is that ID can also invalidate the paper if you want it to. We know that evolutionary search spaces in biology are non-uniform so arguing that they couldn't have been 'found' just begs the answer that they were designed - which is exactly the position of many theistic evolutionists. Laminar
Dembski and Marks can do fancy things with probabilities, but in the applications Dembski really cares about, he cannot get the probabilities he needs. I have always said that active information may have engineering applications. Sal Gal
gpuccio, As Patrick points out, a code is an abstraction. All models are abstractions. You are generally very good about avoiding reification. But here you have slipped up and treated the abstract code as physical, and have gone from there to say that if scientists model genes as encoding proteins, then something must have done the encoding. Consider that you are doing with the genetic code what others do with fitness functions.
It could have been encoded through the amazing works of RV + NS, as darwinists believe, but encoded it is just the same.
This goes to the crux of the matter. In mainstream evolutionary theory, there is no encoder. Merely changing the bases in a chromosome is not encoding. Exchange of DNA among organisms, perhaps mediated by viruses, is not encoding. I commented on probability because you seem to regard complex biological structures and processes as improbable. Tell me if you in fact do not. I mentioned "wonderment" because your degree of wonder at complexity may be converted into a subjective (Bayesian) probability, but not an objective (frequentist) probability. No one can assign an objective probability to the proposition that our empirical observations of the universe may be accounted for strictly in terms of matter, energy, and their interactions. And this spells death for improbability-based arguments that some events have causes attributable to something other than interactions of matter and energy. Sal Gal
Patrick #109:
As for the information/code and rocks discussion: a code is an abstraction. How is a rock in a stream an abstraction?
Thank you, that's helpful and simple. My question now becomes: where is the "abstraction" in DNA? I think we're projecting the abstraction onto it. Before we came along, there was just this mechanism -- if one series of base pairs was present, a cell developed in one way, if another base pair, another course of development. Grossly oversimplified, perhaps (probably), but I don't think it's fundamentally off base. Anyway, we looked at it and conceived of the "abstraction" of "instructions" or "code." pubdef
Some people here seem to be asking, "How does this relate to life, evolution, and ID Theory?" Well, first, life is founded upon the processing of a system of signs/units (measurable in binary digits). The organization of these bits form patterns and a certain percentage of those patterns, when processed produce function which may aid in the survival of the system itself. Since the bits are subject to mutations and transfers of many types, they are subject to bit flips and operations. Thus, the patterns are subject to search (as defined in the first para above in my previous comment). This is why mutations can change one functional or non-functional pattern into another functioning or non-functioning pattern. When a functional pattern is generated, it can be said that that pattern has been found by searching through the possible patterns. Obviously evolution by natural selection provides the ratcheting filter whereby some patterns are kept and other are rejected. Now, "How does this relate to ID?" Since it is just as improbable to find the pattern which can be processed into function as it is improbable to find the match between search procedure and search space which will increase the probability of finding the functional pattern, then the probability of finding the set of laws and initial conditions to allow the structure of life and evolution from a uniform search space is just as improbable as finding the results of life and evolution. So, we can have an infinite regress of fortuitous and highly improbable matchings of search space to search procedure (active information) to ultimately increase the probability drastically of finding a system which can apply foresight (the brain), or we can hypothesize other options. The infinite regress seems too similar to "its turtles all the way down," and I am personally unaware of any evidence of a tested pre-universe evolutionary process, and so this provides no real explanation for highly improbable, functional results seen in life. "Evolution-did-it" doesn't cut it anymore. That evolution must now be explained since it is also just as difficult to find within a higher order search space as the human brain would be to find in a uniform search space -- that is, by chance. If we must ultimately explain everything in terms of law and chance, then it seems that the only real option, other than "turtles (evolutionary algorithms) all the way down" would be to simulate evolution generating itself from background noise (chance) and an arbitrary collection of laws (set of laws put together absent any consideration for future results – absent foresight). But is there another option? Of course, ID Theory says that there is. Systems which are capable of foresight (modeling the future and producing targets) can also increase the probability of finding a given pattern. Intelligent systems do this by applying foresight when matching the proper search procedure to search space in order to increase the probability of finding a given pattern. This has been observed and evidence provided by the NFLT shows that a simulation of evolution will not work unless "problem specific information" about the search space and target is incorporated into the behavior of the algorithm. Basically, this provides evidence that evolution can not even be simulated without some knowledge of the characteristics of the target being incorporated into the programming of the behavior of the interaction between search procedure and search space. So if future knowledge of the target is necessary to even simulate evolution, what happens to the hypothesis that evolution needs no foresight? Put all of this together and we see that the ID Hypothesis is valid and scientific since it is based on observation and is the only verified option that is provided by the math within these two papers. Furthermore, it is falsifiable by showing through testing and observation that just law and chance absent foresight can increase the probability of finding a given target. So far, no one has shown that to be the case and the NFLT provides evidence that it may be practically impossible. CJYman
It seems that some people here are not quite understanding what a search is in relation to these papers. The word "search" in this context has no intrinsic teleology associated with it. It merely describes the flipping of bits (bit operations), while describing the search space in terms of bits. This is where the probability of a pattern comes in to play. The rest of this is kinda lengthy, so I'm gonna post the rest in sections. What is the probability that a given pattern (measured in bits) will be generated by unguided (random) bit operations? That is simply calculated as the pattern's probability -- assuming a uniform search space. Now, the point of these papers is to ask [and answer] the question: "how can the probability of generating a pattern be increased *and what is the probability of finding a way to increase the probability of finding that given pattern*?" Little bit of a tongue twister, but read it a couple times and it'll make sense. Well, basically, there can exist a search procedure (bit flipping operation) which rejects some bit operations/flips and keeps others. However, if the search space is uniform, this does not increase the probability for finding a given pattern. What needs to happen is that the search procedure needs to be matched with the proper search space that will allow the filter to actually improve the probability for finding a given pattern. This has already been proven in the NFLT. Dembski and Mark's paper picks up from there. What is the probability of matching a search space to a search procedure in order to increase the probability of finding a given pattern? They have merely shown that the probability of finding that match can be no less (and apparently increases exponentially with every higher level search) than the probability of finding the given pattern within a uniform search space in the first place. So appealing to a non-uniform search space and an evolutionary filter does not provide a solution to increasing the probability, since the probability (information) is merely moved back a step to the probability of finding the set of laws and initial conditions which will provide a non-uniform search space and ratcheting filter to increase the probability of finding the given pattern. Thus, according to the math provided by Dembski and Marks, it is just as improbable to find a given pattern (measured against a uniform search space) as it is to find the search procedure (and landscape) to increase the probability of finding that given pattern. CJYman
gpuccio @118
So I ask you: how did these two proteins (and all the other hundreds of different proteins, with hundreds of different foldings, 3D structures and functions) arise?
I think you're asking a different question than djmullen was answering. ;-) There are two issues being conflated here. The first, which I think is what you're focusing on, is origin of life. That is, of course, a fascinating area of research. The second, which djmullen discussed so eloquently, is the result of evolutionary mechanisms once life, or at least populations of replicating entities, exists. This is where I have trouble seeing the connection between Dr. Dembski's two papers and ID. Once you have a population of imperfect replicators, djmullen is absolutely correct when he notes that there is no need to search a large genome space. Most replicators will have the same genetic makeup of their parent(s). Some will mutate but still maintain far more similarity than difference. That's equivalent to just looking around the (very) local region of genome space for equivalent or better fitness. Over many generations, the genomic makeup of the population may change, even significantly. At no point, however, is there a need to search prohibitively large regions when starting from a viable point. How to get to that point and the size of the connected, viable regions are interesting questions. JJ JayM
jerry[114], My family tree starts with a guy named Sven in the early 17th century. Doesn't get more Swedish than that. I need to visit you; here in San Antonio it hardly ever freezes and I would love to see some winter. And listen to sarcastic comments... Anyway, yes, I did see your attempt to answer yourself. I don't think we can sort this out until we learn how these results are intended to be relevant to ID/biology. The construction in the "search for a search" paper is very technical (I haven't read the other paper). I understand the technicalities but cannot see the connection to biology. Skal! PO Prof_P.Olofsson
djmullen: Do you really mean what you are saying? Let's start with your (obviously imaginary) bacterium with only one protein. First of all, in case you forgot it, we still have to explain how he got that protein (that is, we still have to solve the problem of OOL). You will probably say that it originated from simpler proteins. But the smallest proteins which can fold and exercise an autonomous biochemical function as enzymes are usually 80-100 aminoacids long (one of the smallest reported subunits is 62). So, where are the simpler proteins (functional proteins) from which the actual ones we see today originated? But let's leave that alone, for now. You have your one protein bacterium. Now you state the most incredible thing: that when a bacterium duplicates its DNA, there is no search necessary. That is incredible, not because it is not true, but because it is so trivial that I really can't understand why you have to state it. Who has ever thought to deny that? It is obvious that DNA duplication is not a search! But let's go on. Now, let's compare your one protein bacterium with a (more realistic) 500 protein bacterium. And let's suppose that the one protein bacterium is the progenitor of the other one (being simpler, that's a fair supposition). Now just tell me: whence did all the other 499 proteins come? You say: it's easy, they came by evolution, just substituting an aminoacid at a time, and always remaining in the sweet spot of functional proteins. And I say, are you kidding? You are simply ignoring a very essential fact: the 500 proteins are completely different one from the other. Do you want an example? In the genome of E. coli, a well known bacterium, we can find at least 2018 proteins described, grouped in 525 groups of similar proteins. The biggest group includes 52 proteins, while the smallest groups include only two proteins each, and there are 291 of them. Now, let's take, randomly, two proteins from two of the smallest groups. One is called dinG, is described as a "Probable ATP-dependent helicase", and is 716 aminoacids long. The second is called clpA, is described as "ATP-dependent Clp protease ATP-binding subunit" (it is indeed a subunit of a more complex protein), and is 758 aminoacids long. Now, I have taken these two examples completely at random, from two of the 291 smaller groups (of two similar proteins) in the E. coli genome. I could have taken thousands of different pairs. I have just selected two proteins with approximately the same length, for a more clear discussion. Then I have blasted the two sequences with blastp, on the NCBI site. That software, which is routinely used for research, looks for possible alignments between two (or more) protein sequences. This is the result: Only four possible partial alignments were found. The best was 49 aminoacids long, and presented 15 identities (15/49). The second was 7/8. The third was 12/30, and the fourth 6/11. In other words, even considering all four possible alignments (which, obviously, are not really compatible one with the other), we get only 40 identities out of two sequences which are each more than 700 aminoacids long. So I ask you: how did these two proteins (and all the other hundreds of different proteins, with hundreds of different foldings, 3D structures and functions) arise? How were these two completely different functional sequences found, remaining always in your imaginary "sweet spot"? How can your concept of "no search" apply to that? Just answer that. And, as you seem to be familiar with combinatorial computing, I need not remind you that the combinatorial space of 700 aminoacid sequences is 20^700. gpuccio
djmullen @116 That was an incredibly lucid explanation, thank you. Of course, it supports my personal contention that the "edge of evolution" (where have I heard that phrase before?) is a rich area for ID research, so I'm a friendly audience for you. While I would like to see more peer-reviewed papers supporting ID, I find your and Professor Olofsson's arguments persuasive. After re-reading Dr. Dembski's two papers, I can't honestly portray them as supportive of ID. There is a gap between the math and what we know about biology. I look forward to someone making the necessary connections. JJ JayM
gpuccio: I'll try to express myself better with an example. Let's use an extremely small bacteria, one with only a single protein gene. "Protein gene" refers to a portion of its DNA whose pattern of DNA bases specifes the pattern of amino acids in a protein. (To simplify the discussion, we'll assume that the stretch of protein gene DNA is continguous and codes for a single protein.) Further assume that this particular bacteria has been happily reproducing for some time and that its offspring are viable bacteria, also capable of reproducing, and that it wants to reproduce again. Where is the bacteria going to get the exact sequence of DNA bases it needs in its protein gene that will specify the correct sequence of amino acids to produce the needed protein? Dembski and Marks seem to think the bacteria has to find this sequence via some sort of a search process, which starts with every conceivable DNA sequence that will fit into the gene and rejects all but the single sequence of DNA bases that will produce the correct protein. At least that's what I gather from reading the words in the paper and ignoring the math, which is beyond me. This would be a gargantuan task. If the protein gene was only 400 base-pairs long, there would be 4^400 or 6.6 E+240 different combinations to try. The universe will burn to a cinder long before the bacteria even gets properly started with this task. What the bacteria does instead is much simpler - it merely makes an exact copy of its DNA, including the section that codes for that protein, and hands that copy down to its offspring. No searching is necessary. Since the bacteria has been using its DNA and the protein it encodes for to successfully reproduce, we know that the protein gene is a good one because successful reproduction is our definition of success. No searching whatsoever is necessary. Now let me show you when the bacteria actually does "search" the search space and how it greatly improves the odds of finding a workable protein gene by searching only areas of the search space that are "close" to itself. Let's use a larger bacteria, one with 500 protein genes. That means that 500 stretches of its DNA encode for proteins and we know that all of those proteins "work" because this bacteria is successfully reproducing. Suppose that a single base-pair in one of the 500 protein genes gets mutated during reproduction. That means that the other 499 protein genes are good ones because they didn't mutate and are identical to their parent's. So the bacteria isn't chosing a point in the search space at random, it's choosing a point that is so close to where it started from that 499 out of 500 protein genes are known to be correct. The odds of the new genome being servicable are obviously hugely more likely than if you choose a point in the search space at random, which would be done by mutating every single base pair. Think of evolution this way: three or four billion years ago, a sub-microscopic molecule was randomly put together that was able to reproduce itself before the same forces that assembled it tore it apart. It's offspring was also able to reproduce itself before it was destroyed. This puts the molecule squarely in the tiny section of the humongus search space that is able to successfully reproduce. The molecule doesn't have to search for the sweet spot, it's already in it, in a section that's likely enough so that natural forces could randomly assemble a molecule that's in it in a few hundred million years or less. All the molecule has to do to keep its offspring in the sweet spot is to make an exact duplicate of itself. No searching necessary. When the molecule is mutated, then the new, different offspring is tested to see if it's also in the sweet spot. It's tests itself by trying to stay intact and reproduce. Again, only nearby sections of the search space are tested. If the molecule is 100 atoms long and we only change a single atom, then 99% of the molecule is "known good" and we're only testing the effect of the single changed atom. This is much more likely to be successful than randomly assembling a 100 atom long molecule and seeing if it works. If it doesn't work, the new molecule is destroyed. But if the new pattern is good, then we now have TWO different species of molecules that successfully self-reproduce and evolution is off and running. This is how evolution works. Every viable species is always in the sweet spot. It is never necessary to search for a successful DNA pattern to hand down to the offspring, you merely duplicate your known good pattern. If that pattern mutates, you are only then searching the genome space for a viable DNA pattern and you are only searching that space in the immediate vicinity of your known good pattern. I honestly don't understand why Dembksi and Marks are so concerned with searching large spaces. That has nothing to do with evolution. djmullen
Jerry Re your post #106. You finished with: This discussion has appeared several time in the last few weeks and the anti ID people think they have scored points by asking for a definition of information when the simplest definition from a dictionary will suffice. This makes a point which runs through much of the discussion. Words such as “information”, “symbol”, “code” and “meaning” are bandied about as though their use was obvious and unambiguous. But they are not. Even the link you provide has multiple different simple definitions of information. I wrote a detailed response to this. It is too long for a reasonable comment so I posted it here Mark Frank
Prof_P.Olofsson, Like everyone from Sweden, i can trace my ancestry back to an Ericsson so I tell my kids they are descendants of Leif Ericsson. I am cool, it has only been above freezing once in the last 10 days where I live. You don't like my sarcasm about the anti ID comments? Some are pretty silly. Did you see my comment at #110 where I try to answer my own question. I am just trying to get some understanding of what all the technical stuff is about. Maybe you can help me. jerry
jerry[97], And you're a descendant of vikings if I remember correctly! To answer your question, no, I am not ready to say that. I see your comment [9] now, I missed it before. I agree. I would also like to know how these results are supposed to be relevant to biology. The only logic I can imagine for the "search for a search" articel is that either the darwinian search algorithm is chosen according to the Kantorovich-Wasserstein probability distribution, or else there is support for ID. I could not argue for such a claim if my life depended on it. For starters, what is it supposed to mean that an organism "searches for a search"? Then, why does it have to search according to the aforementioned probability measure? What does that measure even mean intuitively? I have actually offered some constructive criticism about the math in the "search for a search" paper. As for applications, my point is that if somebody claims that an article is "pro-ID" then I think it is fair to ask how, when there are no pro-ID claims in the actual article. The fact that the authors are pro-ID is not enough for me. So I'm left to guess but I can't come up with anything reasonable that I can even argue against. In that sense, yes, I suppose I am not being constructive but I don't have much to be constructive about. Let your viking blood cool down now and be nice to Mark Frank! He's a very nice person. Prof_P.Olofsson
pubdef et al., In the paper "The origin of biological information and the higher taxonomic categories", Stephen C. Meyer wrote:
Dembski (2002) has used the term “complex specified information” (CSI) as a synonym for “specified complexity” to help distinguish functional biological information from mere Shannon information--that is, specified complexity from mere complexity. This review will use this term as well.
see also
Biological specification always refers to function. An organism is a functional system comprising many functional subsystems. In virtue of their function, these systems embody patterns that are objectively given and can be identified independently of the systems that embody them. Hence these systems are specified in the same sense required by the complexity-specification criterion (see sections 1.3 and 2.5). The specification of organisms can be crashed out in any number of ways. Arno Wouters cashes it out globally in terms of the viability of whole organisms. Michael Behe cashes it out in terms of minimal function of biochemical systems.- Wm. Dembski page 148 of NFL
Joseph
Gpuccio - I will leave the definition of code and symbol and concentrate on simulation. I ask that in a good simulation of it, the simulated NS derive from the fitness fucntion intrinsic in the digital environment where the simulation is run (such as the operating system, and any other software or hardware resource present in the environment), and not from an artificial fitness function introduced by the programmer of the simulation. At last I understand what you want from a simulation although I cannot see why you create this constraint. I don't think Tierra is going to satisfy you. It is hard from the available documentation to understand how Tierra decides which individuals to eliminate but I am willing to bet the programmer didn't just rely on Windows or Linux to do the elimination. There is a reference to code written to eliminate individuals based on their lack of reproductive success. I am guessing this means how many they spawn i.e. fecundity not fitness. But anyway it appears to be written for the purpose. But in any case stop and think. Suppose we do create a simulation where individuals reproduce and die wthout any code written to deliberately kill them off. i.e. it is just the hardware and operating system. The hardware and operating system are designed by someone - well a team of people. They didn't design it to eliminate individuals but it does the job and no doubt by inspecting the code it would be possible to deduce the algorithm it (accidentally) implements for eliminating individuals. Now suppose someone writes code deliberately that implements the same algorithm. They have now written a fitness function. What's the difference? Why do you accept it when the code was a by-product of the OS but not when it was written intentionally for that purpose implementing the identical algorithm? Mark Frank
I have a question since I do not understand the basic argument of this thread. Is the question: Can sexual reproduction amongst members of a population with a specific gene pool find any possible combination of elements in that gene pool? And if so what are the limits for the gene pool? Now there are three sources for alleles or other genetic elements in the gene pool, those already currently in the gene pool in at least one of the members of the population of the gene pool, a mutation on a genomic element of a gamete of a member in the gene pool essentially adding a new element to the gene pool and then there is possible recombination of a DNA sequence during sexual reproduction that can also add something new to the gene pool. Not all combinations of these genomic elements exist in the population but it can be theorized that a currently non existing combination may have some reproductive benefits and would be selected if it showed up. And so the issue is how likely such a combination will show up in a future member of the population. Combinations not available because of the current makeup of the gene pool are just not possible unless there were infusions of new genetic material from some place. So is the question about how easy or how long it will take to produce a combination that has better selection possibilities and secondly whether that combination represents anything significantly new in terms of functional capabilities. In other words is a flying squirrel currently possible within the gene pool of the squirrel population? And if so how long or how difficult would it be to get to the genetic combination that allows a squirrel to fly. Or is the combination of changes in the genome to get there just so implausible or even impossible because no combination of genomic elements in the population could ever lead to it so that naturalistic methods have no chance what soever. And Dembski and Marks are trying to show that whatever combination of fortuitous events took place, it would never reach certain places even if it were theoretically possible. And the term "search" has no meaning because nothing is really searching and the term is just used to describe the results of possible naturalistic processes. Or are we just having intellectual fun with this discussion. I know Darwin just viewed the organism as infinitely malleable and the right combination of environment and luck would lead to almost anything you can imagine. I guess I am trying to answer by own question posed at #9. jerry
As for the information/code and rocks discussion: a code is an abstraction. How is a rock in a stream an abstraction? The positioning of the sun and earth does not encode information in itself. The positioning of ink on a paper does, but the atomic properties of the ink do not inherently encode anything related to the information being outputted. There are real properties involved, which may influence the usage of a code, but the information content is not inherent to these real properties. A series of rocks encoding my name on a plane would be an abstraction that does not rely directly on the properties of the rocks, or streams, or whatever. Patrick
If I breed pigeons I decide which ones survive and on what basis. Suppose I breed pigeons on the basis of speed and the result is a pigeon that has a radically different breast bone structure (I don’t design the breast bone structure - in fact I may not even know it exists). Would this not be an impressive demonstration of Darwinian mechanisms in action? But the selection mechanism (speed) is completely designed.
1. Depends on the information content for this "new" bone structure, does it not? But let's assume it's something radical (not minor) and ONLY Darwinian mechanisms are at work and functional intermediates are always 2-3 steps away. That would at least show that Darwinian mechanisms related to variation are at least up to the task for this one object. 2. As you said the long term goal for the selection mechanism is designed. This gets back to the funneling problem for natural conditions described in #76. Can you think of a hypothetical environment that could produce the same effect in pigeons? This brings to mind my comment on natural selection and squirrels. I could artificially breed squirrels and presumably I could eventually produce a flying squirrel. But what if I could not? Patrick
Mark: # 92: "On the other hand they may depend on some kind of causal relationship other than an arbitary agreement - for example dole queues are a symbol of an economic depression. UCU causes the production of Serine. This is because of biochemistry - not some arbitrary agreement. Therefore it falls into the second category of symbol." No, UCU causes the production of Serine because of an agreement between the code in DNA and the code in the translation system. That agreement is not due to biochemistry, although obviosuly the single steps of the process of recognition (like the coupling of codon and anticodon in the tRNA) follow the laws of biochemistry. You say: "But that is only to say that UCU causes Serine in the context of the transalation system - given the presence of the translation system then there is every necessity that UCU leads to Serine." That's true. But there is no reason of necessity that the same codon (UCU) be connected to Serine both in the information embedded in DNA and in the recognizing codon in tRNA. The relationship between UCU and Serine has no biochemical cause, and is purely symbolical. And that relationship is the same in two completely different parts of the cell. That's all I am saying. For further elaboration on this, please ckeck my answer to Sal Gal at # 104. gpuccio
pubdef, Information gets used in different ways. But we tend to use it in common language as data that mean something. And that is how I believe most use it here. So information is just a piece of data and each nucleotide is a data point. Search the internet for definitions of the word. For example, go here http://www.onelook.com/?w=information&ls=a Use for information such things as news; intelligence; words facts; data; learning; lore Each nucleotide in a DNA string is thus, a piece of information. Each molecule in rock is a piece of information. If one wants to go further down the structure, then be my guess. Each substructure down to the quarks will be pieces of information. Now the DNA and rock are also complex. So each is an example of complex information. However, some units of the information in the DNA specify something else, for example a gene specifies a protein and sometimes RNA. And some of these proteins and RNA have functions and some of these proteins and RNA work together as functional units or systems. So the information in the DNA is complex, specified and and the elements specified are functional. Life is the only place this appears in nature. It appears quite frequently with human intelligence. Now it is quite possible that not all the DNA in a genome is specified and functional and may be just junk but a large part is not. Hence DNA is functionally complex specified information or FCSI. This is what the whole debate is about. This discussion has appeared several time in the last few weeks and the anti ID people think they have scored points by asking for a definition of information when the simplest definition from a dictionary will suffice. jerry
Mark: There is probably still much to say, but I will try to be brief. Please refer also to my other answers to others. #89: Obviously a simulation is not the real thing, because digital entities are not biological entities. But what I ask is that the same logical concept is tested in the simulation, which is believed to act in the real thing. As NS in the darwinian theory acts according to the fitness function of the environment, which is written by nobody, I ask that in a good simulation of it, the simulated NS derive from the fitness fucntion intrinsic in the digital environment where the simulation is run (such as the operating system, and any other software or hardware resource present in the environment), and not from an artificial fitness function introduced by the programmer of the simulation. That is not an absurd or unrealistic request, Indeed, as you can see, JayM has understood it very well, and pointed to an example (Tierra) where my requests seem to be satisfied. Therefore, it is possible to satisfy those requests, and I am convinced that such a premise guarantees a much better simulation of the "concept" of NS. I have no pretense that any digiatl simulation can really model the "substance" of biological reality. You say: "What does an “intrinsic” capacity to survive mean in this context? What is the “true” replicating ability as opposed to any other replicating ability?" It's very simple: I mean the natural capacity of the digital replicators to replicate and survive in the digital environment, not because of some judgement or measurement made on them by an artificial fitness function, but simply because their code can better utilize the resources in the digital environment. You say: "It is almost as if you want the environment and the die/survive mechanism to develop through evolution as well as the individuals that live in that environment." No, that was never my point. The environment is alredy set at the beginning, like any digital environment is. It could change, or not, but not "through evolution", because the environment in natural history is not supposed to change "through evolution". Similarly, the die/survive mechanism is initially set by the programmer, through his programming of digital replicators which can survive and replicate in that environment. That is done by the programmer, because we are not simulating OOL. But any successive variations in survival or death of modified replicators would happen as a consequence of what evolution has done, plus the intrinsic rules of the system. And no, I don't think that artificail selection is "an impressive demonstration of Darwinian mechanisms in action". Your example with pigeons, if it were true, would only demonstrate that a different breast bone structure can arise by RV in pigeons, and that an intelligent observer interested in speed can select it because it helps gain speed. But it tells nothing about NS. To show a complete darwinian mechanism, you should show that the different breast bone structure, after arising by RV, is selected and expanded in the population of pigeons through a reproductive advantage, be it due to speed or to any other associated function. Can you see the difference? gpuccio
Sal Gal (#98 and 99): OK, in the genetic code we observe the decoding. But if something is decodedm it must have been encoded. I am not saying that in itself that proves that it was encoded by a designer. It could have been encoded through the amazing works of RV + NS, as darwinists believe, but encoded it is just the same. I don't want to start here an useless discussion about terms. I just think that you, and pubdef, are charging me of what I have never said. I am saying only two things: a) The genetic code is a symbolic code, in the sense that the information stored in DNA as protein coding genes can only be retrieved by means of a compelx system, the translation system, where exactly the same symbolic correspondences are embedded. And that correspondence is not in any way connected to biochemical laws, but only to a semantic connection between the stored information in DNA and the translating system in tRNAs. b) The information in protein coding genes is functional information, because it is perfectly apt to guide the synthesis of a perfectly functional protein. Please notice that the function is in no way present in the DNA sequence (DNA can never act as an enzyme), but arises only in the final protein, as a consequence of the information of the DNA. The possibility of errors in the process does not change anything of that. And I cannot understand to what you are referring when you speak of my "wonderment at the transformation of sequences of bases into proteins" and to some "argument from improbability" about that which should imply the whole universe and its Creator. I am not wondering at the transformation of anything. I am just saying a very simple thing, that the transformation of a sequence of bases into a protein can never happen because of natural biochemical laws. It requires a very complex system of biochemical machines. In the same way, a newspaper cannot come out of mechanical laws of the universe, if you have not the journalists, the printing machine, and so on. I am very susprised that I must discuss with you about those simple things. What has that to do with the universe, the Creator, and arguments from improbability?. If you think that it is an argument from improbability that a newspaper is the product of journalists and of a printing apparatus, well, I am all for arguments from improbability. The real question for the genetic code, and for the information in DNA, is: Why and how is it there? Why and how is it as it is? I don't think that this is a stupid question, or an "attempt to circumvent absurdity", or that trying to answer it strictly depends on our personal experience of the Creator (although I would probably agree with you that, in a sense, everything does). Darwinists have been trying to answer that question for decades, without directly drawing from their personal experience of the Creator, and I certainly don't blame them for that, even if I don't agree with their answers. ID is just trying to answer it in a different, and IMO much better, way. gpuccio
# 97 pubdef I think you charge yourself unfairly. You asked what is it that makes a rock in a stream not a code, while DNA is a code, other than greater complexity. Gpuccio, being a gentleman, attempted to answer it, and this clarified what he thinks a code is. Seems like it was a good question to ask. Mark Frank
pubdef (#97): "In other words, a physical interaction attains the status of “information” when it can give you “a specific useful information.”" I can't see your problem. The physical interaction of this answer to you post, (all the physical events which cause the bits of my answer to appear on your screen) has the status of information because you can red the words, and understand the meaning (although it does not seem to be particularly useful to you, but that's another story). A protein coding gene is information because it can instruct the translation system about the correct aminoacid sequence. If we cannot agree on such simple concepts, I really don't know what else to say. Again, I am not trying to imply "immediately", with that, that the sequence of nucleotides was in some way written by a designer. That is the conclusion of the whole ID theory, but I don't want in any way to imply that in my concept of code or of information. I am just saying that a protein coding gene contains in its organized sequence, not in its general biochemical structure, the information for a functional protein. That remains true even if that information was produced by unguided mechanisms like RV and NS, as darwinists believe. And the information is stored through a symbolic code. That is simply true. But I am afraid I cannot say it more clearly than I alredy have. You say: "As long as it appears to me that ID relies on how complex something is, I don’t think I’ll be impressed. In my limited experience, evolutionary biologists have been quite aware of how complex nature can be." But the main concept in ID is not complexity, but specification. Complexity is everywhere, but specification is characteristic of designed things. The problem is, some things are designed and specified, but they are simple. In that case, the specification we observe could in alternative be the product of a random process, and we cannot infer design with certainty, even if it was really the cause of what we observe. Those are the famous false negatives, which are part of the ID theory. That's where complexity becomes important. Associated to specification, it becomes the rule to infer design with only virtual (logical) possibility of false positives. But it's specification which is the real characteristic product of design, and not complexity. gpuccio
JayM: I find your last post very balanced, and I don't think that you are confused about my points. In a sense, I think you have understood them very well, and that we agree on much. I will try to clarify what seems to be still not clear. 1) I am not so aware of the current status of GAs, so I apologize for not being able to enter into details. I will use your indications to deepen my knowledge, and maybe we can go further in the discussion. 2) I think it should be clear that I have no objection that well programmed GAs can achieve specific answers. In the same way, I have no objection that, if a GA really does not introduce active information, its results can be considered interesting data, and should be carefully evaluated. I remain of the idea, howevere, that the best simulation of the "concept" of NS is the kind I have suggested, and which seem to be implemented in tierra. In a sense, any artificail introduction of a fitness function could potentially introduce active information, and it could be difficult to demonstarte or understand how in each single case. That's why relying only on the spontaneous properties (or, if you want, fitness function) of an independent digital system seems the best solution. 3) I agree with you that Tierra, at least at first scrutiny, does what I have requested. I am happy with that, but obviously I will try to understand more about it, thanks to your new links. Just give me some time. But I can well accept that Tierra can be the kind of simulation which I have in mind. 4) I am in no way "adding additional constraints about new code and CSI". I have no objection to Tierra, if it really works as I understand. I have only said that its results seem to me, at ny current level of understanding, good simulations of microevolutionary events. I am adding no constraints about CSI: I just think that Tierra has not produced new CSI. I will be more explicit. I think that Tierra has produced some optimization of an existing code, within the range of a random search. That is interesting, and it is ceratinly important data, but it is not CSI. But I don't want to anticipate a discussion about that until I can personally understand the data from Tierra. I am interested essentially in two aspects: a) How much new functional complexity has been generated (that is, how many new functional bits). b) If new functional specifications have emerged (that is, for instance, new algorithms, and not only an optimization of the existing algorithm and code). That was the meaning of my expression "new code", but I understand that it probably was not precise enough. 5) I absolutely agree with you that there is extensive potential for ID research in all that. I am convinced that if ID theorists and darwinist theorists (and anybody else interested) could work together in mutual respect, instead of fighting, and compare their views in the field of active research and theoretical confrontation, much good would ensue for science. gpuccio
gpuccio (and Bill Dembski), Your wonderment at the transformation of sequences of bases into proteins is not an indication of its improbability. You, an entity within the universe in which you and I agree that the process exists, cannot transport the two of us without and demonstrate to me the objective probability of a universe in which the process exists. Arguments from improbability are absurd attempts to circumvent absurdity. (What is the CSI of the semiotic agent in terms of which CSI is defined?) My knowledge of the Creator is experiential, not logical. I can tell other people steps that might lead them to similar experience, but I cannot speak sensibly of the experience itself.
For since in the wisdom of God the world through its wisdom did not know him, God was pleased through the foolishness of what was preached to save those who believe.1 Corinthians 1:21
This verse is not popular with preachers. It never grabbed me until I saw it in an apologetics presentation of Bob Marks. The title of the slide is "Faith before Science." Bob and I have different beliefs, but I greatly admire people who make such statements outright. Sal Gal
gpuccio (and pubdef): In a coding system, as defined by humans, there are both encoder and decoder. We observe cells doing work to frame sequences of discrete objects and transform them into sequences of discrete objects of a fundamentally different kind. This looks like decoding, but with no empirical observation of an encoder, we are not justified in calling it that. Humans recognized the analogy to decoding, and began to perform encoding operations. The fact that there exist genetically engineered organisms decoding human transmissions implies neither that organisms typically decode nor that genetic information is generally due to human-like activity. In fact, if we adopt the perspective that imperfect replication of DNA is erroneous, then algorithmic information usually enters the genome by error. Sal Gal
Jerry -- guilty as charged. It must have been past my bedtime when I posted that last comment. But in all seriousness -- it seems that a large part of the argument of ID relies on the concept of "information," and I'm honestly at sea about what exactly that it. And gpuccio: this is not very helpful.
You ask:
“Where along the continuum of complexity does a physical interaction attain the status of “information?”
The answer is very simple: when some particular complexity assumes a configuration which can give you a specific useful information, which can allow you to do something which otherwise you could never do.
In other words, a physical interaction attains the status of "information" when it can give you "a specific useful information." As long as it appears to me that ID relies on how complex something is, I don't think I'll be impressed. In my limited experience, evolutionary biologists have been quite aware of how complex nature can be. I didn't get very far into "The Blind Watchmaker" before it was overdue, but I did get through the chapter in which Dawkins describes echolocation in bats in greater and greater detail, all the while making analogies to radar developed for planes in WWII. pubdef
Prof_P.Olofsson, Are you ready to say the correspondence of the nucleotides in DNA to the production of functional proteins does not act like a code. If you are, then I will gladly include you with pubdef and Mark Frank. But you should also look at my comment #9 about the current thread. I do detect a small lack of constructive criticism on your part. Yours is one of pointing out the flaws in other's arguments, which is well and good but I do not see any attempt at helping others how to solve the problem other than your criticism. For example, can the problems you raised in the past about the flagellum be solved or ameliorated somewhat if there were probability estimates of the number of potentially functional proteins from the totality of possible proteins. Now I do not know enough about the technicalities of either probability theory or the behavior or random polymers of amino acids to make any intelligent assessment but I bet that there are some that do. By the way I have hardly read all your comments so my assessment could be quite wrong. I was only using the sampling of the ones I have read to make my judgment and like any statistical analysis there is a potential error. The sampling also indicate a cordial and generally nice person. By the way I was a mathematics major in college and went to Duke graduate school on a fellowship for mathematics before leaving for the military and a change of life. My leaving had nothing to do with my work in math but I never really got back into it and eventually went to Stanford for an MBA where I initially was enamored with Operations Research because of my math background. I decided to pursue marketing instead. I had a number of statistics courses in later years as part of a different Ph.D program. So while the technicalities of statistics is a distant memory other things I remember quite well. I can recognize constructive behavior or the lack of it when I see it. That is something I never lost because I see it or the lack of it every day. I look forward to you proving me wrong because I believe you could be extremely helpful. But if you were, you would probably suffer for it with your colleagues. Actually what could be a better way to discredit ID then to actively help it and then have those you helped admit that what they were trying to show was a dead end and then have them thank you for all that you tried to do for them. If such a thing happened to me I would glow with satisfaction. jerry
WilliamDembski[87], Indeed. I also pointed this out in [54]. Minor point: On page 2, there is no need to restrict yourself to finite-dimensional vectors; Tychonoff's Theorem takes care of the infinite-dimensional case which is relevant if you don't want to fix the number of steps in advance. I still wonder about the relevance to evolutionary biology of searching for a search according to the probability meausure induced by the Kantorovich-Wasserstein metric. I'm not trying to annoy you; I think it's a very relevant question. Maybe jerry [93] can give me an answer? Prof_P.Olofsson
gpuccio @86
The problem with many fitness functions in GAs is, IMO, that while they pretend to model NS, they are really introducing active information, and therefore realizing IS.
Do you have any particular fitness functions in mind, other than those like Dawkin's Weasel that do explicitly encode their desired end state? One of my favorite GA examples is the evolved antenna. This is a much better example of what GAs can do than the Weasel. In this model, the environment consists of the physical laws regarding antennas and the other antenna designs. Each genome is rated against the others in that environment. The only thing that the simulation itself does is model random mutation followed by selection where the chance of reproducing is proportional to the fitness in the environment at that time. The only information introduced in this model is the relative fitness of the genome. Although greatly simplified, this is exactly how information from a real world environment is communicated to a population of organisms. I don't see how this requires intelligent selection.
If a GA does not do that, we could analyze its results, and see how much it is really modeling something. But we need to know the code in the details for that.
I believe the paper linked from the page I linked to gives enough information. Would you agree, though, that if the GA behaved as I described, it really is modeling RM+NS?
I have found the link about tierra very interesting. Unfortunately, there is probably not enough detail there to really understand, but my impression is that it could be something more near to what I think should be done. The limit here seems to be that the original organisms were 80 bytes long, if the article is right, and the derived organisms were either not very different from them (79 bytes), or much shorter (45). The shorter ones were essentially a subset of the original ones, and parasitic of them, because they had to use part of the originals’code to replicate.
My apologies, the page I linked to wasn't the best. Here is a much better one, that includes links to many publications on Tierra and alternative implementations. The "organisms" are actually quite different from each other.
In other words, my impression is that in general all the viruses were using approximately the same code, with minor variations. That is very interesting, but it would be essentially a simulation of microevolution, in the range of random probabilistic resources applied to an existing code. Obviously, we should know in detail the individual codes to really understand what happened, and why the modified code was sometimes more efficient than the original one. But I don’t think that any really new code has been generated here, and certainly not with the characteristics of CSI.
I'm a little confused by this. In your original post to which I replied, you said:
No. The replicator must survive or not survive for its intrinsic capacity to survive or not survive in the environment. In other words, the variation must increase the true replicating ability of the replicator: it’s not the designer that has to decide who survives according to a predetermined, and searched for, function.
This is exactly what Tierra does. I'm not sure why you're now adding additional constraints about new code and CSI. What is your objection? In any case, the code of the digital organisms is available, and you can even run Tierra on your own machine to see the behavior you are looking for.
I am sure that we can simulate microevolutionary events in a real simulation. That kind of events is possible, and well documented even in biology (see antibiotic resistance). It is the assumption that a cumulation of simple microevolutionary events can bring about new complex functions (let’s say, a completely new code of at least 500 bits, a new algorithm, and so on), and not just reshuffle or slightly modify existing ones, which should be tested in a true simulation. But if you have further details about that, please let me know.
This is where I think there is extensive potential for ID research. That being said, I believe there are GAs out there that do mathematical theorem proving and have come up with new algorithms. I'm afraid I'm getting more, rather than less, confused about your precise objections to GAs.
You ask for further elaboration on this concept “You understand that it is a very severe restraint to what functions can be selected.” What I mean is that if NS can only expand new replicators which have a detectable reproductive advantage, then not all useful functions can apply, indeed only a tiny subset of them. Many complex functions, while potentially interesting and useful in the long term, would never give an immediate reproductive advantage, and could never be selected by NS.
Ah, I see, thank you. That is definitely a restriction. Much as I might like retractable wheels built-in to my feet, that is not something that can evolve from our current body plan. This ties in with my previous comment about rich areas for ID research. Where can we get to from where we are? JJ JayM
Mark Frank, I don't want to leave you out. So thank you too for your inane arguments. The more we have people like you and pubdef, the easier it is for us. The two of you are setting standards for what can be used against ID and we appreciate your efforts. jerry
Pubdef, Your arguments are fatuous. They seem more like an attempts of a disruptive toddler rather than constructive adult. Comparing a rock in the middle of the stream to an ordered set of molecules which then set in motion a series of steps that end up with a completely different ordered set of molecules that operate in context with several other ordered sets of molecules and each of these ordered sets of molecules have physical properties which are functionally useful for an organism is one of the more inane arguments I have ever heard. Keep up the good work because it is objections to ID such as yours that makes our case easier. Remember we are not trying to convince you because we long ago knew such attempts were useless but we are trying to convince those who are seriously trying to understand the issues. So thank you for your efforts. People like you make our job easier. jerry
Gpuccio Re #90. This all depends on what you mean by "code". Code, meaning and symbol are all words with a number of uses in English. An important distinction is between what Grice call non-natural and natural meaning. They can be associated with an agreement among people to associate one thing with another, for example the letters Au with the metal gold. On the other hand they may depend on some kind of causal relationship other than an arbitary agreement - for example dole queues are a symbol of an economic depression. UCU causes the production of Serine. This is because of biochemistry - not some arbitrary agreement. Therefore it falls into the second category of symbol. You write: The important point is that there is no biochemical reason (law of necessity) why, say, UCU corresponds to Serine. The connection is purely symbolic, and is guaranteed only by the fact that the translation system recognizes the UCU codon and connects it to the aminoacid Serine. But that is only to say that UCU causes Serine in the context of the transalation system - given the presence of the translation system then there is every necessity that UCU leads to Serine. Drinking alcohol causes road accidents - but only in the context of driving a car. Mark Frank
pubdef: Why do you say that DNA is not a code, and that the concept of a genetic code is only an analogy? THat is simply not true. DNA, in its protein coding parts (which, as you probbaly know, are only 1.5% in the human genome)stores information through a specific symbolic code, which works in the same way as the Morse code you cite. I am not implying here thatv the word "code" indictes necessarily that ir is designed: as you say, that would be assuming the conclusion. I am using the word "code" in a very elementary sense (the same as geneticists have always ised it): a symbolic language which bears symbolic information for something else. The genetic code is made of codons (three consecutive nucleotides). Of all the possible 64 combinations, each has a specific meaning: it corresponds to one of the 20 aminoacids, or to a stop signal. The important point is that there is no biochemical reason (law of necessity) why, say, UCU corresponds to Serine. The connection is purely symbolic, and is guaranteed only by the fact that the translation system recognizes the UCU codon and connects it to the aminoacid Serine. The connection is realized by the tRNA molecules, which recognizes the codon in the mRNA, and (at another site of the molecule) links to the right aminoacid and transfers it to the growing protein sequence. So, the connection is purely symbolic: nobody knows why UCU, in particular, corresponds to Serine. UCU is just a word which represents a meaning, the aminoacid Serine. The system works only because all of its parts use the same code or language. And beyond the genetic code, the gene coding sequence is information. The gene coding for the protein myoglobin has a sequence of 154 x 3 = 462 nucleotides which code for the 154 aminoacid sequence of the protein myoglobin. That sequence is unique to that protein, and is the sequence which allows the function of myoglobin. That's what I mean when I say that there is no law of necessity which can output the sequence of myoglobin (or any other functional protein sequence). If you just try to synthesize a random protein sequence from a pool of random aminoacids, you can get any possible sequence. Biochemical laws do not privilege any specific sequence, and least of all a functional sequence like myoglobin. If you want to synthesize myoglobin, you need to know the prinary sequence of myoglobin: you have to know that specific sequence of 154 aminoacids. In other words, you need a specific information. You ask: "Where along the continuum of complexity does a physical interaction attain the status of “information?” The answer is very simple: when some particular complexity assumes a configuration which can give you a specific useful information, which can allow you to do something which otherwise you could never do. gpuccio
Gpuccio No. The replicator must survive or not survive for its intrinsic capacity to survive or not survive in the environment. In other words, the variation must increase the true replicating ability of the replicator: it’s not the designer that has to decide who survives according to a predetermined, and searched for, function. I am really struggling with this. What are you asking for? This is a simulation not the real thing. No simulated life form is going to die or survive unless there is a mechanism in the software for doing that. The programmer must create that mechanism. What does an “intrinsic” capacity to survive mean in this context? What is the “true” replicating ability as opposed to any other replicating ability? It is almost as if you want the environment and the die/survive mechanism to develop through evolution as well as the individuals that live in that environment. Go back to the example of artificial selection. In this case the real world fitness function is the product of a designer. If I breed pigeons I decide which ones survive and on what basis. Suppose I breed pigeons on the basis of speed and the result is a pigeon that has a radically different breast bone structure (I don't design the breast bone structure - in fact I may not even know it exists). Would this not be an impressive demonstration of Darwinian mechanisms in action? But the selection mechanism (speed) is completely designed. Mark Frank
gpuccio:
As Upright BiPed has already said, DNA is a support for information: you cannot have pure software, you always have software written on a hardware support. The “physical configuration of molecules that interacts with other physical objects/particles” is only the biochemical structure of the DNA molecule. But the special sequence of nucleotides which, in a symbolic code, encodes the sequence of aminoacids in a specific protein is pure digital information.
I have to admit that you lost me after this point; I don't really understand what "necessity" is in this context, and don't really care at this moment. But I think there's a problem with the portion of your post that I reproduced here. The "special sequence of nucleotides" -- the nucleotides are physical objects, interacting with other physical objects. You, apparently, are asserting that their sequence is a "code." I maintain that the only difference from a rock in a stream is a matter of degree; they are both physical objects interacting with other physical objects. DNA is much more complicated, but how does that constitute "information" in a way that makes it fundamentally different from the rock? Where along the continuum of complexity does a physical interaction attain the status of "information?" I know that geneticists and others in science refer to DNA as a "code," but I see that as a nomenclature that describes its function by analogy. To argue that the genetic "code" is evidence of ID is to assume the conclusion, i.e., that DNA is a product of intelligence -- a "code" like Morse code or computer source code -- when there is no empirical evidence of teleological origin. pubdef
I should note that our approach subsumes fitness functions but is considerably more general. Fitness functions alter the probabilities in a search. Our measure of active information focuses on that change in probabilities. But there are other ways to alter probabilities than by introducing a fitness function. William Dembski
JayM: thank you for the interesting reflections. A few thoughts: The problem with many fitness functions in GAs is, IMO, that while they pretend to model NS, they are really introducing active information, and therefore realizing IS. If a GA does not do that, we could analyze its results, and see how much it is really modeling something. But we need to know the code in the details for that. I have found the link about tierra very interesting. Unfortunately, there is probably not enough detail there to really understand, but my impression is that it could be something more near to what I think should be done. The limit here seems to be that the original organisms were 80 bytes long, if the article is right, and the derived organisms were either not very different from them (79 bytes), or much shorter (45). The shorter ones were essentially a subset of the original ones, and parasitic of them, because they had to use part of the originals'code to replicate. In other words, my impression is that in general all the viruses were using approximately the same code, with minor variations. That is very interesting, but it would be essentially a simulation of microevolution, in the range of random probabilistic resources applied to an existing code. Obviously, we should know in detail the individual codes to really understand what happened, and why the modified code was sometimes more efficient than the original one. But I don't think that any really new code has been generated here, and certainly not with the characteristics of CSI. I am sure that we can simulate microevolutionary events in a real simulation. That kind of events is possible, and well documented even in biology (see antibiotic resistance). It is the assumption that a cumulation of simple microevolutionary events can bring about new complex functions (let's say, a completely new code of at least 500 bits, a new algorithm, and so on), and not just reshuffle or slightly modify existing ones, which should be tested in a true simulation. But if you have further details about that, please let me know. You ask for further elaboration on this concept "You understand that it is a very severe restraint to what functions can be selected." What I mean is that if NS can only expand new replicators which have a detectable reproductive advantage, then not all useful functions can apply, indeed only a tiny subset of them. Many complex functions, while potentially interesting and useful in the long term, would never give an immediate reproductive advantage, and could never be selected by NS. Moreover, a new function must be well integrated into the existing system of replication, before it can translate into a true advantage. Going back to the biological world, you must understand that protein functions, even when searched by protein engineering algorithms, appear in the beginning as very low biochemical affinities, detectable by some sensitive measurement system, but completely useless in the real cell environment. They have to be intelligently selected and amplified, before a true powerful biochemical function can be reached. And still that function would have to be integrated into what already exists, and carefully regulated (the synthesis of the protein started at the right moment, and stopped when it is no more necessary, the protein concentration regulated at the right level, and so on). Think, for instance, of protein cascades, where all the components of the cascade must be present for the final result to be obtained, and each protein has to be present in different concentrations, from very low to very high, so that the cascade may amplify the original signal. Amd the signal must come from the right source, and be translated to the right effector. And still the effect must be strong enough to give a reproductive advantage, before it can be selected. So, what I mean here is that NS, as it is conceived in the real biological world, and at the molecular level, is a very, very poor oracle. It can do very little, probably almost nothing at the level of complexity which is already present even in the simplest autonomous living beings, bacteria and archea. Because the more an organism is complex, the more difficult it is to obtain an immediate reproductive advantage by a simple step variation. And bacteria and archea are very complex. So, to assume that NS is responsible for the emergence of all the existing proteomes, where each protein is hundreds of aminoacids long, and most proteins are deeply different one from another, and have different functions, and almost none of those functions can be useful by itself, and all that scenario has to be regulated and integrated, not only within a single cell, but among myriads of different cells, in multicellular organisms, and so on, well, to believe that is real folly. So, please remember that ID has never affirmed that RV cannot generate something useful: the ID assumption is that RV cannot generate something useful and complex. That's why the concept of CSI has two parts: the specification and the complexity. It is the complexity which avoids the false positives due to random variation. But it is the specification which connects the complexity to design. gpuccio
gpuccio[84] Evolution is not concerned with CSI as it is defined as a process that has no specific goal. The term belongs to the ID point of view, not Darwin's. My model is what you asked for, which is a darwinian mechanism. From the ID point of view, it generates CSI, because the observer rejects any information that does not lead to his specific requirements which may be quite complex. From the point of view of the box, there is no predefined specific output, but the viewer sees what he requested, CSI. Voting need not be done by human viewers, it could be another black box looking for a mate. Toronto
gpuccio @77 I've been following this discussion with considerable interest. While my personal interests when writing software for my own amusement lean more toward cellular automata, I have implemented a couple of genetic algorithms.
No. The replicator must survive or not survive for its intrinsic capacity to survive or not survive in the environment. In other words, the variation must increase the true replicating ability of the replicator: it’s not the designer that has to decide who survives according to a predetermined, and searched for, function.
First, this has sparked a number of ideas that I probably lack the time to implement fully. Thank you for the mental kickstart. However, I think you might be focusing on the wrong level of abstraction. The fitness function in a genetic algorithm (GA) that is simulating random mutation plus natural selection (RM+NS) is a simplified model of the ability of an individual organism to survive and reproduce in the simulated environment. That is the replicator's "intrinsic capacity to survive or not survive in the environment." That being said, if you want to see a GA where ability to replicate is measured directly, see Thomas Ray's Tierra. It does exactly that.
That’s exactly what is assumed in darwinian theory: the new function must be such that it increases the reproductive ability of the new form.
Yes, and this is exactly what is modeled by many GAs. It is possible to learn about the capabilities and limitations of RM+NS without taking the simulation down to the level of individual atoms.
You understand that it is a very severe restraint to what functions can be selected.
Could you expound on this? I don't see the severe restraint (aside from the limitations of current computing hardware).
So you cannot underestimate this point. Solving the travelling salesman problem in a shorter way does not usually help a software to replicate. It has to be a function inherent to the replication in that environment. Only that kind of function can be selected by NS.
Only that kind of function is selected in the real world of competing organisms. The point of (some) GAs is to show that RM+NS can generate complex, unexpected, and varied solutions to surviving in particular environments. That's a different level of abstraction than simple replication, but the mechanisms being used are the same.
Other kinds of selection are IS.
Again, GAs measure the capabilities and limitations of random mutation plus selection (usually without a known end goal) in various environments. Those environments generally don't mirror the real world, although Tierra reflect a small subset of it. That is immaterial to the discussion, however, because GAs show that the process of random mutation followed by selection does have certain capabilities in a wide variety of environments. My personal view is that GAs could be used to flesh out Dr. Behe's ideas in The Edge of Evolution to provide a better understanding of where the edge lies and what types of problems cannot be solved by RM+NS. I'm a math and software geek, though, so I would think that. I do believe we need significantly more computing horsepower for such research. JJ JayM
Toronto: I think I have answered you #65 in my #74. I cannot see any CSI in your example. I cannot see any function or complexity in the output you describe. And the users are not an environment there, but just conscious observers who project their representations. It has nothing to do with what we are discussing here, but if you want to think differently, you are welcome. gpuccio
Patrick[82] My proposal does not require the properties of a fractal generator. The bit stream could be generated as the result of a function applying a bit mask to n bits of PI starting from location X. The goal is still hidden from the programmer and it will change based on the whims of the environment, the end-user. I too would like to see the output, but I would use it to generate short sentences. If for every million chars of output I got 3 or 4 short runs of proper grammar, say 5 to 7 words each, I would place them in an array. After a few thousand runs, I would have thousands of arrays which I could select again via my black box by simply deciding to interpret the output in a different way. An automated author! Toronto
Toronto, Fractals have been discussed on UD by gpuccio and myself before. Still, I wouldn't mind seeing your particular proposal carried out, if only to see what it results in from an artistic perspective. Patrick
gpuccio[74]
So, as you can see, I insist that to simulate the darwinian mechanism, you have to demonstrate that the results of random variation can generate a ture spontaneous advantage in some replicators, and that such spontaneous advantages can cumulate to the point of generating true new complex functions (CSI).
I think my example at [65] satisfies the above. As a "black box", it generates new CSI according to the environment. The environment supplies the fitness functions in the form of people voting. These fitness functions are constantly changing according to the moods and needs of the voters. The output bit stream need not be used strictly as video or audio but left up to the environment, e.g., a group of unknown users could use the output as indexes into a list of possible stock picks for trading purposes. The "black box" never changes, only the CSI as perceived by the user. As far as the use of a fractal generator is concerned, any process could be used to generate the bit stream. The user is as blind to the internal process as the programmer was to the external goal. Only the single bit change in the "DNA" is required after a successful survival. Toronto
I think that in general I agree with what you say, but I really am not comfortable with the whole concept of fitness function. It looks extremely artificial to me.
I'd say it IS artificial, since fitness functions in GAs are typically not dynamically morphing but are instead statically and uniformly defined at the outset. Although it's possible for the programmer to step in and tweak the function once a plateau is reached. But my point was to try and limit the usage of fitness functions to be more realistic. The problem is that it makes writing the program much more difficult since a large variety of fitness functions will need to be accounted for. Essentially, to make the project feasible you'd need the software to dynamically adjust the fitness functions realistically at runtime without constant programmer intervention. Now I have heard of this being done but only in reference to AI research. In the late 90s a friend of mine wrote a basic AI that then self-modified via a dynamically adjusted GA and other forms of input. Supposedly the resulting AI was pretty smart even on the limited hardware of that time. Unfortunately, the project got axed when the investor died in the World Trade Center. And I'd also agree that I've never heard of a simulation that models all the biological constraints you mention. Patrick
Patrick: I think that in general I agree with what you say, but I really am not comfortable with the whole concept of fitness function. It looks extremely artificial to me. Moreover, I am always thinking in terms of molecular biology. Too many discourses remain generic and useless because they are not brought to the essential level of molecular biology. That's also what Behe tries to remind always. What I wonder is: how can a specific change in a protein bring to a reproductive advantage in all general cases? Obviously, we have very extreme (and artificial) cases where a small change can bring a great survival advantage: antibiotic resistance (of the simpler form) is a good example, and it's not a case that it is almost the only example. And, as Behe correctly points out, it is a case of lucky disease, of loss of information which becomes useful due to exceptional circumstances. But in general, how can a simple mutation in one protein generate a phenotipic change that is good enough to give a true reproductve advantage and to expand? Again, I mean beyond the few cases of microevolution, which always deal with very small changes and adaptations in the same island of functionality? The transition from one protein to a different one is a virtual impossibility in almost all cases. There are no functional intermediates, when you have to change completely the primary sequence, and the folding, and the active site. And beyond that, in almost all cases, how could the appearance of a new protein with some elementary biochemical activity be useful in an integrated and complex cellular system? We know very well that you don't need just a new protein, but a correct regulation of its transcritpion, translation and post-translational events, and a series of finely tuned interactions of that protein with a lot of other proteins and cascades in the cell, before you get functionality. In other words, most cellular functions are IC from the beginning. That's why I have always found very strange the focus on the flagellum, just because Behe used that model in his book. Almost everything is IC in a cell! I would like to repeat here that the focus of darwinists on duplicated genes as the basis for evolution is rather symptomatic. We should remember that a duplicated gene is the only way to work at developing a new functional protein without losing the original function. Indeed, that's what any programmer does when he wants to work at some part of the code and change it: he copies it, and works on the copy, not to destroy the original. But, apart from the lucky circumstances of having the right genes in copy for our evolutionary experiments, we should remember that applying variation on a non functional copy of a gene has an unpleasant consequence which is often underestimated: that negative selection cannot any more control what is happening. In other words, if the duplicated gene is no more transcribed and functional, no negative selection can eliminate the bad mutations which compromise the original function. So, all the original useful information will be quickly lost, and unless and until a new functional configuration arises, no positive selection can apply. In other words, mutations on a non functional gene become neutral, and we are in the full ocean of non functional possibilities, from which probably nothing has ever emerged. gpuccio
I think I'll attempt to summarize in one sentence: The fitness function must not contain a long term goal that is not applicable to short term goals, and these goals are very general in scope applying only to competitiveness against other replicators and not "functionality for function's sake". Patrick
Mark: "Would this satisfy you?" No. The replicator must survive or not survive for its intrinsic capacity to survive or not survive in the environment. In other words, the variation must increase the true replicating ability of the replicator: it's not the designer that has to decide who survives according to a predetermined, and searched for, function. That's exactly what is assumed in darwinian theory: the new function must be such that it increases the reproductive ability of the new form. You understand that it is a very severe restraint to what functions can be selected. So you cannot underestimate this point. Solving the travelling salesman problem in a shorter way does not usually help a software to replicate. It has to be a function inherent to the replication in that environment. Only that kind of function can be selected by NS. Other kinds of selection are IS. That is the weakest point in the concept of NS, and you cannot get rid of it so simply. Moreover, you are underestimating the importance of the metrics. A measurement is a very sensitive metrics, where you can put the threshold as low as you want or can. In NS, that is completely different. the threshold of measurement of a function is necessarily very high, becasue that function must provide some real reproductive advantage. Only a few functions can do that, and only at really functional levels of activity. That is a big, big problem for NS. Finally, you are underestimating the difference between the two kinds of selection: negative and positive. Negative selection has the role of eliminating failures (by far the most likely results, with RV). But positive selection must expand the mutated individual so that it comes to represent the whole population, or most of it. For a mutated individual to expand, it has to acquire a true reproductive advantage vs the previous form, because the previous form is still perfectly functional, and the single mutated individual must compete with all the others so that they are suppressed, and it expands. That is no simple result to be obtained. It is not a case that the best examples of "positive" selection are those of antibiotic reistance, where a single and very powerful artificial aggressive agent can suppress the normal population and let only the lucky carriers of a mutation survive. But that is not certainly the general scenario for all supposed cases of NS. So, as you can see, I insist that to simulate the darwinian mechanism, you have to demonstrate that the results of random variation can generate a ture spontaneous advantage in some replicators, and that such spontaneous advantages can cumulate to the point of generating true new complex functions (CSI). The necessity of overcoming the barrier of better survival is the biggest obstacle for darwinian mechanism. Such a barrier, in a complex replicator, can be overcome only by complex functions and adjustments, which cannot derive from simple random variations. And, if the darwinian model is true, why shouldn't it work in a digital environment? There are infinite ways in which a replicating software can profit of the digital environment where it runs: better programs can occupy better spaces of the digital environment, compete for the computing resources, or directly attack competitors. Computer viruses do that. But I suppose that they are usually designed. Why cannot "better" computer viruses, or anything like that, come out of random digital noise in a digital environment? Why cannot we simulate that? The truth is that we all know that such a simulation would never work. Because the model it is based on has never worked, and never will work. In the same way that plasmodium falciparum has never acquired the capacity to survive at certain temperatures, or to survive in carriers of sickle cell anemia. You just don't acquire that kind of functions that way. By the way, have you an example of GA solving the travelling salesman problem in a shorter way (and, possibly, with a new and different algorithm, which would be similar to generating a new protein)? I would be interested in that. gpuccio
1. The active information in a fitness function acts as a long term "funnel". This funnel can contain an explicit target, like with the weasel example. As gpuccio said, "It is obvious that GAs, being intelligently designed problems, can find solutions which the designer did not know in advance." This is because the funnel contains a specific target that is general in scope and is applied over the long term regardless of short term considerations. For example, you could be looking for the best shape for an antenna and these could be several different fitness functions. I'm just guessing but I'm presuming the landscape for an antenna GA should be very smooth in the best case scenarios. This example starts with an antenna that gets 1 dBi. a) tested against explicit set of shapes b) specific test by sampling signals and any increase, however small, in forward gain is rewarded c) the more complex the shape the better but everything is rejected except those shapes that have a gain of 1 dBi higher than the previous generation (as in, the steps in the pathway are in 1 dBi incremental jumps). d) the more complex the shape the better but everything is rejected except those shapes that have a gain of 12 dBi or higher. e) complexity and length is rewarded. Each generation is checked for a better antenna, but is not rejected. a should find the target very fast, with performance getting worse and worse with each version. I'd be surprised if d gets anything. And as a funnel e is so overly generalized that it may never find anything useful. Although you'd probably end up with a monster of an antenna. 2. The main issue is this:
The environment must be totally blind to the replicator, in other words in no way it must be designed to favor some specific solutions in the replicator. It can, obviously, favor whatever solution arises “naturally”, as a consequence of necessity deriving from its internal rules and structure. But no “fitness function” must be purposefully written in the environment by the designer. The only “fitness function” will be a description of how the environment works, or is structured, or changes, exactly as it should happen in darwinian theory.
If I may re-interpret you: a) fitness functions in nature are not static, and will often not apply uniformly over many generations. Some fitness functions may not exist at all for certain problems, or they're so overly generalized that the search is not properly funneled b) neutral mutations are allowed but no fitness function can target specific configurations of them or guide them toward long term goals. If a series of neutral mutations manages to hit upon a configuration without any guidance some fitness functions will activate at that point. The weasel program is a good example since many intermediates do not have a function (comprise no english word) in the short term yet they're selected for anyway. c) most fitness functions must be limited to a short term target acheiveable in single step pathways. A new set of fitness functions will be generated based upon the new functional configuration of the replicator. d) in order to be realistic the fitness function must not contain a long term target that is specific in scope yet cannot currently effect the organism/creature in the short term. For example, you cannot have a fitness function targeting the long term functionality of the flagellum but you could have hypothetical fitness functions for intermediates. e) in nature there are examples where deleterious/destructive mutations may provide benefit in limited environments. There must be no preference for fitness functions that imply both benefit and constructiveness. f) some overly generalized long term fitness functions can be used but there must be a maximum ceiling balanced against other considerations. For example, for a hunter a generalized fitness function could be "an increase in speed in order to catch my prey". Problem is, this must be balanced against things like an extremely high metabolism and increasing speed at the expense of strength required to take down the prey. Patrick
Gpuccio You have written a lot. But I think it will be enough to concentrate on this bit: And it is not only “the problem” which “stems form the simulator’s brain”, but especially the process of solution, even if the solution itself is not known in advance. And the process of solution is an intelligent creation, a fruit of design, and it incorporates not only the active informatiout about the problem, but also the intelligent elaboration of the designer, his intuitions, his patient work, his purpose, his general knowledge and view of the world, and who knows how many other things. The process of solution of a GA has three parts: (a) The initial conditions (b) The variation mechanism (c) The elimination mechanism I am going to leave aside (a) and (b). I am sure there are GAs where these are done at random with no attempt to tune them to the end result. The real issue is (c). You use grand phrases about the designers’ work, intuition and intelligence but in the end this work is going to manifest itself as an elimination mechanism – a set of criteria for deciding who will survive and who will not. It doesn’t matter a toss whether the designer laboured on it for 20 years with the genius of Leonardi Di Vinci or stumbled on it while doodling and decided to give it a go. We need to understand what is it about a GA elimination process which makes GA an unacceptable simulation of NS? It isn’t the amount of work that was done beforehand! You hint at one difference when you write in the context of protein design: I recognize any approach to my function, by my capacity to define and measure it, and not by a spontaneous advantage which the function implies. . In other words the designer has worked out that the fitness function = selection criteria = measurement system will lead to the required function. But note that it is the measurement that decides what survives and thus plays the role of natural selection. It looks like your concern is that the measurement leads to an end objective over and above satisfying the measurement while NS does not. But not all GAs work that way. Some simply have a fitness function. Suppose for example you use a GA to solve the travelling salesman problem. Then the fitness function is simply the shorter solution survives – the end objective is the same as the fitness function. To summarise. Suppose we have a GA with the following properties: (a) Initial states are selected randomly within the domain space (b) Variation is not in way related to fitness function or any external objective (c) Survivors are selected in a manner chosen by the programmer on a whim and with no other end than to generate survivors that do well on the measurement used Would this satisfy you? Mark Frank
Toronto (#65): Your example, as I understand it, is interesting, but absolutely not pertinent to our discussion. First of all, a fractal generator is a necessity algorithm with some random seed. It is an example of "organization" based on necessity, but not of "functional information". Your fractals perform no function. Second, as you say: "Lifeforms die based on voter input. If too many visitors on the website don’t like the song or picture, it’s removed from the environment. If an output gets a lot of thumbs up, it survives and modifies a bit." And so on. In other words, the results are selected by intelligent consious people, according to their intelligent and consious sensibility. That has nothing to do with NS, and has nothing to do with function. It is just a collective rorschach test where people select what they like (or recognize) more. In a sense, it is an interesting variation of the weasel example (and I mean Shakespeare, not Dawkins). "In the US, enough generations might produce a picture of an eagle, while in Italy, you might end up with something that sounds like opera." That's exactly my point. The audience is modeling the content according to its projected representations. In a way, it's a form of (artistic) design. If I have missed something of what you mean, please let me know. gpuccio
Mark: So, let's try to draw some conclusions. You say: "If so, surely this is a significant piece of evidence in favour of the Darwinian process? All that remains is to show that it is possible for RM to generate complex and unanticipated solutions to problems which stem from the natural world rather than the simulators’ brains." Here is where your reasoning is not correct, in the light of what we have said before. The fact that highly sophisticated pieces of engineering can find solutions to specific problems, even by intelligently using random search, is in no way evidence in favour of the darwinian process. And it is not only "the problem" which "stems form the simulator's brain", but especially the process of solution, even if the solution itself is not known in advance. And the process of solution is an intelligent creation, a fruit of design, and it incorporates not only the active informatiout about the problem, but also the intelligent elaboration of the designer, his intuitions, his patient work, his purpose, his general knowledge and view of the world, and who knows how many other things. How could all that "stem for the natural world"? The natural world is as it is. Nothing stems from it, not problems, not processes of solution, not solutions. The best we can say is that, form the interaction between the replicator (which, at least, is a functional entity) and the natural world (which is only a passive scenario) the functions inherent in the replicator can be favoured, or suppressed, according to how they fare in that scenario. And that's exactly what I had requested in my proposed simulation: a passive scenario (although an organized one, with its laws and structures), a functional replicator, and random variation. That's all that is needed. In other words, the simulator must put his intelligence in building that scenario, but not in solving it. " While it may be difficult to create such problems in a simulation it is not clear why they the differ in principle." I hope I have explained why they differ in principle. "What are the key differences between a designed fitness function and a “blind” one such as NS? They both eliminate some individuals in a systematic way." Indeed, the fitness function is an ambiguous abstraction. Let's say that it is the replicator itself which survives or not survives, according to the resources it has (in NS); while it survives or not survives according to the planned expectations of the programmer (in IS). That is a lot of difference. The "fitness function" of the natural world, whatever it nay be, incorporates no design and no intelligence. The fitness function in an algorithm is a well defined product of design, and incorporates a lot. "Imagine someone did discover (a la Douglas Adams) that the earth was in fact a giant machine designed to generate complex life forms through RM and a complex environment. Would this suddenly invalidate the Darwinian logic?" Well, maybe I was wrong and you really are a theistic evolutionist in disguise! You had fooled me with all that false reasoning about a materialistic view of the world. :-) Really, I don't want to discuss here the inconsistencies of TEs. But if you are just implying that the earth is a giant machine designed to produce life by ETs, then I can answer: no, it isn't. If it were, we would see that. We would see the active information somewhere. Now, I am not saying that the earth is not tailored to favour life: I do believe it is. If it were not, life would simply be impossible. In a sense, the whole universe is tailored to favour life (see the fine tuning argument). But that is the most I can concede to TEs. That "favouring" does not include the generation of specific protein sequences, or of all the other information we observe in the biological world. That's why the biological world is so different from the non living world. For it, another level of design is necessary. Finally, I can admit that biological information could be generated by pure random variation and Intelligent Selection, like in GAs. But that is exactly an ID scenario (although not my favourite one). gpuccio
Mark: Your objections allow me to go into further detail about very important aspects of the question. First of all, I agree with you not to get tied into semantics about the word "process". I think we agree about what NS is in the darwinian theory, and that's the important thing. So, let's go on. You say: "It appears that for you it is a key issue that the fitness function in a GA is designed, while the fitness function in the real world is not. This is the big thing that invalidates GAs as simulations of the Darwinian process as far as you are concerned. Correct?" Yes, it's perfectly correct. But I would like to add immediately that it is not only a question of fromal difference: it is a question of utter substance, because in GAs the fitness function (and the algorithm which uses it) are designed in a way which incorporates a lot of "active information" (or, if we don't want to get tied into semantics, of useful knowledge) about the problem to be solved, plus a lot of intelligent planning about how to best solve it. It is not a small difference. More about that later. "Now do you accept that some GAs do generate complex solutions to designed fitness functions? So they establish that RM can generate complex and unanticipated solutions – even if the problem they address is anticipated?" Here the matter becomes more tricky. Let's try to analyze it better. GAs, like any intelligent software, can certainly generate solutions. And the solution is by definition unanticipated, while the problem is certainly anticipated. IMO, what differentiates GAs from other softwre is that GAs use a random search as part of the algorithm. That's why I have always compared GAs to other engineering processes which do the same, like protein engineering and antibody maturation. Now, to understand the question better, let's take an example which uses only necessity: a software program which can calculate succesive digits of pi. Now, let's pretend you are calculating the nth digit of pi: the solution is unanticipated (we don't know it in advance), but the problem is anticipated (we have to know the right algorithm which can correctly calculate that digit). The same algorithm can calculate many digits of pi (I suppose...). So, we have here an example of a program based on pure necessity (a mathemathical formula) which can give us a solution which becomes ever more complex with each new digit. But please, take notice of two important things: a) the algorithm is of pure necessity, and does not imply any random search. b) even if the complexity of the solution can increase, the specification remains the same, and is linked to the mathematical definition of pi. No new specification is ever created by the program. Now, let's go to intelligent algorithms which incorporate some random search as part of the process. Apart from examples which are pure propaganda, like Dawkins' weasel, where the program could as well write down the solution ibstead of looking for it by random search, that kind og algorithms, like those used in protein engineering, have a definite reason to exist: in that case, again, the designer knows the problem but not the solution, but he knows no algorithm based on necessity to reach the solution. In other words, he cannot calculate the solution. That's typically our situation with protein functions: we may know what function we are searching, but we have no idea of which aminoacid sequence can implement it. That's why, instead of a process of calculation based on necessity, we can adopt a process of trial and error, based on limited random search and some form of intelligent selection, usually a very sensitive measure of the desired function after each step of limited random variation. Such a method, if correctly designed, works. It is not easy, it is not quick, but it works. We know that. But it is important to understand why it works. We have a lot of intelligent programming and active information here, at different steps: a) First of all, there is usually a careful selection of the sequences we start from. That can be very important, and is based on what we understand of protein function. b) The random variation step is usually tailored as much as possible. For instance, in antibody maturation, it is applied only to the part of the sequence which has to be "improved", and not to the whole protein coding gene. In other words, it is random, but "targeted" variation. c) The engineer never asks random variation to do what it cannot do. In other words, the process of variation has the role of achieving as much variation as statistical laws allow to achieve with the available resources, and not more. That's why each step of limited variation has to be followed by a step of becessity (measurement and selection). d) Finally, and most important, the selection is intelligent selection, and not natural selection. That does not mean only that it is implemented by an intelligent engineer, but also, and especially, that it is completely based on our intelligent understanding of the problem: here is where most of the active information slips in. The difference is fundamental. I decide what function I am looking for: I am not expecting "any possible useful function", but a specific solution to a specific problem. I "measure" it, even at levels which could have no significant relevance in the general context of the environment. In other words, I recognize any approach to my function, by my capacity to define and measure it, and not by a spontaneous advantage which the function implies. My selection is therefore artificail, intelligent, and guided by a lot of active information about the result. More in the next post. gpuccio
2- Whether or not “evolution” is goal-oriented is being debated.
Toronto: gpuccio @[48] seems to agree that the process of evolution has no goal.
yes "evolution" as it is currently accepted and applied.
ROb: Correct me if I’m wrong, but ID scenarios seem to imply the latter. That is, the designer was not using evolution to find a solution to a problem, but rather to instantiate an already-known solution.
ID doesn't say. All ID says is that there are some things in the universe (and perhaps the universe itself) which show signs of being intelligently designed. As for the implementation that would be another question that can be pursued after design is determined. Joseph
Gpuccio The real world, in darwinian thought , does not “include” any “process” for selection. NS is just a consequence of how the real world is, and of how a replicator works. It is not a “process”. It is a blind effect, due to blind laws of necessity. First a comment on terminology: a process does not have to be designed. Stalactites grow through a process; stars are born, shine and then die through a process. These are both blind. They just happen. But let’s not get tied into semantics. It appears that for you it is a key issue that the fitness function in a GA is designed, while the fitness function in the real world is not. This is the big thing that invalidates GAs as simulations of the Darwinian process as far as you are concerned. Correct? Now do you accept that some GAs do generate complex solutions to designed fitness functions? So they establish that RM can generate complex and unanticipated solutions – even if the problem they address is anticipated? If so, surely this is a significant piece of evidence in favour of the Darwinian process? All that remains is to show that it is possible for RM to generate complex and unanticipated solutions to problems which stem from the natural world rather than the simulators' brains. While it may be difficult to create such problems in a simulation it is not clear why they the differ in principle. They may be more more complex but they don't present any new problems in principle. What are the key differences between a designed fitness function and a “blind” one such as NS? They both eliminate some individuals in a systematic way. Why should a process that can solve one type of problem not be able to solve the other? Imagine someone did discover (a la Douglas Adams) that the earth was in fact a giant machine designed to generate complex life forms through RM and a complex environment. Would this suddenly invalidate the Darwinian logic? Mark Frank
Mark: "Any simulation of RM+NS has to deal with both RM and NS. RM is random in the sense that the mutation is independent of the selection criteria. But NS, which is the simulated by the fitness function, is far from random." Well, that's exactly my point. Exixting GAs deal certainly with RM, but they absolutely don't deal with NS. All of them deal with Intelligent Selection (IS). Let's speak a little of the famous "fitness function". In reality, no fitness function exists. Functions are just our intelligent creations. The problem is, the fitness functions created in GAs don't represent in any way what is assumed in darwinian theory. And that's not only because biological reality is different from digital reality: that is a basic problem of all simulations, both GAs and the one I am proposing. That is certainly a limit, but it is not my point. My point is that teh fitness function in GAs is an intelligent function, devised exactly to obtainin in some more or less indirect way, what the designer of the simulation wants to obtain. That's only design in a cheap tuxedo. In darwinian theory, NS is only a blind effect which derives form the interaction of two different realities: the environment, or landscape, or whatever you want to call it, and the functional reality which we call the replicator. Two points are essential to define some process as NS: a) The environment must be totally blind to the replicator, in other words in no way it must be designed to favor some specific solutions in the replicator. It can, obviously, favor whatever solution arises "naturally", as a consequence of necessity deriving from its internal rules and structure. But no "fitness function" must be purposefully written in the environment by the designer. The only "fitness function" will be a description of how the environment works, or is structured, or changes, exactly as it should happen in darwinian theory. b) The replicator can be as functional and as complex as we want: in my simulation, it represents the result of OOL (which we are not simulating here). The only requirement is that nothing must be frontloaded in the replicator to "guide" or help the future variations. In other words, any future variation must be truly random (and, as I have said, the variation mechanism can apply any statistical distribution you like, uniform or not, provided that no information is inputted about some specific functional solution). That is the concept of NS as it is outlined in darwinian theory. That's what we have to simulate. Anything which does not have those two properties, is neither NS nor a simulation of it. It is just some form, more or less in disguise, of IS. "Nevertheless it is not random." I have never said that NS is random. It is a process of necessity. But it has to satisfy the two criteria I have detailed, otherwise it is not NS. In other words, NS is a "blind" (not "random") process of necessity. "Then the mechanism of RM+NS would lead to species with ever greater P/W ratios." In that simulated world, you would only obtain (if you are lucky) the same replicators, with slightly greater P/W ratios (as far as that is possible without damaging the existing functions, and by means of simple random variations, which would never approach the level of CSI). In other words, you attain (if you are lucky) exactly what you specified in your intelligent fitness function. If you specify greater P/W ratios, you can obtain that and nothing else. If you specify flying objects (whatever that may mean in a digital environment) you can obtain that and nothing else. If you don't specify anything, you obtain nothing. That's exactly, IMO, the intuitive meaning of the concept of active information. And specification, as you well now, is the first requirement for CSI and design. "But, just like the real world, any simulation must include a process for selection based on something." The real world, in darwinian thought , does not "include" any "process" for selection. NS is just a consequence of how the real world is, and of how a replicator works. It is not a "process". It is a blind effect, due to blind laws of necessity. Maybe a theistic evolutionist could argue that the environment is designed to produce life, but that's another story. I believe, to your merit, that you are not a theistic evolutionist. gpuccio
Sal Gal and Mark: As usual, very interesting discussion with you. I offer my counter arguments to what you say, and in defense of my "thought simulation". Sal Gal: "But you are have not spoken to death of virtual organisms." Death can well be included in the original programmimg of the replicators. You can program the replicators as you like, provided that you don't use any explicit or implicit form of frontloading, which could be anyway easily revealed by an impartial scrutiny of the original software. "There are an estimated 5 X 10^30 bacteria on earth at present. I won’t guess how many elementary self-replicators there might have been prior to cellular life, but will suggest that they may have reproduced more rapidly than bacteria. In any case, our computational resources are relatively tiny." Well, I am not so sure they are so tiny, although I have not made the calculations. Even if we work with less replicators, in a digital environment we can greatly increase the times of variation and of replicating fitness evaluation. And anyway, I am not saying that such a simulation is easy, quick, or that it could be done with small computational resources. I am only saying two things: a) It is possible in principle to realize that simulation. b) It is the only appropriate kind of digital simulation for darwinian theory. "As long as the virtual environments are uninteresting, there is no reason to expect virtual organisms to be anything but the same. One thing that makes an environment “interesting” is that there are many ways to garner resources and reproduce." You can make the operating system as interesting and varied as you like, provided that it contains no specific code to recognize and ecourage any specific thing which you want to attain. In other words, the selection must be self-selection, deriving from the interaction between the functional replicators and the rules of the environment, and must in no way be programmed by the desginer of the simulation. "This is not sensational, but it is qualitatively what you said you wanted to see." I have no problem in admitting that GAs can do a lot of things, like any other software can. The only thing they cannot do is simulate the logical mechanism of darwinian evolution. What I wanted to see is some form of evolution of CSI in simulation of the kind I have suggested. It is obvious that GAs, being intelligently designed problems, can find solutions which the designer did not know in advance. If a computing software is designed to solve an equation, that's exactly because the designer does not know in advance the solution. But that does not make it a simulation of darwinian theory. And finally, I do believe that you should play a little with real proteins. It's rather easy, and it could be an amazing experience. Mark, next post is for you. gpuccio
But you are have not spoken to death of virtual organisms. You also have not addressed the massive parallelism of life. Gpuccio Sal Gal is spot on. Any simulation of RM+NS has to deal with both RM and NS. RM is random in the sense that the mutation is independent of the selection criteria. But NS, which is the simulated by the fitness function, is far from random. What confuses the issue is that the fitness function for real species are very complicated, hard to determine, and keeps on changing over time. Nevertheless it is not random. If the criteria for survival in each generation had no relation to the criteria for survival in the previous generation then evolution would not get off the ground. GAs typically specify a simple and unchanging fitness function. This is like a world where there is only one attribute that matters in the struggle to survive. For example, imagine a world where the only organisms to survive were those with the greatest power/weight ratio. Then the mechanism of RM+NS would lead to species with ever greater P/W ratios. This world has just as much knowledge of its target as a GA, but would still be a Darwinian process. You see something like this when artificial selection limits the fitness function to one or two criteria desired by man. You seem to be demanding that the fitness function arise spontaneously in the simulation. But, just like the real world, any simulation must include a process for selection based on something. Mark Frank
gpuccio, I have agreed with your general argument, though not the details, for years. But you are have not spoken to death of virtual organisms. You also have not addressed the massive parallelism of life. There are an estimated 5 X 10^30 bacteria on earth at present. I won't guess how many elementary self-replicators there might have been prior to cellular life, but will suggest that they may have reproduced more rapidly than bacteria. In any case, our computational resources are relatively tiny. Various folks, including me, who are interested in both information and evolution think of organisms as modeling their environments. As long as the virtual environments are uninteresting, there is no reason to expect virtual organisms to be anything but the same. One thing that makes an environment "interesting" is that there are many ways to garner resources and reproduce. I should mention that GAs have discovered bugs in application-specific software the GA developers had not examined. That is, the GAs came by fit individuals that exploited errors made by programmers other than those who wrote the GAs. This is not sensational, but it is qualitatively what you said you wanted to see. P.S.--Thanks for pointing me to SCOP and BLAST. I'm interested, but I just can't follow up at the moment. Sal Gal
gpuccio[62] How about this? A 256 bit "DNA" based fractal generator. char DNA[32]; where DNA[0] to DNA[15] select any one of f1() to f16() in any order with the parameters from DNA[16] to DNA[31]. x=f1(DNA[16],0){yadadyada..}; x=f2(DNA[17],x){yadayada...}; x=.... r=f16((DNA[31],x){yadayada...}; The result r contains the seed and fractal generator. The final output, determined at birth, is used either to generate a video bitmap or fed to a DA converter for sound output which means each DNA/life, doesn't know what it's generating as an output, audio or video. Lifeforms die based on voter input. If too many visitors on the website don't like the song or picture, it's removed from the environment. If an output gets a lot of thumbs up, it survives and modifies a bit. If it survives long enough, (100 votes?), it gets to mate with another successful DNA/life whose offspring have attributes of both parents. In the US, enough generations might produce a picture of an eagle, while in Italy, you might end up with something that sounds like opera. Toronto
R0b: I will try to clarify. 1) The system must not be designed to run the simulation, just as the environment (in darwinian theory) is not designed to produce life (or any higer form of life). The system will obviously have its rules of functionality (its "fitness functions"), but those rules must not have been written in view of the simulation. Just as it would have happened in natural history, according to darwinian theory. 2) Yes, we are simulating biological evolution in a computer environment. Obviously, a computer system cannot have the characteristics of our terrestrial environment, neither in my simulation nor in any other GA simulation. And, least of al, can the digital replicators have the characteristics of true biological replicators. But that's not the point. The point is that we are simulating the model according to which random noise, plus a natural selection deriving from the unprogrammed interaction between modified replicators and a fucntional environment, can build up new functional information and generate more complex and functional replicators. That's the point of the simulation. Even if the digital context is obviously different from the biological one (and there is no escape for that, in any digital simulation), the logical model is the same. So the simulation, although not perfect, is appropriate, while traditional GAs are not appropriate because they are not modeling a blind natural selection, and therefore those models are logically completely different from what they are trying to simulate. 3) The point is exactly to verify if what you call "the replicators that are better aligned to the environment" will really arise in this model. I believe and expect they never will. 4) You say: "I don’t know where the line is between “more functional” and “new function”, so that’s something you’ll need to define for me.". There is no line. Any new replicator which is better aligned to the environment and spontaneously expands in the environment to the expense of the previous forms is more functional and has developed a new function. There is no difference between the two concepts. 4) Personally, I do believe that those more functional replicators will never arise in such a system, as I believe that they have never arised in natural history by a similar causal mechanism. But I am ready to analyze any possible outcome of such a simulation. Morever, if a new form of replicator is really selected, it would be easy to verify what the new function is, how it gave the new replicator a reproductive advantage, and above all what is the statistical boundary which has been overcome (how many functional bits have been added). According to darwinian theory, even if the first advantages may be trivial (within the range of statistical expectations), in time the accumulation of such useful variations should give functional variations of the order of CSI (500 bits). My belief and expectation, on the contrary, is that we would never ackowledge even the first, microevolutionary steps, or if we do they will be absolutely trivial variations, and they will stop at that level. I hope that is clear. I really believe that, if we want to simulate the logical model of darwinian theory, we have to proceed that way. Otherwise, we are only playing games, doing one thing and claiming we are doing another. And, by the way, that's what I meant when I said that GAs don't model anything. I meant that they don't model anything even distantly related to the logical model of darwinian evolution. But it is true that they can model an intelligently designed way to intelligently find a solution. gpuccio
gpuccio, first of all, I don't understand what you mean when you say that the system must not be designed to run the simulation. What are we simulating? Biological evolution? If so, then the system should have the characteristics of our terrestial environment. It needs to simulate whatever it's supposed to simulate. Regardless, selection will be determined by the rules and characteristics of the system. The replicators that are better aligned to the environment will reproduce more. Such replicators are more functional, in terms of that environment, than other replicators. I don't know where the line is between "more functional" and "new function", so that's something you'll need to define for me. R0b
R0b: "Active information is defined in terms of a target. What is the target of the proposed simulation?" That's exactly the point: my proposed simulation has no target, out of verifying if a simulation which has no target can obtain some new functional target. That's exactly the point, and that's exactly what darwinian evolution is believed to be: a process with no target, which can find some extraordinarily functional targets. I don't believe that can happen. But darwinists believe exactly that. So why don't they try to simulate that kind of process, instead of trying to realize processes which see very well, and then pretending that they are simulations of a blind process? Again, and to be clear to the extreme, I admit that the simulation can exercise natural selection, but not intelligent selection. In other words, the selection must arise spontaneously from the interaction between the replicator and the system, without any previous programming. In that way, and only in that way, a true (blind) natural selection can take place versus any possible new useful function which may arise from true blind random variation. AQny possible useful new function is the target. No specific target at all, the hreatest possible target of all: any possible new useful function. True blindness, true random variation, true blind natural selection. Isn't that what darwinists have been preaching for all these years? gpuccio
gpuccio:
GAs know a lot about what they want to find, and how to find it.
I said that GAs typically know nothing about the location of the target. But yes, what they are trying to find and how to find it is incorporated into the GA.
The weasel is just a gross example, but wouldn’t you agree that the weasel algorithm knows exactly the solution it is searching for?
No, I wouldn't, but I think our disagreement is semantic. You seem to see the objective function as part of the GA, while I see the two as separate. The weasel search algorithm, for instance, starts with a random string precisely because it has no idea what the target string is. The fitness function (oracle), on the other hand, knows the target string.
I am not specially interested in GAs, being absolutely convinced that they are completely useless, and that they don’t model anything, except for the ability of their programmer.
And yet GAs are able to find solutions that their programmers don't know beforehand. And I'm not sure what you mean when you say that they don't model anything. If you're referring to biological evolution, then obviously nobody claims that computers can model that with any degree of fidelity, although they can provide us with some insight into the variation+selection process on a small scale.
...and, above all, no active information introduced in the system.
Active information is defined in terms of a target. What is the target of the proposed simulation? R0b
R0b: "GA’s typically know nothing in advance about the location of optima, but they do act on the assumption that the landscape is reasonably smooth." I don't agree. GAs know a lot about what they want to find, and how to find it. The weasel is just a gross example, but wouldn't you agree that the weasel algorithm knows exactly the solution it is searching for? Other GAs may not know exactly the solution, but they do know a lot of other precious information. They don't certainly know (or assume) only "that the landscape is reasonably smooth". I am not specially interested in GAs, being absolutely convinced that they are completely useless, and that they don't model anything, except for the ability of their programmer. But many times, here and elsewhere, I have declared what a true computer simulation of darwin's theory should be, usually receiving no answer or comments. IMO, such a simulation should work more or less as follows: 1) Take a digital environment: a computer running some operating system, and any software we like. The digital environment can be stable or change, but the important point is that it should in no way be programmed for the simulation. 2) Take a program which generates digital replicators, and let it run in the system. The program can incorporate some system of random variation, which can be regulated as we like, but it has to be completely random, and we can apply any random probability distribution we want, but again then important point is that no programming must be introduced which has any relationship with the simulation. The whole system will so be blind to the simulation, exactly like darwinian theory assumes. 3) Just let the system and the software run, and wait. For what? For any variation in the replicator which is spontaneously selected by the system as useful. That would really be an evolutionary simulation. You have everything: a system which is not designed to run the simulation (in other words, a blind system), but which has its own rules and characteristics, like any true landscape; replicators which start as functional (we are not simulating OOL here), and are subject to random variation, adjustable as we like; and, above all, no active information introduced in the system. That, and only that, would be a real simulation of darwinian evolution. Can you see the differences with existing GAs? gpuccio
Joseph:
2- Whether or not “evolution” is goal-oriented is being debated.
Interestingly, Marks and Dembski's work cannot address that question, because you have to know that there is a goal and know what it is before you can even begin an analysis in their framework.
3- In any goal-oriented scenario the goal is being searched for.
Actually, we need to distinguish between a goal in the codomain of the objective function vs. a goal in the domain. If, as in a prototypical optimization problem, the goal is defined in terms of the codomain, it makes sense to describe the process as a search for a solution to a problem. If the goal is defined in terms of the domain, then we're not searching for anything, but simply moving toward a goal whose location is already known. Correct me if I'm wrong, but ID scenarios seem to imply the latter. That is, the designer was not using evolution to find a solution to a problem, but rather to instantiate an already-known solution. R0b
Joseph[50]
2- Whether or not “evolution” is goal-oriented is being debated.
gpuccio @[48] seems to agree that the process of evolution has no goal.
..because a true evolutionary model must know nothing in advance of the solution it finds.
If evolution had a goal it would be a form of ID. Toronto
gpuccio:
The concept is probably that the author of a simulation of evolutionary search would want, and indeed does want, to hide how his algorithm performs well only because of the specific information about the solution which has gone into the programming of the algorithm.
The active information metric measures the algorithm's performance relative to the given problem. To say that an algorithm performs well because of X bits active information is to say that it performs well because it performs well.
The idea is simple: that makes the algorithm efficient, but it makes it a very bad evolutionary simulation, because a true evolutionary model must know nothing in advance of the solution it finds. Is that clear?
Actually, no. Efficient algorithms have lots of active information, by definition. Does that make them bad? "True evolutionary models", assuming they perform better than random sampling, have plenty of active information. GA's typically know nothing in advance about the location of optima, but they do act on the assumption that the landscape is reasonably smooth. The correctness of that assumption constitutes active information. R0b
Thanks Prof. Olofsson. On reading II.B, I see that I was completely misinterpreting their vertical NFLT examples. It appears to be perfectly general. That'll teach me to try to speed-read technical papers. R0b
Prof. Olofsson: "As for the kind of papers you ask for, you can check out Rick Durrett’s publications for a start." Does he show how background noise (chance) combined with an arbitrary collection of laws (set of laws put together absent any consideration for future results -- absent foresight) will produce a system of signs and the mechanisms to process them, a search space amenable [where functions aren't spaced too far apart] to a ratcheting filter (natural selection), and the laws necessary to create environments that will allow life to evolve to ultimately highly improbable and functional machines and intelligence (systems with the ability to model the future and generate targets). Basically does he show that life and evolution to produce intelligence will occur from any arbitrarily chosen set of laws and initial conditions? I would definitely be interested in seeing such research. CJYman
ROb[44], As you can see, there is no mention of fitness functions in the "search for a search" paper. A "search" is defined as a probability measure over the search space (see the construction on page 2, B); this probability measure may or may not involve the use of fitness functions. Prof_P.Olofsson
Joseph[51], You're avoiding th issue. If you claim these papers to be "pro-ID" you ought to demonstrate what bearing they have upon biology. As for the kind of papers you ask for, you can check out Rick Durrett's publications for a start. Prof_P.Olofsson
djmullen: No, things are very different from what you are saying. You say: "But every living organism that exists is the product of organisms that are already in that tiny proportion of the possible genomes and if they successfully reproduce then their DNA is already in it too." Absolutely not. First of all, we cannot speak of the whole genome (that would be beyond any possible analysis). We have to analyze some single portion of genomes. As I have said (see my answers to Sal Gal) the only portion we can really compare (because it is the only portion we really understand) is the protein coding genes. Now, while there are some similarities between some proteins in different species, others are completely different. Practically each species has proteins which are absolutely peculiar to that species. I quote here form one previous post of mine: "The proteins we do know (and we know a lot of them) are really interspersed in the search space, in myriads of different and distant “islands” of functionality. You don’t have to take my word for that. It’s not an abstract and mathematical argument. We know protein sequences. Just look at them. Go, for example, to the SCOP site, and just look at the hyerarchical classification o protein structures: classes (7), folds (1086), superfamilies (1777), families (3464). Then, spend a little time, as I have done, taking a couple of random different proteins from two different classes, or even from the same superfamily, and go to the BLAST site and try to blast them one against the other, and see how much “similarity” you find: you will probably find none. And if you BLAST a single protein against all those known, you will probably find similarities only with proteins of the same kind, if not with the same protein in different species. Sometimes, partial similarities are due to common domains for common functions, but even that leaves anyway enormous differences in term of aminoacid sequence." Just to give you an idea: bacteria have about 500-2000 protein genes, while humans have 20000-25000. Drosophila has about 14000. Each protein, and each protein fucntionality, is a different island in the sae of possib le sequences. We have hundreds of thousands of different proteins. Some of the most complex proteins are more than 2000 aminoacids long. So, maybe you are not familiar with biology (no problem there), but when you say: "This keeps the results of their so-called “search” either inside the sweet spot or very close to it. No other “search strategy” is necessary." you really don't know what you are speaking of. gpuccio
To all of those who doubt these papers are pro-ID: Can you point to any peer-reviewed papers that demonstrate the power of blind, undirected processes pertaining to biology? Joseph
Evolution is a blind, unguided process. ID is a goal-driven process. Since ID has it’s goal and evolution doesn’t care for one, neither are searching for anything.
1- "Evolution" is not being debated 2- Whether or not "evolution" is goal-oriented is being debated. 3- In any goal-oriented scenario the goal is being searched for. Joseph
What I don't understand is what does searching have to do with evolution? It looks to me like the ultimate concern of the authors is searching through all possible genomes for the tiny proportion of those genomes that will construct and operate an organism that can successfully reproduce. But every living organism that exists is the product of organisms that are already in that tiny proportion of the possible genomes and if they successfully reproduce then their DNA is already in it too. In practice, they do this by making only small changes to the DNA they hand down to their offspring. This keeps the results of their so-called "search" either inside the sweet spot or very close to it. No other "search strategy" is necessary. Evolution says that every living organism since the first self-reproducing molecule is already in the genomic sweet spot and all they have to do is keep their offspring in it too. djmullen
R0b: "What I find interesting is the idea that someone would want to hide how well their algorithm performs." The concept is probably that the author of a simulation of evolutionary search would want, and indeed does want, to hide how his algorithm performs well only because of the specific information about the solution which has gone into the programming of the algorithm. The idea is simple: that makes the algorithm efficient, but it makes it a very bad evolutionary simulation, because a true evolutionary model must know nothing in advance of the solution it finds. Is that clear? gpuccio
Bill Dembski, When you first announced your publications here at UD, you linked them to complex specified information. Yet I can find no explicit reference to CSI. And I genuinely do not see it between the lines. Would you explain the connection to CSI? Sal Gal
gpuccio:33 "Well, NS is a very limited oracle, and it can only judge if a reproductive advantage has been achieved or not." Whether an agent reproduces or not is always a binary event, obviously. How NS differnetiates between agents that reproduce or not is not a binary event, it is a probability. Roughly speaking the fitter the individual the greater their probability of reproducing and, depending on the type of GA, higher fitness individuals have a greater chance of producing greater numbers of offspring. "NS can do only two things: expand a genome if a perceptible reproductive advantage has been achieved (positive selection); or eliminate it if there has been a significant loss of function (negative selection)." Genomes with a lower fitness than their parents still have a probability of reproducing, it is just lower. Genomes with the same fitness as their parents have the same probability of reproducing, even if they are different. And that all assumes that the fitness landscape stays the same, which in biology and in some GA implimentations is not always the case - you can end up fitter than your parents because of an environmental change even if at birth you were less fit. Either way we are talking (and the paper is talking) about search algorithms and my point still stands - A binary fitness evaluation presents no slopes, even if they actually exist in the fitness landscape. Removing the ability to detect any graduated differences in fitness between individuals or iterations effectively converts a variable fitness landscape into a flat one. A hill climbing algorithm is ONLY an hill climbing algorithm if there is a hill to climb and a GA is ONLY a GA if there is some way of ranking individuals. Otherwise you just have an individual taking a random walk, or a population of individuals taking a random walk. To imply that these strategies are 'adding information' is like arguing that an aeroplane is no better at flying than a car because the aeroplane can't fly without wings. Laminar
A comment on the last conclusion of the 1st paper:
To have integrity, search algorithms, especially computer simulations of evolutionary search, should make explicit (1) a numerical measure of the difficulty of the problem to be solved, i.e., the endogenous information, and (2) a numerical measure of the amount of problem-specific information resident in the search algorithm, i.e., the active information.
So whoever publishes an algorithm for solving a particular problem should make explicit (1) the size of the problem space and (2) the efficiency of the algorithm applied to that problem. What I find interesting is the idea that someone would want to hide how well their algorithm performs. R0b
Not that anyone cares, but I answered my own question from #21. The piece I was missing was the fact that the amount of active information changes over the course of the search. The fact "the fitness function is smooth" contains no active information initially, since by itself it provides no guidance as to the location of the target. That's why the first query of genetic algorithms is typically random. It's only after we start querying that the distribution starts to become biased. So maybe the vertical NFLT is general enough to handle all kinds of information about the target and fitness landscape. I would be interested to see it formally applied it to a smooth fitness landscape. R0b
Sal Gal: Long time no see, friend. It's always a pleasure to meet you. You say I may have missed your point, but I don't believe that. I have purposefully avoided to discuss Dembski and Marks'points, mainly for lack of competence on my part. But I have commented on what you were saying. I quote again: "There is huge inconsistency among IDists on the matter of functionality of DNA sequences." That's not true, as I have tried to show. We do believe that, in general DNA sequences are functional, but in different ways: protein coding genes are functional as "storage" of aminoacid sequences; non coding DNA is functional as a regulatory information, at present poorly understood, and probably much more complex than the "simpler" protein coding information. Let's say that protein coding DNA codes for the "effectors", while non coding DNA codes, in some mysterious way, for the procedures. Or for part of them (other unknown components will probably be discovered in time). Let's remember that at present we know almost nothing about the procedures. We don't know how transciption is regulated, how and why cells differentiate in myriad of forms, how multicellular plans are achieved, how intercellular communication is controlled at higher levels of integration, and so on. Procedures. In any software, those are the real thing. I quote again: "But when they want to portray evolution as utterly improbable, they say that a sequence of bases is categorically fit if it codes for a prespecified protein, and is categorically unfit otherwise. Now which is it? Can a non-coding sequence contribute to fitness, or not?" As already said, we reason on the protein coding genes, because that is the part we understand (and that "we", for once, includes both darwinists and IDists). We will reason about the regulation part when we will understand it. And believe me, that will not be good news for darwinists. So, leaving alone Dembski and Marks, where is it that I have missed your point? And is my point clear? To answer explicitly your question, a non coding sequence can certainly contribute to fitness, but only to the fitness related to its specific function, probably a regulatory one. It cannot certainly contribute to the fitness inherent in a protein coding gene, which is another thing, works in another way, has another symbolic code, and so on. So, for protein coding genes, we have to go back to the only possible oracle if we are to believe in unguided, non designed evolution: NS. I have dealt with that in my previous post, so I refer you to that. In other words, all "warmer or colder" effect must be attributed only to natural selection (or to design). And it is a very "threshold" effect: only if the message is "very warm" (the new replicator replicates better than the old ones, and can expand in the population) can the variation in genome be fixed, and we have something different from a random search on a uniform distribution. Or, in alternative, if the message is "very cold", and the new genome is terminated. In all other cases, there is no oracle, and no other supporting or miraculous force. We are back to random search over an uniform distribution of possibilities. Or to design. gpuccio
pubdef: "What I’m trying to say is: notwithstanding the characterization of DNA as “digital information,” isn’t it really a physical configuration of molecules that interacts with other physical objects/particles, leading to a far flung range of physical consequences? Another stab at it: DNA seems to be viewed as software, but isn’t it really hardware?" As Upright BiPed has already said, DNA is a support for information: you cannot have pure software, you always have software written on a hardware support. The "physical configuration of molecules that interacts with other physical objects/particles" is only the biochemical structure of the DNA molecule. But the special sequence of nucleotides which, in a symbolic code, encodes the sequence of aminoacids in a specific protein is pure digital information. There is no physical or biochemical law which determines that sequence. The sequence is preserved in DNA, and it must have been created in some way. Darwinists believe that it was created through RV + NS. We (IDists) believe that it is the product of designe. But in no case it can be the product of necessity. There is no law of necessity which can output the correct sequence of nucleotides which corresponds to the aminoacid sequence of, say, myoglobin, according to a definite symbolic genetic code (the genetic code we observe in biological beings). Darwinists, and even many biologists, are really confused when they speak of non uniform distributions, or of landscapes, or of fitness functions, as though those concepts could apply to the random generation of new nucleotide sequences. There are no laws which can generate randomly a sequence of nucleotides in a pattern which is really distant from an uniform distribution, least of all in a pattern which may have anything to do with the genetic code or with functional aminoacid sequences. All that talk about evolutionary algorithms is just smoke in your eyes. The truth is that any component in darwinian theory which can apply to the possible generation of information in a way which differs from a random search on a uniform distribution must be related to NS. And NS is a form of necessity. Call it evolutionary algorithm or any other name, it is only NS which can make a difference. But NS is not an omnipotent principle. As all laws of necessity, it must be modeled according to necessity. It is indeed a potential oracle, but an oracle of which we know all too well the limits: it can only expand genomes with a reproductive advantage or eliminate genomes with important loss of function. Out of those limits, NS is non existent. So, again, darwinists should be able to show where NS can act to "join" informational leaps which are in the range of what a random search over an uniform distribution can empirically accomplish. If they cannot do that, they have nothing: not a model, not a theory, nothing. And believe me, they cannot do that. gpuccio
P.S. to 39: Free associating on Toms -- that's Schneider, not Ray. Sal Gal
gpuccio, Long time no see, friend. I think you may have missed my point, which was a criticism of modeling the fitness of a DNA sequence as all-or-nothing. I am saying that I think it's inappropriate for Dembski and Marks to attribute all "warmer or colder" information to an "oracle" or an "assistant." Sal Gal
So if you’ve got a problem with the applicability of the research at the Evolutionary Informatics Lab to real-life biological evolution, take it up with Schneider and Pennock.
Evolution of Avida programs is not combinatorial optimization. When the running time for Avida is bounded, proceeding from shorter programs to longer programs is generally superior to proceeding from longer programs to shorter programs. Avida affords itself of that free lunch. Only you and Bob, to my knowledge, have tried to bring "no free lunch" results to bear on Avida. So I believe it is reasonable to take up the issue with you. And I'm curious as to whether you and he elected to remove the attacks on ev and Avida from the first paper, or if the reviewers required it. As for ev, Ray has contended that you have not understood what he meant to demonstrate. I have not reread his paper to see if he's right, and I won't bother, because it seems to me that you're beating a dead horse. Computational studies today are more sophisticated than ev. Sal Gal
"Another stab at it: DNA seems to be viewed as software, but isn’t it really hardware?" No more than a CD-ROM is hardware; is has a physical reality, but it is the information contained within that is its purpose. DNA is a vessel of information (of the functionally specified variety) that animates and organizes inaimate matter into living tissue. It is the only thing in the universe that does so. Upright BiPed
#32:
No, the rock is not information. And your example has nothing to do with DNA. DNA (at least, the protein coding genes) is digital information which stores a sequence which, when translated into an aminoacid sequence, has a specific function. We know the code, we can read it. We know, in many cases, the product (the protein) and its function. What has all that to do with your rock?
I will concede that my point may be "clear as mud," and I appreciate that anyone makes an attempt to answer it. What I'm trying to say is: notwithstanding the characterization of DNA as "digital information," isn't it really a physical configuration of molecules that interacts with other physical objects/particles, leading to a far flung range of physical consequences? Another stab at it: DNA seems to be viewed as software, but isn't it really hardware? pubdef
gpuccio:
Natural laws do not imply any new information except obviously the possible information in the original setting of the laws themselves: that is an aspect of the fine tuning argument, which is very different, and separated, form the main ID argument of information in biology.
Since Dembski and Marks put no restriction on what processes can be modeled as searches, any process that isn't uniformly random involves active information. (In fact, even uniformly random processes can be modeled as active-info-rich searches by simply modeling the search space in such a way that the sampling isn't uniform. If a process generates random circles whose diameters are uniformly distributed, simply define the search space in terms of the areas rather than the diameters.) So natural laws are active-information-rich according to Marks and Dembski's definitions. Their claim is that this active info must be referred to a higher-level search (or, presumably, to intelligence, although they don't say that in the papers), which can in turn be referred an even higher-level search, ad infinitum. Regularities that just are don't seem to be an option in this framework.
That equals, obviously, knowing the searched phrase. In other words, you can get the weasel phrase only of you already know it. But then why search for it?
The active information metric is defined in terms of a prespecified target T. If prespecified targets are a problem, then they're a problem for Marks and Dembski's whole framework.
That’s why we speak of a search: because darwinian evolution needs a system which acts as a search.
Who is "we"? The idea that biological evolution is searching for something is certainly not the mainstream view.
That’s why you will never obtain better digital replicators just by random noise in a digital system. The concept does not simply work. Mathematicians can argue about why it does not work, but we do know that it does not work: you just have to try to detail it or put it to test, both in a computer environment or in a biological environments. The only space where that system seems to work is in the mind of darwinists.
Genetic algorithms certainly do work. Whether they scale well enough to produce our earthly biota is another question. R0b
gpuccio @30 Evolution performs no search, it is in itself, a result. Evolution is not the trip, it's the destination you arrive at. Whether implied by Darwin or anyone else, there is no search process involved. Since evolution has no goal, we can't tell where it's going or how easy it will be to get there. Toronto
WilliamDembski[26], I skimmed the "search for a search" paper. It's a nice paper. I think there is a simpler proof of "conservation of uniformity" for finite Omega, one that uses elementary methods and does does not invoke weak convergence of measures. Consider the vector [p(1),...,p(n-1), p(n)] where [p(1),...,p(n-1)] is uniformly distributed and p(1)+...+p(n-1)+p(n)=1. Integrating over the simplex gives the joint pdf and the marginal of each p(i), and the expected value in the marginal is 1/n. I understand that the authors are pro-ID but I don't see how the paper itself can be labeled pro-ID." What is the logic behind such a claim? Because the darwinian search obviously is not uniform, it has not been chosen according to the uniform distribution induced by the Kantorovich-Wasserstein metric, hence there is support for ID? I also don't understand your comment. What do Schneider, Pennock, and Behe have to do with how your paper is relevant to biology? Prof_P.Olofsson
Laminar: "This is an example of a random search with a binary evaluation criteria (it either is the works of Shakespeare of it isn’t) There is no descent with modification and no fitness metric, just a halting condition. It is NOT an evolutionary search." Well, NS is a very limited oracle, and it can only judge if a reproductive advantage has been achieved or not. I really think that all that talking of landscapes and fitness functions creates a lot of confusion. Let's be simple: NS can do only two things: expand a genome if a perceptible reproductive advantage has been achieved (positive selection); or eliminate it if there has been a significant loss of function (negative selection). That is the only oracle you have in the darwinian model, nothing else. And I am still expecting that a credible molecular model for macroevolution be detailed with that kind of oracle correctly utilized, and the random variation part correctly computed. Is that asking too much from such an extraordinary scientific theory as darwinian evolution is supposed to be? gpuccio
pubdef (#19): "Suppose the course of a creek is altered by a rock that falls into it. Better, suppose that the new course leads the creek into country that is more amenable to creek-dom, and the creek becomes a river. Is the rock “information?” My real point: how is this different from DNA (or RNA — I don’t have the specifics of that science)? A physical substance affecting another physical substance, with extensive consequences in the long run?" No, the rock is not information. And your example has nothing to do with DNA. DNA (at least, the protein coding genes) is digital information which stores a sequence which, when translated into an aminoacid sequence, has a specific function. We know the code, we can read it. We know, in many cases, the product (the protein) and its function. What has all that to do with your rock? gpuccio
PO--I wrote “relevant” which is not the same as “related.” Relevant is relevant to related and related is related to relevant Don’t you read Merriam-Webster? It has an awful plot :-) tribune7
R0b: Just a few thoughts on what you say: 1) Natural laws do not imply any new information except obviously the possible information in the original setting of the laws themselves: that is an aspect of the fine tuning argument, which is very different, and separated, form the main ID argument of information in biology. Once a necessity mechanism is detailed, there is no addede information there. The mechanism has to follow the laws. Functional information can be superimposed on contingent structures, not on necessary ones. The information in biology is very definite: it is digital, and it is functional. Necesiity has no role in creating it, except for the possible role of the NS model. NS is a model of necessity, but it is a model which has never been detailed, except for very trivial microevolutionary events. And I am not only saying that it has not been proved: I am saying that no real molelular model has ever been detailed, for any important protein transition, for example, where the role of NS is substantiated. 2) I will not answer your weasel argument becasue, like you, I am not a mathemathician. But I will only say that I don't agree with your concept. If you have to get the final phrase, you have to know the solution for eaxh search, and the correct relationship netween those solutions. That equals, obviously, knowing the searched phrase. In other words, you can get the weasel phrase only of you already know it. But then why search for it? It's not the same if you have to look for one protein with one function you need. Take protein engineers, for instance. They know what they want (the function), but they do not know how to achieve it (they don't know the sequence which will have that function, notwithstanding all the knoweledge we have about proteins). So, they utilize partial random search, and function measuring for selection. And still, they have to work a lot to get some results. And function measuring is much more sensitive than just NS (selecting a fucntion only when it is strong enough to provide, by itself, a reproductive advantage). 3) Always to get back to biology, in evolution there is IMO no search model: just replicators and an environment, and random variation. NS is only a consequence of the interaction between the replicator and the environment. But the theory of darwinian evolution implies a search, because it assumes that the growing information in higher beings is derived from the information in lower beings through unguided mechanisms. In other words, darwinian theory interprets a completely neutral system (the replicator and its environment) as a system which can find new information and new intelligent patterns. That's why we speak of a search: because darwinian evolution needs a system which acts as a search. The only problem is that such a system is not really effecting any search, and that's exactly why the theoy does not work. That's why you will never obtain better digital replicators just by random noise in a digital system. The concept does not simply work. Mathematicians can argue about why it does not work, but we do know that it does not work: you just have to try to detail it or put it to test, both in a computer environment or in a biological environments. The only space where that system seems to work is in the mind of darwinists. gpuccio
tribune[27], I wrote "relevant" which is not the same as "related." Don't you read Merriam-Webster? Prof_P.Olofsson
Sal Gal (#24): "I don’t think it’s a particularly well kept secret that the genetic sequences that are prime for evolution into genes coding for new proteins are duplicate genes and pseudogenes." It is not a particularly well kept secret that darwinists do think that way. They have to use the hypothesis of duplication as a first step because that helps to keep the previous function at the level of the originary gene. That seems useful, if you are supposing that unguided evolution goes from protein A with function A1 to protein B with function B1. But, at the same time, operating on a duplicate gene implies that negative selection cannot any more act. "There is huge inconsistency among IDists on the matter of functionality of DNA sequences. They often push the point that most or all of the genome is designed, and that we simply have a great deal to learn about what non-coding sequences do. But when they want to portray evolution as utterly improbable, they say that a sequence of bases is categorically fit if it codes for a prespecified protein, and is categorically unfit otherwise. Now which is it? Can a non-coding sequence contribute to fitness, or not?" I understand that you are not a biologist, but please, let us go back to real examples. Protein coding genes are only 1.5% of the human genome. The non coding DNA is almost certainly functional, and its functions are almost certainly regulatory. But we still understand poorly what those functions are and how they work. There is absolutely no inconsistency in ID about that matter. We do believe that non coding DNA is functional (at least a graet part of it; personally, I belive it is almost completely functional). And we do believe that its functions are regulatory. But, when we "want to portray evolution as utterly improbable", we do choose the model of protein coding genes, because that's the model we know about, and that's the model on which darwinist theory has been built. In other words, we do not deal with the information in non coding DNA for exactly the same reason why darwinist do not deal with it: because we don't know where it is, and how it is encoded. But where is the inconsistency? Let's say that we are dealing with how that 1.5% of the genome which codes for proteins was generated. That is more than enough to prove that darwinian evolution is "utterly improbable". In other words, in tha 1.5% we have information fro about 20,000 proteins (at least), and we have to explain how it emerged. Moreover, a protein coding gene is a protein coding gene. It encodes a protein sequence. And there is only one kind of fitness for a sequence of nucleotides which encodes a protein sequence: the protein must be functional. Or are you suggesting that a protein coding gene evolves first as a regulatory gene, and then is "coopted" as a protein coding gene? Are you sharing the darwinist folly to that point? In the end, the problem is simple: we have two proteins, A and B, completely different one from the other, and someone says that B is derived from A, or that both are derived from a common ancestor. Well, the information in those proteins is digital, and for me that means that any search based on random variation has tp traverse the combinatorial space defined by proteins sequences of approximately that length. That is the problem with random variation an a digital search. And if you say that NS changes the facts, that may be true, but NS is a model of necessity, and you have to build that model, to detail where and how NS can work. Simply hoping that it can work anyway will not do. Regulatory functions are all another question. They are still vastly unknown. But, if we knew more about them, they would certainly be a much greater problem for darwinists, because indeed regulation is a higher level of information. gpuccio
how they are relevant to evolutionary biology. Everything is related to evolutionary biology, Professor. If it wasn't for Darwin, we'd still be walking around in animals skins and rubbing two sticks together to make a fire. Don't you read, Newsweak? Oh, I misspelled it. Horrors. tribune7
As for the relevance of this work to biology, let me remind commenters that Thomas Schneider used his ev program to argue against Behe and for the power of natural selection in biological evolution and that Rob Pennock cited his work on AVIDA likewise to argue against Behe and for evolution (Pennock cited this not in his NATURE article but in his Dover expert witness report). So if you've got a problem with the applicability of the research at the Evolutionary Informatics Lab to real-life biological evolution, take it up with Schneider and Pennock. William Dembski
Glad to hear about your articles Mr. Dembski! You seem to be successfully pulling into the area that neo-Darwinists seem to deem the most important, that is, peer reviewed articles. Once more ID themed papers get into the peer reviewed sections of science, who's to say neo-Darwinism will hold up at all? Especially considering, if I understand the idea behind the conservation of information theory correctly, it will literally be the death-knell of neo-Darwinism! Domoman
Wolpert and Macready stated outright that their work applied to combinatorial optimization. No one in his right mind would model biological evolution as combinatorial optimization. Setting up as a target, say, the set of all length-100 sequences of nucleotides coding for a particular protein is highly unrealistic. Sure, that gives you a combinatorial optimization problem, but to say that any natural process sought for an encoding of a particular length is absurd. I don't think it's a particularly well kept secret that the genetic sequences that are prime for evolution into genes coding for new proteins are duplicate genes and pseudogenes. There is huge inconsistency among IDists on the matter of functionality of DNA sequences. They often push the point that most or all of the genome is designed, and that we simply have a great deal to learn about what non-coding sequences do. But when they want to portray evolution as utterly improbable, they say that a sequence of bases is categorically fit if it codes for a prespecified protein, and is categorically unfit otherwise. Now which is it? Can a non-coding sequence contribute to fitness, or not? Geneticists believe that some pseudogenes serve functions, even though they don't code for proteins. So why should the fitness of a genetic sequence be dichotomous? Why should we not consider that a non-coding sequence may pass through a succession of functions before coding for a protein? Just as the IDists say, who knows what functions we have yet to discover? Sal Gal
Evolution is a blind, unguided process. ID is a goal-driven process. Since ID has it's goal and evolution doesn't care for one, neither are searching for anything. How do you measure the efficiency of a function they're not actually performing? In other words, what do these two papers contribute to the resolution of the ID/evolution argument? Toronto
Getting papers published is always nice so congratulations to the authors! However, I don't see how these 2 papers qualify as"pro-ID" until it is demonstrated how they are relevant to evolutionary biology. Remember that such claims were made for the original paper by Wolpert et al. until it was pointed out that it has assumptions that do not apply to evolutionary biology. Prof_P.Olofsson
For anyone who has read the second paper: To my math-challenged eyes, it appears that the only type of information that their vertical NFLT addresses is what they call "importance sampling" in the first paper. I must be missing something, so I need someone to hold my hand here. I would expect the phrase "search for a search" to refer to a search for an better-than-random search algorithm. But the meta-search space (M(omega) on page 4) is a space of probability distributions, not a space of algorithms. If we're restricting ourselves to stateless search algorithms and no fitness/cost function, then the only information we can use is a probability distribution to bias our sampling. But this leaves out common search strategies like genetic algorithms. For example, if the active information I'm given is: - Fitness function f is smooth - Target T is at the maximum of f then that information helps immensely in choosing a search algorithm. But neither of those items of information are probability distributions, so neither of them fit into Marks and Dembski's meta-search space. Again, I'm math-challenged and I know I'm missing something. R0b
CJYman, thanks for the comments. I think we're going to have to talk far more concretely if we want to avoid talking past each other. With regards to weasel, you say:
Which metric are you using to determine when each search locks on a letter? Is this metric possible without any knowledge of the target? IOW, trying simulating the search you propose without imposing problem specific information about when each search is to stop and which search goes in which position. Without knowledge about the final target, your search procedure seems even harder for chance and law to accomplish since now there are two steps instead of only one.
I need to describe my model better. SearchAlgorithm1 sends a guess to Oracle1, and Oracle1 responds with "yes" if the guess is "M", or "no" otherwise. Once SearchAlgorithm1 gets a "yes" back, it stops searching. This is a blind search as Dembski describes it, and it involves no more problem-specific information than any other blind search. Likewise with SearchAlgorithm2, SearchAlgorithm3, etc. Oracle2 says "yes" for the letter "E", Oracle3 for the letter "T", etc. When all searches are done, SearchAlgorithm1 has the letter "M", SearchAlgorithm2 as the letter "E", etc. You might ask, how does each SearchAlgorithm know which Oracle to query? Isn't that problem-specific information? I would respond, how does any search algorithm know what oracle to query? How does it know to query at all? How does it know what search space to sample from? How does it know to stop when the oracle tells it that it found the target? How does it know to not stop before that? All of this information is built into the search model, and Marks and Dembski do not count it as active information. The only info that counts as active info is that which allows the search to perform better than random search. My weasel model performs random searches, so it has no active info, according to Marks and Dembski's definition. I think the same principle applies to your response to the stick-in-the-river example. According to my model of that physical process, the stick finds the target far faster than it would through random sampling. You may think that a different model is more appropriate, and that's exactly my point. Marks and Dembski's framework provides no criteria for determining whether a model is appropriate or not, so the choice is arbitrary. This means that the active information metric is, to some degree, arbitrary when applied to real processes rather than pre-specified search models. Note that the two papers referenced in the OP apply the metric only to pre-specified models. Note also that neither paper makes any attempt to connect the active info concept to intelligence, design, teleology, etc. So there's a gap there that Dembski needs to fill in order to support his claim that the papers are pro-ID. R0b
Pardon me for jumping in, and for not having read any of this post or comments, but I thought this might be an opportune moment to ask a question I've had for a while, on the general subject of "material" and "information." Suppose the course of a creek is altered by a rock that falls into it. Better, suppose that the new course leads the creek into country that is more amenable to creek-dom, and the creek becomes a river. Is the rock "information?" My real point: how is this different from DNA (or RNA -- I don't have the specifics of that science)? A physical substance affecting another physical substance, with extensive consequences in the long run? pubdef
As an aside, active information is caused by the organization of the algorithm (matching of search procedure to search space) and this is why the algorithm can not generate active information any more than it can create itself. Thus, if it can not create itself absent previous foresight, the active information is generated by a higher level search ad infinitum or by a system which can model the future and generate targets (foresighted system). CJYman
Rob (#8): "Now try modeling the search the found our current natural laws. What is the search space? What’s the target? Is active info a property of actual processes, or of the choices we make when we model those processes as searches?" Active info is a property of the organization necessary for the actual processes to produce better than chance performance. Taking this back a step further, if chance and law absent previous foresight will not cause an organization of law and chance to produce active information, then active information is also a property of foresight. So far, no one has shown that active information can be generated absent previous foresight (applying knowledge of the problem to be solved or search space into the behavior/organization of the search algorithm as it relates to the search space). Of course, that obviously adds credibility to the ID Hypothesis. Rob: "Another example: Suppose my target is a lake. A stick dropped in a river tends to find its way to the target without searching every corner of the universe, so we’re talking massive amounts of active info. How does the stick know which way to go? Obviously, the same gravity that determines the location of the lake (at a local minimum) also guides the stick. The target location and the search algorithm are not independent. Is this dependency intelligent?" I don't think your example models the concept of active information correctly, since no one can create an evolutionary algorithm based on whichever principles you are saying are at work. I think the flaw in the example would be in actually calculating for active information -- how may sticks fall into rivers naturally and then what percentage of those sticks would actually make it to any lake. Then we would have to randomly drop sticks in random rivers and see what percentage make it to your target lake compared to any lake. This would seem to be akin to assuming that any collection of background noise and laws will eventually produce life and an evolutionary algorithm and then generate intelligence at better than chance performance. Well, sure you can make that as your hypothesis, but good luck with trying to test that or even coming up with a mathematical model to show that is theoretically possible. Rob: "As far as the weasel searches, if Search1 searches for the first letter, Search2 searches for the second letter, etc. then all letters will be found. Marks and Dembski’s weasel algorithm is nothing more than all of these searches happening in parallel. One model has lots of active info, and the other has none. So again it seems that the active info metric depends heavily on how we choose to model a process." The problem is not merely finding the letters. The problem is first locking the letters in place when they are found and then placing those locked positions in to the correct position to reach the final phrase -- ie: search 1 goes in position a, search 2 goes in position b, etc. Which metric are you using to determine when each search locks on a letter? Is this metric possible without any knowledge of the target? IOW, trying simulating the search you propose without imposing problem specific information about when each search is to stop and which search goes in which position. Without knowledge about the final target, your search procedure seems even harder for chance and law to accomplish since now there are two steps instead of only one. CJYman
Rob (#8): "If you try to model the evolution of the human brain as a search, you’re immediately faced with some choices that seem arbitrary. For instance, what is the target?" As far as I can tell, there is no need for arbitrary measurement. Out of all possible combinations of the chemical constituents to make a brain, the brain is an extremely improbable combination of chemicals. In fact, with enough calculation power, an objective number of probability could be given to the brain. I'm sure for the purposes of debate, no one would object to the that number being unfathomable small, seeing that every scientists acknowledges that it is indeed the most complex system by a far shot within our universe (that we are aware of) -- even compared to a space shuttle which is a pretty darn complex system. If we wouldn't expect as a matter of probability, in the history of our planet, for the constituent chemicals which make up a brain to randomly coalesce into a brain, then we both understand that we need some type of fine tuned laws to allow the brain in the first place and then a ratcheting mechanism based on those same laws to bring us closer to the brain. Thus we have a filtering process which is needed to discover the brain. Now, we both seem to realize that the brain itself won't come together via chance and law absent that filtering process. So the question becomes, can that filtering process and the organization of laws which account for it be discovered via merely an arbitrary collection of laws and initial conditions (chance and law)? This becomes a search for that filter (search process which raises the probability for discovering/generating the brain). SO we still haven't accounted for the ability to find the brain at better than chance performance by saying "evolution did it." We need to still search for that evolutionary algorithm -- the incredibly fine tuned laws and initial conditions of our universe -- those highly improbable physical constants and variables which drive some physicists to marvel that it seems a "superintellect has mokeyed with physics." Rob: "Have we reached it? Maybe we’re stuck in a local optimum. If natural laws prevent us from ever reaching the ultimate target, then they have negative active information." It doesn't matter if we are stuck in a local maximum, it only matters if we have indeed reached one of the targets. Even if we don't reach all of the targets (which we will never know since we are within the program) then that makes no difference for us having reached at least one of the targets. Not knowing all the targets makes no difference in calculating the positive active information for targets which have been reached. I think that a good example for the point that is being made is this: take some background noise and an arbitrary collection of laws (definitely only law and chance absent any previous foresight to organize the laws and initial conditions in any specific way for any specific target) and run a program which causes the noise and laws to interact with each other an see if any active information is generated. Are any specified or pre-specified patterns discovered at better than chance performance? CJYman
Hello Rob, I definitely enjoy discussing these things with you as well. Your style of conversation is a breath of fresh air (no obfuscation, honest responses, and no personal attacks). I apologize if I can't follow this topic for that long, as I am back to work now, but here goes for now. And I apologize for my extreme inability to be brief and to the point. Sometimes, I just think that a little in depth rant is necessary to get the idea across. So I will be posting responses to your #8 in sections. Rob, you state: "It’s interesting that simple natural laws can be more information-rich than human brains, but such is the concept of active information." Yes, the laws themselves are simple mathematical descriptions of regularities, however the relation of laws to each other is not simple and it is in this complex organization of law and initial conditions where the active information "resides." In fact, if you understand the fine tuning argument from physics you will understand that our set of laws and initial conditions (cosmo constant) are part of an extremely small set that would allow life, much less even provide enough time for evolution to occur. Life's existence is teetering on a knifes edge of a combination of laws among a vast majority of possible mathematical laws which wouldn't support any notion of life as an evolving information processing machine. That is what information is about. It is not about the laws as merely descriptions of regularity themselves but about the *organization* of those laws, ie: which laws are being utilized, which initial conditions are being utilized, and what values are the laws set at? That is what ID critics fail to realize. There is objectively more to our universe than law and chance. There are organizations of law and chance. This is known as information and it can be assigned a probability -- its information content in bits. In fact, the organization of natural laws may be so complex that they operate in such a way as to provide a framework for consciousness. The organization of laws themselves may even be so complex as to cause the universe itself to be an intelligent system. If a chess program, "merely" a sufficiently organized collection of logic gates, can be intelligent (model the future and move toward a target of winning the game) then there is no theoretical reason why our universe can't be intelligent in at least that same way. To say that a chess program is only law and chance and thus has no need for previous intelligence is blatant misinformation. It's not necessarily the law and chance involved that requires previous intelligence (law and chance is just what it is), it is the highly specific and improbable organization of law and chance -- the information -- that required previous intelligence (foresight). CJYman
Abstract: Conservation of information theorems indicate that any search algorithm performs on average as well as random search without replacement unless it takes advantage of problem-specific information about the search target or the search-space structure.
The notion of "performance" in the context of NFL is nonstandard. Performance is defined as the quality of a sample of n points in the search space and their associated fitness values. The running time of a search algorithm is entirely ignored. This is reasonable only if the time to evaluate fitness does not vary much from one point in the search space to the next -- e.g., as in typical combinatorial optimization problems. It is obvious that the time required for "fitness evaluation" in biological evolution varies enormously from one type of organism to the next. Evolution is not just a matter of information, but also of time, and extant NFL analyses do not tell us about "performance" in any conventional sense of the term.
To pose the question this way suggests that successful searches do not emerge spontaneously but need themselves to be discovered via a search.
When time is taken into account, generally superior "searches" do emerge quite naturally [paper in review]. Sal Gal
From the CoS paper: "In evolutionary search, a large number offspring is often generated, and the more fit offspring are selected for the next generation. When some offspring are correctly announced as more fit than others, external knowledge is being applied to the search giving rise to active information." Deciding that an individual has achieved the search target is to decide that it is more fit than one that hasn't, all search algorithms require some knowledge of the solution, namely what qualifies as a solution. The fitness function in this case is just binary or 'all or nothing' choice. The problem here is that reducing the evaluation criteria of a GA or other type of hill climbing algorithm to an all or nothing choice converts these algorithms into random searches - If a 'hill climber' can't measure the slope then it is just doing a random walk. Providing an evaluation criteria that is non-binary does not necessarily imply or require knowledge about the search in hand, it just requires that the search space has certain properties in order for the search to be effective. In other words if your fitness landscape is flat with a single pinnacle of fitness then a graduated fitness function is of no help. "A “monkey at a typewriter” is often used to illustrate the viability of random evolutionary search." This is an example of a random search with a binary evaluation criteria (it either is the works of Shakespeare of it isn't) There is no descent with modification and no fitness metric, just a halting condition. It is NOT an evolutionary search. Laminar
Well, I wonder who the NCSE will "Richard Sternberg" (verb) after they are published. William Wallace
On one leveI agree with Dr. Dembski, and it is hard to imagine how anyone possibly could not. He is stating in a more complicated way the following: The probablity of a bitstring is proportional to the number of bits in the smallest program that will generate it. Thats why a "search for a search for search" cannot be more probable than the original search. However, what is kind of amazing to me, is that there is still a very strong commitment to Dualism in Dr. Dembski's work which I think is complicating his whole endeavor. He demands we take away a certain part of a computer algorithm and say, "That's not really part of the search algorithm - that was added into it." And the inevitable source for this essential ingredient divorced from the algorithm proper is of course the miraculous, inscrutable and decidly uncomputational (in Dr. Dembski's mind) marvelous Human Mind. (And please remember to capitalize, out of respect.)
"Computers, despite their speed in performing queries, are completely inadequate for resolving even moderately sized search problems without accurate information to guide them" "Such information does not magically materialize but instead results from the action of the programmer who prescribes how knowledge about the problem gets folded into the search algorithm. " "Recognition of the inability of search algorithms in them selves to create information is very useful" "when the results of an even a moderately sized search problem are too good to be true or otherwise overly impressive, we are faced with one of two inescapable alternatives: The search problem under consideration, as gauged by random search, is not as difficult as it first appears. The other inescapable alternative, for difficult problems, is that problem-specific information has been successfully incorporated into the search." "No Free Lunch Theorems (NFLT’s) [28], [51], [52], show that without prior information the search environment or the target sought, one search strategy is, on average, as good as any other." "Over 50 years ago, Leon Brillouin, a pioneer in information theory, wrote “The [computing] machine does not create any new information, but it performs a very valuable transformation of known information. When Brillouin’s insight is applied to search algorithms that do not employ specific information about the problem being addressed, one finds that no search performs consistently better than any other"
I can't imagine Leon Brillouin thinking his observation was any more than obvious. A computer is an extremely simple device, everyone should understand that (think Turing Machine). But a computer agorithm can be as arbitrarily complicated as needed. And you can't come in and strip part of it out and say, "That's not part of the algorithm! That's the result of a Human Mind!" It is part of the algorithm. And even if it came from a human mind, you haven't shown a human mind isn't an algorithm too. Don't really want to start a huge heated debate though. JT
Well this would be interesting.... except the abstracts show that the papers will constitute a slew of technobabble and idiosyncratic terminology, and layered levels of abstraction, no doubt never even mentioning theism once. Not surprising. NS http://sciencedefeated.wordpress.com/ notedscholar
Is anyone interested in providing how these search algorithms are related to ID and evolution? What are they searching? And how does this relate to biological processes. If anyone can put this in simple language, it would be appreciated. I realize that it has something to do with transforming one DNA string into another and for the new DNA string to generate something functional and useful but while the mathematical rigor may be necessary for these journals what is it in simple terms? Anyone? jerry
CJYman, thank you for your comment.
Basically, the laws and environment which allow the ratcheting filter to discover the brain are just as hard to find by search as the human brain would be.
It's interesting that simple natural laws can be more information-rich than human brains, but such is the concept of active information. If you try to model the evolution of the human brain as a search, you're immediately faced with some choices that seem arbitrary. For instance, what is the target? Have we reached it? Maybe we're stuck in a local optimum. If natural laws prevent us from ever reaching the ultimate target, then they have negative active information. Now try modeling the search the found our current natural laws. What is the search space? What's the target? Is active info a property of actual processes, or of the choices we make when we model those processes as searches? What is the connection between active info and intelligence or design? Consider the following information about an objective function: "It's smooth." (Assume that smoothness is well-defined, eg neighboring nodes differ by no more than 1.) The amount of active information in that simple fact depends on the size of the search space. If active info implies intelligence, then we have arbitrarily high intelligence associated with a simple fact. Another example: Suppose my target is a lake. A stick dropped in a river tends to find its way to the target without searching every corner of the universe, so we're talking massive amounts of active info. How does the stick know which way to go? Obviously, the same gravity that determines the location of the lake (at a local minimum) also guides the stick. The target location and the search algorithm are not independent. Is this dependency intelligent? As far as the weasel searches, if Search1 searches for the first letter, Search2 searches for the second letter, etc. then all letters will be found. Marks and Dembski's weasel algorithm is nothing more than all of these searches happening in parallel. One model has lots of active info, and the other has none. So again it seems that the active info metric depends heavily on how we choose to model a process. I always enjoy talking with you, CJYman. R0b
Hello Rob, The claim is legitimately defended since one of the papers proves that it is just as difficult to find the problem as it is to find a match between landscape and search procedure to find the problem any faster. IOW, if it is extremely improbable that the chemical constituents for a human brain will randomly coalesce to form that functioning human brain, then it is at least as extremely improbable (if not more so) for the evolutionary process (encoded in the environment and laws of nature) to be discovered. Basically, the laws and environment which allow the ratcheting filter to discover the brain are just as hard to find by search as the human brain would be. Information is continually moved back a search. Hence, the fine tuning argument. Furthermore, if the problem were modeled as 28 independent searches, how would you guarantee successfully finding and locking on any one letter and then positioning the results of the independent searches into the correct phrase? Will 28 blind searches truly be able to perform such a feat? CJYman
Can you provide a publication date? critter
A hearty congratulations to Drs. Dembski and Marks. These papers obviously represent quite a bit of work. The claim that these papers are pro-ID will, of course, be immediately challenged. Personally, I don't see how that claim can be legitimately defended. It seems that information shared between the landscape/target and the search algorithm can be accounted for by deterministic processes, or simply by the way that the search is modeled. For example, Marks and Dembski model their understanding of Dawkins's weasel algorithm as a "partitioned" search, but it could be equivalently modeled as 28 independent searches, one for each position in the phrase. In that case, the 28 searches are blind, so there is no active information involved. Regardless, I wish the EvoInfo Lab all the best in its research. R0b
Great, I've been waiting for this paper for a long time. bornagain77
Congratulations to both Dr. Dembski and R. J. Marks. This is brilliant stuff. skynetx
Well Congradulations. That is good news. BTW are there 2 Marks (I and II?). It seems like math/engineering journals, perhaps being more grounded in the real world, are willing to follow where the evidence leads. Collin
Congratulations Dr Dembski! How soon before the whole Darwinian edifice crumbles? Here is a poem to celebrate There was a young man named Bill Who found science to be quite a thrill But he hated old Charlie Found his "Origins" gnarly And so swiftly moved in for the kill sallyann

Leave a Reply