Uncommon Descent Serving The Intelligent Design Community

NEWS FLASH: Dembski’s CSI caught in the act

Share
Facebook
Twitter
LinkedIn
Flipboard
Print
Email

Dembski’s CSI concept has come under serious question, dispute and suspicion in recent weeks here at UD.

After diligent patrolling the cops announce a bust: acting on some tips from un-named sources,  they have caught the miscreants in the act!

From a comment in the MG smart thread, courtesy Dembski’s  NFL (2007 edn):

___________________

>>NFL as just linked, pp. 144 & 148:

144: “. . . since a universal probability bound of 1 in 10^150 corresponds to a universal complexity bound of 500 bits of information, (T, E) constitutes CSI because T [i.e. “conceptual information,” effectively the target hot zone in the field of possibilities] subsumes E [i.e. “physical information,” effectively the observed event from that field], T is detachable from E, and and T measures at least 500 bits of information . . . ”

148: “The great myth of contemporary evolutionary biology is that the information needed to explain complex biological structures can be purchased without intelligence. My aim throughout this book is to dispel that myth . . . . Eigen and his colleagues must have something else in mind besides information simpliciter when they describe the origin of information as the central problem of biology.

I submit that what they have in mind is specified complexity, or what equivalently we have been calling in this Chapter Complex Specified information or CSI . . . .

Biological specification always refers to function . . . In virtue of their function [a living organism’s subsystems] embody patterns that are objectively given and can be identified independently of the systems that embody them. Hence these systems are specified in the sense required by the complexity-specificity criterion . . . the specification can be cashed out in any number of ways . . . “

Here we see all the suspects together caught in the very act.

Let us line up our suspects:

1: CSI,

2: events from target zones in wider config spaces,

3: joint complexity-specification criteria,

4: 500-bit thresholds of complexity,

5: functionality as a possible objective specification

6: biofunction as specification,

7: origin of CSI as the key problem of both origin of life [Eigen’s focus] and Evolution, origin of body plans and species etc.

8: equivalence of CSI and complex specification.

Rap, rap, rap!

“How do you all plead?”

“Guilty as charged, with explanation your honour. We were all busy trying to address the scientific origin of biological information, on the characteristic of complex functional specificity. We were not trying to impose a right wing theocratic tyranny nor to smuggle creationism in the back door of the schoolroom your honour.”

“Guilty!”

“Throw the book at them!”

CRASH! >>

___________________

So, now we have heard from the horse’s mouth.

What are we to make of it, in light of Orgel’s conceptual definition from 1973 and the recent challenges to CSI raised by MG and others.

That is:

. . . In brief, living organisms are distinguished by their specified complexity. Crystals are usually taken as the prototypes of simple well-specified structures, because they consist of a very large number of identical molecules packed together in a uniform way. Lumps of granite or random mixtures of polymers are examples of structures that are complex but not specified. The crystals fail to qualify as living because they lack complexity; the mixtures of polymers fail to qualify because they lack specificity. [[The Origins of Life (John Wiley, 1973), p. 189.]

And, what about the more complex definition in the 2005 Specification paper by Dembski?

Namely:

define ϕS as . . . the number of patterns for which [agent] S’s semiotic description of them is at least as simple as S’s semiotic description of [a pattern or target zone] T. [26] . . . . where M is the number of semiotic agents [S’s] that within a context of inquiry might also be witnessing events and N is the number of opportunities for such events to happen . . . . [where also] computer scientist Seth Lloyd has shown that 10^120 constitutes the maximal number of bit operations that the known, observable universe could have performed throughout its entire multi-billion year history.[31] . . . [Then] for any context of inquiry in which S might be endeavoring to determine whether an event that conforms to a pattern T happened by chance, M·N will be bounded above by 10^120. We thus define the specified complexity [χ] of T given [chance hypothesis] H [in bits] . . . as  [the negative base-2 log of the conditional probability P(T|H) multiplied by the number of similar cases ϕS(t) and also by the maximum number of binary search-events in our observed universe 10^120]

χ = – log2[10^120 ·ϕS(T)·P(T|H)]  . . . eqn n1

How about this (we are now embarking on an exercise in “open notebook” science):

1 –> 10^120 ~ 2^398

2 –> Following Hartley, we can define Information on a probability metric:

I = – log(p) . . .  eqn n2

3 –> So, we can re-present the Chi-metric:

Chi = – log2(2^398 * D2 * p)  . . .  eqn n3

Chi = Ip – (398 + K2) . . .  eqn n4

4 –> That is, the Dembski CSI Chi-metric is a measure of Information for samples from a target zone T on the presumption of a chance-dominated process, beyond a threshold of at least 398 bits, covering 10^120 possibilities.

5 –> Where also, K2 is a further increment to the threshold that naturally peaks at about 100 further bits. In short VJT’s CSI-lite is an extension and simplification of the Chi-metric. He explains in the just linked (and building on the further linked):

The CSI-lite calculation I’m proposing here doesn’t require any semiotic descriptions, and it’s based on purely physical and quantifiable parameters which are found in natural systems. That should please ID critics. These physical parameters should have known probability distributions. A probability distribution is associated with each and every quantifiable physical parameter that can be used to describe each and every kind of natural system – be it a mica crystal, a piece of granite containing that crystal, a bucket of water, a bacterial flagellum, a flower, or a solar system . . . .

Two conditions need to be met before some feature of a system can be unambiguously ascribed to an intelligent agent: first, the physical parameter being measured has to have a value corresponding to a probability of 10^(-150) or less, and second, the system itself should also be capable of being described very briefly (low Kolmogorov complexity), in a way that either explicitly mentions or implicitly entails the surprisingly improbable value (or range of values) of the physical parameter being measured . . . .

my definition of CSI-lite removes Phi_s(T) from the actual formula and replaces it with a constant figure of 10^30. The requirement for low descriptive complexity still remains, but as an extra condition that must be satisfied before a system can be described as a specification. So Professor Dembski’s formula now becomes:

CSI-lite=-log2[10^120.10^30.P(T|H)]=-log2[10^150.P(T|H)] . . . eqn n1a

. . . .the overall effect of including Phi_s(T) in Professor Dembski’s formulas for a pattern T’s specificity, sigma, and its complex specified information, Chi, is to reduce both of them by a certain number of bits. For the bacterial flagellum, Phi_s(T) is 10^20, which is approximately 2^66, so sigma and Chi are both reduced by 66 bits. My formula makes that 100 bits (as 10^30 is approximately 2^100), so my CSI-lite computation represents a very conservative figure indeed.

Readers should note that although I have removed Dembski’s specification factor Phi_s(T) from my formula for CSI-lite, I have retained it as an additional requirement: in order for a system to be described as a specification, it is not enough for CSI-lite to exceed 1; the system itself must also be capable of being described briefly (low Kolmogorov complexity) in some common language, in a way that either explicitly mentions pattern T, or entails the occurrence of pattern T. (The “common language” requirement is intended to exclude the use of artificial predicates like grue.) . . . .

[As MF has pointed out] the probability p of pattern T occurring at a particular time and place as a result of some unintelligent (so-called “chance”) process should not be multiplied by the total number of trials n during the entire history of the universe. Instead one should use the formula (1–(1-p)^n), where in this case p is P(T|H) and n=10^120. Of course, my CSI-lite formula uses Dembski’s original conservative figure of 10^150, so my corrected formula for CSI-lite now reads as follows:

CSI-lite=-log2(1-(1-P(T|H))^(10^150)) . . . eqn n1b

If P(T|H) is very low, then this formula will be very closely approximated [HT: Giem] by the formula:

CSI-lite=-log2[10^150.P(T|H)]  . . . eqn n1c

6 –> So, the idea of the Dembski metric in the end — debates about peculiarities in derivation notwithstanding — is that if the Hartley-Shannon- derived information measure for items from a hot or target zone in a field of possibilities is beyond 398 – 500 or so bits, it is so deeply isolated that a chance dominated process is maximally unlikely to find it, but of course intelligent agents routinely produce information beyond such a threshold.

7 –> In addition, the only observed cause of information beyond such a threshold is the now proverbial intelligent semiotic agents.

8 –> Even at 398 bits that makes sense as the total number of Planck-time quantum states for the atoms of the solar system [most of which are in the Sun] since its formation does not exceed ~ 10^102, as Abel showed in his 2009 Universal Plausibility Metric paper. The search resources in our solar system just are not there.

9 –> So, we now clearly have a simple but fairly sound context to understand the Dembski result, conceptually and mathematically [cf. more details here]; tracing back to Orgel and onward to Shannon and Hartley. Let’s augment here [Apr 17], on a comment in the MG progress thread:

Shannon measured info-carrying capacity, towards one of his goals: metrics of the carrying capacity of comms channels — as in who was he working for, again?

CSI extended this to meaningfulness/function of info.

And in so doing, observed that this — due to the required specificity — naturally constricts the zone of the space of possibilities actually used, to island[s] of function.

That specificity-complexity criterion links:

I: an explosion of the scope of the config space to accommodate the complexity (as every added bit DOUBLES the set of possible configurations),  to

II: a restriction of the zone, T, of the space used to accommodate the specificity (often to function/be meaningfully structured).

In turn that suggests that we have zones of function that are ever harder for chance based random walks [CBRW’s] to pick up. But intelligence does so much more easily.

Thence, we see that if you have a metric for the information involved that surpasses a threshold beyond which a CBRW is a plausible explanation, then we can confidently infer to design as best explanation.

Voila, we need an info beyond the threshold metric. And, once we have a reasonable estimate of the direct or implied specific and/or functionally specific (especially code based) information in an entity of interest, we have an estimate of or credible substitute for the value of – log2(p(T|H)); especially if the value of information comes from direct inspection of storage capacity and code symbol patterns of use leading to an estimate of relative frequency, we may evaluate average [functionally or otherwise] specific information per symbol used. This is a version of Shannon’s weighted average information per symbol H-metric, H = –  Σ pi * log(pi), which is also known as informational  entropy [there is an arguable link to thermodynamic entropy, cf here)  or uncertainty.

As in (using Chi_500 for VJT’s CSI_lite [UPDATE, July 3: and S for a dummy variable that is 1/0 accordingly as the information in I is empirically or otherwise shown to be specific, i.e. from a narrow target zone T, strongly UNREPRESENTATIVE of the bulk of the distribution of possible configurations, W]):

Chi_500 = Ip*S – 500,  bits beyond the [solar system resources] threshold  . . . eqn n5

Chi_1000 = Ip*S – 1000, bits beyond the observable cosmos, 125 byte/ 143 ASCII character threshold . . . eqn n6

Chi_1024 = Ip*S – 1024, bits beyond a 2^10, 128 byte/147 ASCII character version of the threshold in n6, with a config space of 1.80*10^308 possibilities, not 1.07*10^301 . . . eqn n6a

[UPDATE, July 3: So, if we have a string of 1,000 fair coins, and toss at random, we will by overwhelming probability expect to get a near 50-50 distribution typical of the bulk of the 2^1,000 possibilities W. On the Chi-500 metric, I would be high, 1,000 bits, but S would be 0, so the value for Chi_500 would be – 500, i.e. well within the possibilities of chance.  However, if we came to the same string later and saw that the coins somehow now had the bit pattern of the ASCII codes for the first 143 or so characters of this post, we would have excellent reason to infer that an intelligent designer, using choice contingency, had intelligently reconfigured the coins. that is because, using the same I = 1,000 capacity value, S is now 1, and so Chi_500 = 500 bits beyond the solar system threshold. If the 10^57 or so atoms of our solar system, for its lifespan, were to be converted into coins and tables etc, and tossed at an impossibly fast rate, it would be impossible to sample enough of the possibilities space W to have confidence that something from so unrepresentative a zone T,  could reasonably be explained on chance. So, as long as an intelligent agent capable of choice is possible, choice — i.e. design — would be the rational, best explanation on the sign observed, functionally specific, complex information.]

10 –> Similarly, the work of Durston and colleagues, published in 2007, fits this same general framework. Excerpting:

Consider that there are usually only 20 different amino acids possible per site for proteins, Eqn. (6) can be used to calculate a maximum Fit value/protein amino acid site of 4.32 Fits/site [NB: Log2 (20) = 4.32]. We use the formula log (20) – H(Xf) to calculate the functional information at a site specified by the variable Xf such that Xf corresponds to the aligned amino acids of each sequence with the same molecular function f. The measured FSC for the whole protein is then calculated as the summation of that for all aligned sites. The number of Fits quantifies the degree of algorithmic challenge, in terms of probability [info and probability are closely related], in achieving needed metabolic function. For example, if we find that the Ribosomal S12 protein family has a Fit value of 379, we can use the equations presented thus far to predict that there are about 10^49 different 121-residue sequences that could fall into the Ribsomal S12 family of proteins, resulting in an evolutionary search target of approximately 10^-106 percent of 121-residue sequence space. In general, the higher the Fit value, the more functional information is required to encode the particular function in order to find it in sequence space. A high Fit value for individual sites within a protein indicates sites that require a high degree of functional information. High Fit values may also point to the key structural or binding sites within the overall 3-D structure.

11 –> So, Durston et al are targetting the same goal, but have chosen a different path from the start-point of the Shannon-Hartley log probability metric for information. That is, they use Shannon’s H, the average information per symbol, and address shifts in it from a ground to a functional state on investigation of protein family amino acid sequences. They also do not identify an explicit threshold for degree of complexity. [Added, Apr 18, from comment 11 below:] However, their information values can be integrated with the reduced Chi metric:

Using Durston’s Fits from his Table 1, in the Dembski style metric of bits beyond the threshold, and simply setting the threshold at 500 bits:

RecA: 242 AA, 832 fits, Chi: 332 bits beyond

SecY: 342 AA, 688 fits, Chi: 188 bits beyond

Corona S2: 445 AA, 1285 fits, Chi: 785 bits beyond  . . . results n7

The two metrics are clearly consistent, and Corona S2 would also pass the X metric’s far more stringent threshold right off as a single protein. (Think about the cumulative fits metric for the proteins for a cell . . . )

In short one may use the Durston metric as a good measure of the target zone’s actual encoded information content, which Table 1 also conveniently reduces to bits per symbol so we can see how the redundancy affects the information used across the domains of life to achieve a given protein’s function; not just the raw capacity in storage unit bits [= no.  of  AA’s * 4.32 bits/AA on 20 possibilities, as the chain is not particularly constrained.]

12 –> I guess I should not leave off the simple, brute force X-metric that has been knocking around UD for years.

13 –> The idea is that we can judge information in or reducible to bits, as to whether it is or is not contingent and complex beyond 1,000 bits. If so, C = 1 (and if not C = 0). Similarly, functional specificity can be judged by seeing the effect of disturbing the information by random noise [where codes will be an “obvious” case, as will be key-lock fitting components in a Wicken wiring diagram functionally organised entity based on nodes, arcs and interfaces in a network], to see if we are on an “island of function.” If so, S = 1 (and if not, S = 0).

14 –> We then look at the number of bits used, B — more or less the number of basic yes/no questions needed to specify the configuration [or, to store the data], perhaps adjusted for coding symbol relative frequencies — and form a simple product, X:

X = C * S * B, in functionally specific bits . . . eqn n8.

15 –> This is of course a direct application of the per aspect explanatory filter, (cf. discussion of the rationale for the filter here in the context of Dembski’s “dispensed with” remark) and the value in bits for a large file is the familiar number we commonly see such as a Word Doc of 384 k bits. So, more or less the X-metric is actually quite commonly used with the files we toss around all the time. That also means that on billions of test examples, FSCI in functional bits beyond 1,000 as a threshold of complexity is an empirically reliable sign of intelligent design.

______________

All of this adds up to a conclusion.

Namely, that there is excellent reason to see that:

i: CSI and FSCI are conceptually well defined (and are certainly not “meaningless”),

ii: trace to the work of leading OOL researchers in the 1970’s,

iii: have credible metrics developed on these concepts by inter alia Dembski and Durston, Chiu, Abel and Trevors, metrics that are based on very familiar mathematics for information and related fields, and

iv: are in fact — though this is hotly denied and fought tooth and nail — quite reliable indicators of intelligent cause where we can do a direct cross-check.

In short, the set of challenges recently raised by MG over the past several weeks has collapsed. END

Comments
KF:
For the particular computer sim cases of interest, I contend that we are looking at information transformation not novel creation. A Mandelbrot set has in it no more information than was specified by the inputs put in to start it up and the algorithm.
The very same thought has passed my mind.PaV
April 18, 2011
April
04
Apr
18
18
2011
03:33 PM
3
03
33
PM
PDT
As to biological reality, functional proteins are specified. How do we know? Because if they weren't specified then random protein sequences would exist, and the kinds of shapes and bonding needed for transcription and translation of proteins could not take place. So this is basically a 'given'. Then, knowing that DNA nucleotide bases are equiprobable given known chemistry/quantum numbers, then any nucleotide base has a 0.25 chance of being at any position along the DNA strand. For a sequence of length, 2X = -log_2{(10^150)}, or,log_2 (10^149), CSI would be present. Q.E.D. That is, the event, E, is the specified sequence; the looked for pattern = target, is the very same specified sequence. There is only one way to specify it (this isn't strictly true since some a.a. are not critical, and can be substituted for [with the effect that X must be longer for CSI to be present---but this is nothing for biological systems]), and the chance hypothesis is that the bases are drawn at random, there being 4 such bases to select from. See how easy it is for biological systems?!?!?PaV
April 18, 2011
April
04
Apr
18
18
2011
03:29 PM
3
03
29
PM
PDT
VJT: I contend that the case already in view, with Durston, shows that the material issue is long since answered. Chi can credibly be calculated for biological systems, as we have done so. For the particular computer sim cases of interest, I contend that we are looking at information transformation not novel creation. A Mandelbrot set has in it no more information than was specified by the inputs put in to start it up and the algorithm. But, it would be interesting to see how much info is being output, how fast, on the relevant algorithms. As shown above, we have a quick way to compare info values to a threshold value. GEM of TKIkairosfocus
April 18, 2011
April
04
Apr
18
18
2011
01:38 PM
1
01
38
PM
PDT
PAV: Interesting approach. Of course once we see that Chi = Ip - (398 + K2) . . . we can easily enough show how the results fall short of the threshold of sufficient complexity to be best explained by design not chance and/or blind necessity. GEM of TKIkairosfocus
April 18, 2011
April
04
Apr
18
18
2011
01:30 PM
1
01
30
PM
PDT
Mathgrrl (#37) In response to my earlier post, you write:
You have certainly made the greatest effort and come up with the most interesting and detailed discussions (which I look forward to you continuing in future threads), but even you have had to recognize that Dembski’s formulation of CSI is not calculable.
I'd like to clarify. In my original post on the CSI scanner, I argued that Dembski's CSI was calculable, but not computable. In a subsequent post, I then provided you with a simplified version of CSI, which I christened CSI-lite. CSI-lite is both calculable and computable. I think that kairosfocus' posts at #44, #45 and #47 above meet your requirements for a CSI calculation for the four scenarios you described. But if you aren't satisfied with those answers, then here's a challenge I shall issue to you. Please provide us with a two- or three-page, detailed but completely jargon-free description of the four scenarios you are describing and post it up on UD. No references to other papers by biologists, please. Describe the problems in your own words, as you would to a non-biologist (which is what I am). Then I might be able to help you. I have no intention of undertaking a course in evolutionary algorithms in order to answer your question; I'm afraid I simply don't have the time.vjtorley
April 18, 2011
April
04
Apr
18
18
2011
12:54 PM
12
12
54
PM
PDT
MathGrrl [38]:
What seems clear now is that Dembski’s CSI has not been and cannot be calculated for real biological or digital systems. That means that the claims of ID proponents that CSI, being specified complexity of more than a certain number of bits, is an indicator of intelligent agency are unsupported.
This is stupidity of the highest order. Look, sweetheart, when the ev program produces less than 96 bits of actual information (that's right, we're dealing with 16 sites each six bases long, and perfect positioning doesn't take place). Per Dembski's definition, this doesn't rise to the level of CSI. To then go on and determine the actual "chance hypothesis" serves no usefulness whatsoever. It would be an exercise in masochism, and no more. IOW, all we have to do is to simply ASSUME that all 96 positions have been perfectly specified; i.e., 2 bits of information for each base, then the "chance hypothesis" is that each position is capable of being held by any of four nucleotide bases, with equal probability for each. Then, for a length of 96 bases, P(T|E) [i.e., of a particular "specified" set of sites = T, given a set of sixteen sequences each composed of six bases = E] = 4^96 = 2^-192. The specificity of this complex ensemble is: -log_2{P(T|E)} = -log_2(2^-192) = 192.......which is, of course, well below the standard for CSI. Hence, we can conclude, using Dembski's definition of CSI, that whatever the ev program produced can be explained invoking chance alone. Now, do you want to dispute the above statement? Do you want to assert that what the ev program did---this digital phenomenon---it did using intelligence? Is this your assertion you wish to make?
Intellectual integrity dictates, therefore, that ID proponents stop making those claims.
You say you want to "learn" how to use CSI. Why not be intellectually honest and admit your true purposes? You're the one who, in the face of legitimate answers given to you over and over, keep insisting that we have it wrong here at UD. It's time for you to show some integrity, my dear, and not just talk about it.PaV
April 18, 2011
April
04
Apr
18
18
2011
12:28 PM
12
12
28
PM
PDT
MathGrrl: Because something is difficult to demonstrate, doesn't mean it doesn't exist. I already compared your 'demand' to that of a undergrad telling his professor that unless he can demonstrate that the Law of Conservation of Angular Momentum, did, indeed, apply to the collapse of World Trade Center Buildings, then he would consider the 'Law' unsubstantiated. Would you agree that to "demonstrate this in a rigorous mathematical fashion" would be incredibly difficult, if not downright impossible? Does this invalidate the 'Law'? Yet there are many instance where the 'Law' can be demonstrated. Believe it or not, professors usually use simplifying assumptions. Can you believe that? (a touch of sarcasm here) I've already walked you through an example of how CSI works. Do you dispute that I had done that? No. Instead you continue with your DEMAND that these four scenarios be analyzed.......or else!!! Take the last sentence. Let's analyze the CSI. What's the event? It's: the word pattern give by: "Instead you DEMAND that your four scenarios be analyzed.......or else!!!" Since the 'alphabet' making up this pattern involves capital letters, exclamation marks and periods, it's composed of at least 54 elements. (I could have chosen to use capital letters all the way through; but I didn't for 'effect'). The sentence contains 71 characters (including spaces). So, the chance of getting this particular combination of characters is 54^71, well above the UPB of 10^150. Thus, it is of intelligent origin. [In this case, highly intelligent origin ;-)] Do you dispute this? But, this isn't enough for you, is it? So, how about your first scenario: "A simple gene duplication, without subsequent modification, that increases production of a particular protein from less than X to greater than X. The specification of this scenario is “Produces at least X amount of protein Y.” First, why do you think "Produces at least X amount of protein Y" is a "specification". CSI deals with events. So, please tell us, what is the event. The obvious answer: the event E actually is: "An increase of protein Y above X". Well, we need to know how this "event" came about. It came about because of a duplicated gene (per you scenario). Thus, event E_1 is: "A gene is duplicated.". Obviously, E=E_1. So, let's substitute E_1 as the actual event. Now we have to ask the question: what's involved in gene duplication? Obviously, cellular machinery. How, then, does this cellular machinery work? It takes a sequence of nucleotide bases A,C,G and T, of a given length, L, and it produces an identical sequence of length L. This sequence is then inserted somewhere in the chromosome/s. Now, of course, for this kind of work to be done, intelligence is certainly operative; but this would require an analysis of all of the cellular machinery involved and the regulatory mechanisms by which they function in tandem. But that is not your real concern. Your real concern is determining whether this newly produced sequence represents CSI or not. Well, we know what the event is: E_1. We know that it involves a "specific" nucleotide sequence of length L. Now we ask, per Dembski's CSI, what is the pattern? Well, the pattern is this same "specific" sequence. So this is our target, T, given E_1. Now, what is our chance hypothesis? Since we're dealing with nucleotide bases, and since there are four of them, and because there is no chemical/quantum reasons for any preferred sequencing of these bases, the "chance" of randomly producing such a sequence is 4^L. If L is sufficiently large, then this would put us above the UPB. However, note this: the sequence ISN'T produced randomly. It's produced using particular protein machines that are able to take nucleotide base X, at position Y, and reproduce (faithfully) nucleotide base X at position Y. Therefore, there is a one-to-one mapping taking place between the original, and the duplicated sequence. So, since there is only ONE possibility for a particular base to be at a particular location/position, the probability of this occurring for a sequence of length L, is: P(T|E_1) = 1^L, which is equal to 1. And, the "complexity" of this "chance hypothesis" is, per Dembski's definition of CSI, -log_2{P(T|E_1)}, = -log_2(1)= 0. There is no information gained. Thus, obviously, no CSI. So, please MathGrrl, let's get down to at least 3 scenarios that you're now "demanding". ______________________________________ For onlookers, note this: MathGrrl, if she had diligently tried to understand CSI as presented in Dembski's "No Free Lunch" book, could have worked this all out herself. But, she's disinclined. Why? Because her whole point is to make the assertion that CSI is ill-defined, and hence useless. This is the strange position we're placed in. MathGrrl claims she wants to "learn" how it's used. And then recites scenarios that patently don't contain CSI, e.g., the ev program, and which can easily be worked out using Dembski's formulism. She won't attempt them. And then she demands that we do. Where will this silliness end? And when?PaV
April 18, 2011
April
04
Apr
18
18
2011
11:25 AM
11
11
25
AM
PDT
k@34
I prefer 1,000 bits
I like nice round numbers, so I vote for 1024 bits.Mung
April 18, 2011
April
04
Apr
18
18
2011
11:16 AM
11
11
16
AM
PDT
Joseph:
Then you count and check the specification- i.e. how much variation can it tolerate.
That is pretty close to what Durston et al did. Hence their Fits values are for islands of function. And as we can show from 11 above, that can then be plugged into the Chi metric, in reduced form. Which of course means we have in hand 35 specific indubitably biological values of Chi, by combining Dembski and Durston. GEM of TKI _________________ F/N: For ease of reference, I have promoted the Durston-Dembski Chi values to the original post, point 11 in sequence, by happy coincidence.kairosfocus
April 18, 2011
April
04
Apr
18
18
2011
08:17 AM
8
08
17
AM
PDT
Joseph: MG is so far simply repeating the talking points she has been making all along. I have yet to see any responsiveness to the cogent information and mathematics provided by ID proponents in response to her challenges, with a partial exception being VJT. Who was reduced to exasperation several times, including above. UNTIL AND UNLESS MG IS ABLE TO SHOW WHY THE LOG-REDUCTION DERIVED EQN FOR CHI ABOVE IS WRONG, WE NEED NOT TAKE SUCH TALKING POINTS PUSHED IN THE TEETH OF MANIFEST EVIDENCE OF THEIR FALSITY SERIOUSLY. And, since:
1: By common definition, in an information context - log (pi) = Ii 2: Similarly log (p*q) = log p + log q, and 3: 10^120 ~ 2^398
then we may reduce Chi = - log2(10^120*D2*p) to: Chi = Ip - (398 + K2) where on reasonable grounds, (398 + K2) tends to max out at about 500, as VJT showed. So, the challenge MG has put collapses, as there is nothing exotic in what was just again done. Chi is simply a metric of info beyond a threshold of complexity. We may debate the threshold but the rest is standard high school algebra applied to info theory. (Here is my always linked note's briefing on basic info theory.) Okay, GEM of TKIkairosfocus
April 18, 2011
April
04
Apr
18
18
2011
08:09 AM
8
08
09
AM
PDT
4 possible nucleotides = 2^2 = 2 bits of informtion per nucleotide 64 possible coding codons = 2^6 = 6 bits of information per amino acid (including STOP) Then you count and check the specification- ie how much variation can it tolerate.Joseph
April 18, 2011
April
04
Apr
18
18
2011
08:02 AM
8
08
02
AM
PDT
8: This was followed by a lot of words, but nothing that actually addresses my four scenarios. This in the teeth of what was done in 19 - 20 above. It is so grotesquely false that this is verging on delusional. It seems to me you did not actually look at what was done but simply repeated your favourite dismissive talking points. That is being out of responsible contact with reality on a serious matter where you have gone on to impugn people by making slanderous assertions. 9: If, as seems increasingly likely, you can’t provide a rigorous mathematical definition of CSI as described by Dembski and show how to apply it to my four scenarios, please just say so and we can move on. that AND operator does ineresting work, as both halves have to be separately true for an AND to be so. But it turns out that as has been shown in the OP, there is a "rigorous" definition of Dembski's Chi metric, by dint of carrying out the log, converting - log (p) to Information explicitly (per Shannon-Hartley) and noting that the subtracted values define a threshold of reasonable complexity. That having been done, it is relatively easy to find where novel FSCI has been claimed to have been originated.
i: gene dup -- copying is not origin of info, but the scale of the copy may make chance duplication implausible, so the implied copier entity is a manifestation of FSCI. ii: Ev -- faces the longstanding problem of starting within an island of function and working on processes that transform but do not de novo create info. In any case the search seems to max out at 266 bits was it, and so it is well within the threshold. An exemplary value was calculated. iii: tierra -- same problem in essence. Where a bit value was given, a demonstration calculation was done, but the issue is that this transforms information as opposed to originating it, within a pre-targtetted island of function. iv: steiner -- same problem with logic, and a demo bit value was also calculated on the for argument assumption that this was a genuinely new bit value. Well within the CSI threshold.
So, both sides of the AND are satisfied and the answer you think does not exist is there. 10: some ID proponents seem to consider CSI to be proportional to the length of the genome. Overall, the existence of COPIES implies a mechanism that will carry forth this. In that context it makes sense to take the length including such copies as may exist in a simple brute force calculation. You will also see that the lengths in question in the relevant calculations are usually for protein coding zones, and so whichever copy is used, the FSCI is there. As in, the repeated case 300 AA --> 900 bases --> 1,800 bits, beyond the 1,000 bits threshold. 11: Why would you be suspicious about the gene duplication scenario? It seems like an obvious known mechanism that must be addressed by any CSI definition. Perhaps, because of your consistent brushing aside of cogent answers on the point, as just happened again. ____________ In short, MG's response shows that she is more or less repeating certain talking points regardless of evidence, response and calculations or derivations and associated or resulting definitions in front of her. In particular, she has shown no responsiveness to the derivation that has been on the table for a week -- which, BTW, makes the exchanges with GP moot -- and which is in the OP:
CHI = - log2[10^120 ·phi_S(T)·P(T|H)]. How about this: 1 –> 10^120 ~ 2^398 2 –> Following Hartley, we can define Information on a probability metric: I = – log(p) 3 –> So, we can re-present the Chi-metric: Chi = – log2(2^398 * D2 * p) Chi = Ip – (398 + K2) 4 –> That is, the Dembski CSI Chi-metric is a measure of Information for samples from a target zone T on the presumption of a chance-dominated process, beyond a threshold of at least 398 bits, covering 10^120 possibilities. 5 –> Where also, K2 is a further increment to the threshold that naturally peaks at about 100 further bits. (In short VJT’s CSI-lite is an extension and simplification of the Chi-metric.) 6 –> So, the idea of the Dembski metric in the end — debates about peculiarities in derivation notwithstanding — is that if the Hartley-Shannon- derived information measure for items from a hot or target zone in a field of possibilities is beyond 398 – 500 or so bits, it is so deeply isolated that a chance dominated process is maximally unlikely to find it, but of course intelligent agents routinely produce information beyond such a threshold. 7 –> In addition, the only observed cause of information beyond such a threshold is the now proverbial intelligent semiotic agents. 8 –> Even at 398 bits that makes sense as the total number of Planck-time quantum states for the atoms of the solar system [most of which are in the Sun] since its formation does not exceed ~ 10^102, as Abel showed in his 2009 Universal Plausibility Metric paper. The search resources in our solar system just are not there. 9 –> So, we now clearly have a simple but fairly sound context to understand the Dembski result, conceptually and mathematically [cf. more details here]; tracing back to Orgel and onward to Shannon and Hartley . . .
Sadly revealing and deeply disappointing. GEM of TKIkairosfocus
April 18, 2011
April
04
Apr
18
18
2011
07:56 AM
7
07
56
AM
PDT
MG: I see you have now replied, but -- given the reduction Chi - Ip - 500 bits beyond a threshold -- I find the unresponsiveness of that reply disappointing. It is noteworthy that you do not seem to have registered the significance of the reduction of the Chi metric to a threshold measure in bits beyond a threshold. I will comment on points clipped in order: 1: the fact that no ID proponent can calculate CSI for my scenarios You plainly did not look at the posts at 19 above. Scenario 1, the doubling of a DNA string produced no additional FSCI, but the act of duplication implies a degree of complexity that might have to be hunted down, but would be most likely well beyond 500 bits or 73 bytes of info. Scenarios 2 - 4 were computer sims, and as PAV long since noted Ev was within 300 bits, far below the significance threshold. 266 bits - 500 bits = - 234 bits lacking. Scenario 3, top of no 20, point 14, on the cited bit number I found, and corrected: Chi_tierra = 22 [oops bytes, cf 20 above] – 500 = – 478 Chi_tierra = 22*8 - 500 = -324, below threshold Scenario 4, Steiner has numbers up to 36 bits I find just now: Chi_steiner = 36 - 500 = - 464 IN SHORT NONE OF THE SCENARIOS MEASURES UP TO A LEVEL OF SIGNIFICANCE. This is in addition to the logical problem of having started within an island of function instead of finding it. Since numbers were provided per a calculation on the mathematical meaning of Chi, your declaration to Joseph is false. 2: Claiming that you have, when all of these threads are available to show that you cannot, is not intellectually honest. making false accusations in the teeth of easily accessible evidence to the contrary raises questions about YOUR honesty and civility, MG. 3: you have had to recognize that Dembski’s formulation of CSI is not calculable. In fact, it is easily transformable (once one thinks of of moving forwards instead of backwards) into a form that is calculable once we can have a bits measure on the usual negative log probability metric used since Hartley and Shannon. 4: What I find particularly interesting is that so many ID proponents have blithely claimed that CSI is an unambiguous indicator of the involvement of intelligent agency, despite obviously never having calculated it for any biological system. Another falsehood. Let me simply clip from 11 above, where I converted the Durston metrics of information in fits, to the Dembaki Chi values in bits beyond the threshold for three cases from the table for 35 protein families published in 2007:
PS: Using Durston’s Fits from his Table 1, in the Dembski style metric of bits beyond the threshold, and simply setting the threshold at 500 bits: RecA: 242 AA, 832 fits, Chi: 332 bits beyond SecY: 342 AA, 688 fits, Chi: 188 bits beyond Corona S2 445 AA, 1285 fits, Chi: 785 bits beyond. –> The two metrics are clearly consistent, and Corona S2 would also pass the X metric’s far more stringent threshold right off as a single protein. (Think about the cumulative fits metric for the proteins for a cell . . . ) –> In short I am here using the Durston metric as a good measure of the target zone’s information content, which Table 1 also conveniently reduces to bits per symbol so we can see how the redundancy affects the information used across the domains of life to achieve a given protein’s function; not just the capacity in storage unit bits [= no AA's * 2 4.32 (oops, I plainly had bases in mind there)]
5: What seems clear now is that Dembski’s CSI has not been and cannot be calculated for real biological or digital systems. That means that the claims of ID proponents that CSI, being specified complexity of more than a certain number of bits, is an indicator of intelligent agency are unsupported. Please see the just above, which has ben there in post no 11 which is specifically addressed to you. I take it that 35 protein families are biological systems. Going beyond that, the Dembski metric in transformed form is easily applied to DNA etc. And, all along, the simple brute force X-metric of FSCI was applied, routinely, to biological systems and has been for years; as can be seen from my always linked as just again linked. You were notified of this repeatedly and plainly willfully refused to acknowledge it. 6: Intellectual integrity dictates, therefore, that ID proponents stop making those claims. On the evidence of this post, I think you need to look very seriously in the mirror before casting such false accusations that are without factual foundation again. I repeat: the Durston metric has been integrated with the Chi metric to produce CSI values as just clipped. The Dembski Chi metric in transformed form is -- as just shown -- readily applied to biological systems. The X-metric did all of this on a brute force basis years ago. 7: If you can provide a riogorous mathematical definition of CSI as described by Dembski and show how to apply it to my four scenarios, please do so. Please see the Op and no 19 ff above, with side-lights on no 11 too. I repeat, CSI a la Demsbki is -- by simply applying the log function and the log of a product rule -- Information in standard bit measures beyond a threshold of at least 398 bits, and in praxis up to above 500. None of this is exotic. We may debate how Demsbski got to the threshold values he chose, but those values are reasonable, and have long been shown to be reasonable. The infinite monkeys analysis as extended by Abel shows that indeed something that is that isolated would not be reasonably discoverable by a random walk dominated search on the gamut of our solar system. I prefer a larger threshold, 1,000 bits, as that shows that the whole universe could not sample more than 1 in 10^150 of the config space in question, removing all reasonable doubt. [ . . . ]kairosfocus
April 18, 2011
April
04
Apr
18
18
2011
07:55 AM
7
07
55
AM
PDT
MathGrrl:
The reasons already provided appear to boil down to the fact that no ID proponent can calculate CSI for my scenarios.
MathGrrl, I have provided the references to support my claims about CSI. OTOH you have presented nothing to refute those claims. Also you have been provided with a defnition of CSI that has rigorous mathematical components.
If that were so, you could simply point to the rigorous mathematical definition and show how to apply it to my four scenarios.
I have and I have applied it to one of them.Joseph
April 18, 2011
April
04
Apr
18
18
2011
07:52 AM
7
07
52
AM
PDT
MathGrrl:
The first scenario, gene duplication, is included because some ID proponents seem to consider CSI to be proportional to the length of the genome.
Perhaps to the length of the minimal genome required for a living organism to do what living organisms do. But even then it depends on variational tolerances- the specificity.
The other three come from my academic interests, combined with a discussion with gpuccio on Mark Frank’s blog. In that thread we touched on the importance of historical contingency in CSI calculations, and I want to make sure we cover that here.
And I provided an answer to one of your scenarios.
If you can provide a riogorous mathematical definition of CSI as described by Dembski and show how to apply it to my four scenarios, please do so.
CSI is a specifed subset of Shannon Information. Shannon information has mathematical rigor. Specified information is Shannon information with meaning/ function. Complexity also has a rigorous mathematical component. So I have applied it one of your scenarios. You can do the rest or you can pay me-fund your project- to look into the rest.Joseph
April 18, 2011
April
04
Apr
18
18
2011
07:49 AM
7
07
49
AM
PDT
Mung,
Has anyone considered that there is an underlying logic behind the four questions? Is there a common thread behind them all other than the (I just want to see CSI in action)?
Me! Me! I know the answer! The first scenario, gene duplication, is included because some ID proponents seem to consider CSI to be proportional to the length of the genome. I wanted to find out if I was misunderstanding that position. The other three come from my academic interests, combined with a discussion with gpuccio on Mark Frank's blog. In that thread we touched on the importance of historical contingency in CSI calculations, and I want to make sure we cover that here.
I will say that the first challenge appears to be a thinly veiled attempt to get an admission that information in the genome can increase through a ‘simple’ gene duplication even. IOW, it wasn’t really about CSI at all.
My participation here is solely so that I can understand CSI well enough to be able to test whether or not known evolutionary mechanisms can create it. Why would you be suspicious about the gene duplication scenario? It seems like an obvious known mechanism that must be addressed by any CSI definition.MathGrrl
April 18, 2011
April
04
Apr
18
18
2011
06:52 AM
6
06
52
AM
PDT
kairosfocus,
MG’s four “CSI challenge” scenarios, addressed:
This was followed by a lot of words, but nothing that actually addresses my four scenarios. If, as seems increasingly likely, you can't provide a rigorous mathematical definition of CSI as described by Dembski and show how to apply it to my four scenarios, please just say so and we can move on.MathGrrl
April 18, 2011
April
04
Apr
18
18
2011
06:51 AM
6
06
51
AM
PDT
Mung, I am not interested in jumping to yet another site to discuss this. It's already difficult enough to follow the half-dozen threads that have been started. If you can provide a riogorous mathematical definition of CSI as described by Dembski and show how to apply it to my four scenarios, please do so.MathGrrl
April 18, 2011
April
04
Apr
18
18
2011
06:50 AM
6
06
50
AM
PDT
kairosfocus,
1: you still can’t provide a rigorous mathematical definition of CSI As by now you know or should know, CSI is not a primarily mathematical concept or construct, like say a Riemann integral or a limit.
What seems clear now is that Dembski's CSI has not been and cannot be calculated for real biological or digital systems. That means that the claims of ID proponents that CSI, being specified complexity of more than a certain number of bits, is an indicator of intelligent agency are unsupported. Intellectual integrity dictates, therefore, that ID proponents stop making those claims.MathGrrl
April 18, 2011
April
04
Apr
18
18
2011
06:50 AM
6
06
50
AM
PDT
vjtorley,
“you still can’t provide a rigorous mathematical definition of CSI.” Surely you jest, Mathgrrl. You have already been given a few rigorous mathematical definitions of CSI.
No, I haven't. You have certainly made the greatest effort and come up with the most interesting and detailed discussions (which I look forward to you continuing in future threads), but even you have had to recognize that Dembski's formulation of CSI is not calculable. What I find particularly interesting is that so many ID proponents have blithely claimed that CSI is an unambiguous indicator of the involvement of intelligent agency, despite obviously never having calculated it for any biological system.MathGrrl
April 18, 2011
April
04
Apr
18
18
2011
06:50 AM
6
06
50
AM
PDT
Joseph,
Your 4 examples are bogus for the reasons already provided.
The reasons already provided appear to boil down to the fact that no ID proponent can calculate CSI for my scenarios.
Also you have been provided with a defnition of CSI that has rigorous mathematical components.
If that were so, you could simply point to the rigorous mathematical definition and show how to apply it to my four scenarios. Neither you nor any other ID proponent has been able to do so. Claiming that you have, when all of these threads are available to show that you cannot, is not intellectually honest.
So stop blaming us for your obvious inadequacies. Ya see that is why there are posts dedicated to you. Now grow up…
Ah, more of the civility expected on UD, I see.MathGrrl
April 18, 2011
April
04
Apr
18
18
2011
06:49 AM
6
06
49
AM
PDT
Reminder to MG (and co): This CSI News Flash thread presents: 1: In the OP, a transformation of the Dembski Chi metric that shows that -- per analysis first presented Sunday last in VJT's LGM thread, one week ago and counting . . . -- it in effect measures information in bits beyond a threshold:
Chi = – log2 (10^120 * phi_s(T) * p (T|H) becomes, on transformation and rounding up: Chi = Ip – 500, in bits beyond a threshold of complexity
2: From 19 above, another response to her four challenges, in light of the just above. Given the vigour with which the challenges were issued, and the additional declarations of "meaninglessness" etc, I think a reasonable response is warranted, and silence instead would be quite telling. GEM of TKIkairosfocus
April 18, 2011
April
04
Apr
18
18
2011
06:01 AM
6
06
01
AM
PDT
10 --> So, we observe how N does not cogently address the explanation in UD WAC 30, but acts as though it does not exist; even using language that suggests that the clinging to it is a mere matter of stubbornly refusing to concede defeat. Remember, the WAC 30 has been there, for years, just a click or two away. 11 --> Let's cite, remembering that this has sat there on the record for years:
30] William Dembski “dispensed with” the Explanatory Filter (EF) and thus Intelligent Design cannot work This quote by Dembski is probably what you are referring to:
I’ve pretty much dispensed with the EF. It suggests that chance, necessity, and design are mutually exclusive. They are not. Straight CSI is clearer as a criterion for design detection.
In a nutshell: Bill made a quick off-the-cuff remark using an unfortunately ambiguous phrase that was immediately latched-on to and grossly distorted by Darwinists, who claimed that the “EF does not work” and that “it is a zombie still being pushed by ID proponents despite Bill disavowing it years ago.” But in fact, as the context makes clear – i.e. we are dealing with a real case of “quote-mining” [cf. here vs. here] — the CSI concept is in part based on the properly understood logic of the EF. Just, having gone though the logic, it is easier and “clearer” to then use “straight CSI” as an empirically well-supported, reliable sign of design. In greater detail: The above is the point of Dembski’s clarifying remarks that: “. . . what gets you to the design node in the EF is SC (specified complexity). So working with the EF or SC end up being interchangeable.”[For illustrative instance, contextually responsive ASCII text in English of at least 143 characters is a “reasonably good example” of CSI. How many cases of such text can you cite that were wholly produced by chance and/or necessity without design (which includes the design of Genetic Algorithms and their search targets and/or oracles that broadcast “warmer/cooler”)?] Dembski responded to such latching-on as follows, first acknowledging that he had spoken “off-hand” and then clarifying his position in light of the unfortunate ambiguity of the phrasal verb dispensed with:
In an off-hand comment in a thread on this blog I remarked that I was dispensing with the Explanatory Filter in favor of just going with straight-up specified complexity. On further reflection, I think the Explanatory Filter ranks among the most brilliant inventions of all time (right up there with sliced bread). I’m herewith reinstating it — it will appear, without reservation or hesitation, in all my future work on design detection. [. . . .] I came up with the EF on observing example after example in which people were trying to sift among necessity, chance, and design to come up with the right explanation. The EF is what philosophers of science call a “rational reconstruction” — it takes pre-theoretic ordinary reasoning and attempts to give it logical precision. But what gets you to the design node in the EF is SC (specified complexity). So working with the EF or SC end up being interchangeable. In THE DESIGN OF LIFE (published 2007), I simply go with SC. In UNDERSTANDING INTELLIGENT DESIGN (published 2008), I go back to the EF. I was thinking of just sticking with SC in the future, but with critics crowing about the demise of the EF, I’ll make sure it stays in circulation.
Underlying issue: Now, too, the “rational reconstruction” basis for the EF as it is presented (especially in flowcharts circa 1998) implies that there are facets in the EF that are contextual, intuitive and/or implicit. For instance, even so simple a case as a tumbling die that then settles has necessity (gravity), chance (rolling and tumbling) and design (tossing a die to play a game, and/or the die may be loaded) as possible inputs. So, in applying the EF, we must first isolate relevant aspects of the situation, object or system under study, and apply the EF to each key aspect in turn. Then, we can draw up an overall picture that will show the roles played by chance, necessity and agency. To do that, we may summarize the “in-practice EF” a bit more precisely as: 1] Observe an object, system, event or situation, identifying key aspects. 2] For each such aspect, identify if there is high/low contingency. (If low, seek to identify and characterize the relevant law(s) at work.) 3] For high contingency, identify if there is complexity + specification. (If there is no recognizable independent specification and/or the aspect is insufficiently complex relative to the universal probability bound, chance cannot be ruled out as the dominant factor; and it is the default explanation for high contingency. [Also, one may then try to characterize the relevant probability distribution.]) 4] Where CSI is present, design is inferred as the best current explanation for the relevant aspect; as there is abundant empirical support for that inference. (One may then try to infer the possible purposes, identify candidate designers, and may even reverse-engineer the design (e.g. using TRIZ), etc. [This is one reason why inferring design does not “stop” either scientific investigation or creative invention. Indeed, given their motto “thinking God's thoughts after him,” the founders of modern science were trying to reverse-engineer what they understood to be God's creation.]) 5] On completing the exercise for the set of key aspects, compose an overall explanatory narrative for the object, event, system or situation that incorporates aspects dominated by law-like necessity, chance and design. (Such may include recommendations for onward investigations and/or applications.)
12 --> So, a fairer and truer view of the EF is that it was conceived in response to a pattern of how people infer to design in real life, seeking to model their decision process, which highlightes the role of CSI as a sign of design. On further reflection, it is necessary to emphasise the per aspects view implicitly involved, So, he filter was improved, by elaborating it somewhat, to bring out that per aspects view and onward synthesis, cf here. 13 --> This actually shows the provisional, progressive process of science at work, in this case, design science. Fundamentally correct work is refined, and leads to a deeper understanding and a more sophisticated tool. 14 --> This thread's OP presents a similar case, where Dembski's analysis of the CSI criterion began with a simple statement on complexity beyond a threshold as we see from p. 144 of NFL, then in 2005 was elaborated into a Chi-metric, and on debates over its challenges it has been deduced that the metric is actually a measure of information in bits beyond a threshold of sufficient complexity that RW based processes are not a credible explanation for being found in a highly specific zone of interest:
Chi = - log2 (10^120 * phi_s(T) * p (T|H) becomes, on transformation and rounding up: Chi = Ip - 500, in bits beyond a threshold of complexity
15 --> We may debate the best threshold [I prefer 1,000 bits as that so vastly overwhelms the resources of the cosmos that the point is plain . . . ], or how Dembski estimated his threshold, or the approximations he made in so doing, but the idea of measuring info as negative log probability is a longstanding one, and the further aspect of looking beyond a threshold is premised on a reasonable point that to be in a zone of interest when on a simple RW you would overwhelmingly be expected to be elsewhere, is plainly reasonable. 16 --> H'mm that reminds me of the laws against loitering, that in effect infer that if you are hanging around in a zone of interest, you are not likely to be there by chance, but are "lying in wait" or are "casing the joint." A design inference in law, based on being found in an unusual and significant location, instead of where you would be expected to be "at random." 17 --> So, we can see that what is going on rhetorically is that one can caricature scientific progress on clarifying and improving or elaborating a fundamentally sound approach across time as though it constitutes successive admissions of failure and abandonment of one's core view. 18 --> But, a fair reading of NFL, e.g as cited above will show the core coherence and continuity in the design view, and the process of progressive improvement and elaboration, not abandonment. GEM of TKIkairosfocus
April 18, 2011
April
04
Apr
18
18
2011
02:01 AM
2
02
01
AM
PDT
Talking point: Dembski has abandoned CSI, apart from rehashing earlier claims to get some verbiage for a new book, and has turned to active information . . . . Folks here are such “true believers” [link added, this is a serious piece of accusatory namecalling] that they generated a piteous outcry when Dembski admitted that the explanatory filter was dead . . . CSI did not pan out. Dembski has moved on, but there is a strong disincentive for him to admit that he has. 1 --> Declaring victory and walking away is one way to get out of a sticky situation. So, when one sees the sort of agenda-serving, spin-filled dismissive reconstruction of history as just cited, one's caution-flags should go up. And in this case, for very good reason: a strawman caricature is being presented to claim victory in the face of the OP's demonstration of the collapse of MG's attempt to project the view that CSI was mathematically [and conceptually] meaningless, contradictory tot he earlier use by Orgel etc, inapplicable to the real world, and incoherent. 2 --> Active information, of course, is BASED on the thinking that underlies CSI and addresses how to exceed the performance of random-walk driven search:
a: it was first shown that there is no free informational lunch. b: That is, when an evolutionary algorithm search strategy in a particular situation succeeds above RW, there is a match to the setting that is information rich, or on average the algorithms will be just as bad as RW search. c: Often, this injected active info is lurking in the implications of an intelligently constructed map of the fitness function sitting on the config space and/or in metrics of success that tell searches warmer/colder; allowing approach to oracles. (All of this, BTW, is foreshadowed in the analysis of Dawkins' Weasel and kin in Ch 4 of NFL, accessible at Google Books here.) d: It turns out that the number of possible fitness maps etc cause an exponentiation of the difficulty of search, and it is much harder to randomly find a good fit of fitness map to the config space than to simply do a RW search of the space. Unless, you are intelligent. As Dembski and Marks note:
"Needle-in-the-haystack problems look for small targets in large spaces. In such cases, blind search stands no hope of success. Conservation of information dictates any search technique will work, on average, as well as blind search. Success requires an assisted search. But whence the assistance required for a search to be successful? To pose the question this way suggests that successful searches do not emerge spontaneously but need themselves to be discovered via a search. The question then naturally arises whether such a higher-level “search for a search” is any easier than the original search. We prove two results: (1) The Horizontal No Free Lunch Theorem, which shows that average relative performance of searches never exceeds unassisted or blind searches, and (2) The Vertical No Free Lunch Theorem, which shows that the difficulty of searching for a successful search increases exponentially with respect to the minimum allowable active information being sought." . . . where, active information is defined and contextualised here: "Conservation of information theorems indicate that any search algorithm performs on average as well as random search without replacement unless it takes advantage of problem-specific information about the search target or the search-space structure. [notice, how D & M are building on the conserv of info principles suggested in NFL] Combinatorics shows that even a moderately sized search requires problem-specific information to be successful. Three measures to characterize the information required for successful search are (1) endogenous information, which measures the difficulty of finding a target using random search; (2) exogenous information, which measures the difficulty that remains in finding a target once a search takes advantage of problem-specific information; and (3) active information, which, as the difference between endogenous and exogenous information, measures the contribution of problem-specific information for successfully finding a target. This paper develops a methodology based on these information measures to gauge the effectiveness with which problem-specific information facilitates successful search. It then applies this methodology to various search tools widely used in evolutionary search."
e: Accordingly, as just cited from abstracts [the links point onwards to the full papers] the Dembski-Marks metrics of active info work by measuring the performance improvement over RW based search; on the demonstrated premise that there is no free lunch when looking for a needle in a very large haystack indeed [one that is beyond the RW search capacity of the observed cosmos] and that the cost in effort of searching for a search is proved to be much stiffer than the cost of a direct search. f: So, the talking point that CSI has been abandoned in favour of active information is a complete misrepresentation of the truth. g: unfortunately, it is a clever one, as the math is hard to follow for those not having the relevant background, and it is easy to point to recent papers on active info, and say there you have it, CSI is dead.
3 --> However, N has provided a good parallel of this tactic of spinning to claim a victory, i.e. the case where WmD said indeed that he had "dispensed with" the use of the explanatory filter, and was using straight CSI as the equivalent. 4 --> Underlying this is the explanatory cluster, where cause traces to chance and/or necessity and/or design, each leaving empirically observable signs on various aspects of a process, phenomenon or object -- causes seldom act alone but we can isolate effects of different factors on aspects of what happens or what we see:
a: Mechanical necessity leads to law-like regularities [e.g. dropped heavy objects fall], b: chance leads to statistically distributed contingencies that "map" a probability distribution [a fair die tumbles to read from 1 to 6 with each face uppermost about 1/6 of the time] c: design often leaves the trace of functionally specific, complex information [e.g. text in posts in this thread] d: to identify the differential impact, it is wise to analyse objects and phenomena in terms of aspects, then bring together the overall picture by synthesising an overall causal account. e: to illustrate, this post on your screen is not wholly explicable on the mechanics of LCD or CRT screens, though that is one aspect. f: Similarly, when the post is pulled form the hard drive where this blog post is stored and transferred over the Internet, noise will inevitably impose a statistical scatter on the signal pules. The Internet has mechanisms for recovering from noise corrupted pules. (That scatter is explained on the noise in communication systems) g: But, to explain the text in English, one has to infer to design. h: so all three aspects are involved, and play a role in the overall explanation for what you see on your screen.
5 --> Once you know that sort of step-by-step analysis on aspects of an object or phenomenon, you can often directly take the short-cut of looking for the signs, instead of explicitly ruling out regularities on high contingency, then noting that we are in a peculiar and specific zone of interest that we should not be in if this were just a statistically controlled random walk. (There are some outcomes that are suspiciously unusual . . .) 6 --> As explained here in UD WAC 30, that is a big part of what Dembski meant to say, though in part he was highlighting the problem of the way that simply speaking of the three causal factors can lead to missing the complementary way they work together in a real object or process. 7 --> The explicit introduction of the per aspect view and the following synthesis of the overall causal account addressed this genuine problem wit the explanatory filter as originally presented. 8 --> A balanced view of what happened would acknowledge the above: by focussing on aspects for analysis and then synthesising an overall causal account, the explanatory filter remains a useful context, and it identifies the significance of CSI as a sign of design. 9 --> But,unfortunately, it can be hard to resist the temptation to pounce on words that look like a concession and triumphantly trot them out to claim victory. [ . . . ]kairosfocus
April 18, 2011
April
04
Apr
18
18
2011
02:01 AM
2
02
01
AM
PDT
Mung: It's a pity you are in mod. I Do clip one point from your Kuppers quote:
we must take leave of the idea of being able, one day, to construct intelligent machines that spontaneously generate meaningful information de novo and continually raise its complexity . . .
Since we are plainly derivative info processors and seemingly do innovate info, I think I differ. It is manifestly plausible that something like us is possible. The issue is to figure out how to build a self-moved, truly intelligent entity. Such an entity would be capable of imagination, motivation, common sense, decision and action, not mere algorithm execution. I think the trick to that is a sufficiently broad "common sense" knowledge base that allows for self- and world- modelling and projections with dynamics that reflect chance, necessity and a model of the decisions of others. Such an imaginative capacity then can be fed into actual decision-making and planning. That will require massive parallel-processing power, probably on neural net architectures. I am of course speculating here. How to do it is a challenge but I think one key observation is the point made by Eng Derek Smith in his model: the MIMO loop must have a two-tier controller, one tier handing the processing homework and I/O management, the other supplying a supervisory level. As to getting self-awareness, I have my doubts. I don't think that simply having processing loops and memory are enough, as some seem to be suggesting: spontaneous emergence of consciousness. GEM of TKIkairosfocus
April 18, 2011
April
04
Apr
18
18
2011
12:20 AM
12
12
20
AM
PDT
That may need to be hunted down, but the inference to design on observing a big enough duplicate, is enough to support an inference to design.
Quoting Bernd-Olaf Kuppers from Information and the Nature of Reality:
But if there are no meaning-generating algorithms, then no information can arise de novo. Therefore, to understand a piece of information of a certain complexity, one always requires background information that is at least of the same complexity. This is the sought-after answer to the question of how much information is needed to understand some other information. Ultimately, it implies that there are no "informational perpetual motion machines" that can generate meaningful information out of nothing. At least it is certain that we must take leave of the idea of being able, one day, to construct intelligent machines that spontaneously generate meaningful information de novo and continually raise its complexity.
Mung
April 17, 2011
April
04
Apr
17
17
2011
07:34 AM
7
07
34
AM
PDT
Unfortunately it appears from what others have told me that ARN is not accepting registration for new members. I've tried to keep it updated with the relevant links to posts here at UD but I've probably failed miserably, lol. It doesn't help that I'm under moderation here and have to wait forever for my posts to show up. Maybe it's time for me to open my own forums :).Mung
April 17, 2011
April
04
Apr
17
17
2011
07:22 AM
7
07
22
AM
PDT
Mung: Thanks a million! Your links in 17 to the original challenges are very welcome. Your immediately following response is helpful, but I don't see how to link within the thread at ARN. (Here is a printoff link for the first response, onlookers.) Your objection is quite cogent:
my first observation about these four "challenges" is that none of them has anything to do with a real biological system, and only the first challenge, involving a gene duplication scenario, even comes close. So all the direct and implied claims that the inability to calculate CSI for any real biological system is what the "MathGrrl debate" is about are off base . . .
You then quote PAV:
Patently, none of those four scenarios rises to the level of actual CSI. The case of gene duplication is crystal-clear nonsense, easily dispensed with via recourse to Chaitin-Kolmogorov complexity, which is clearly incorporated into Dembski’s description of “specified complexity”. The case of ev, as I have pointed out to you at least five times, is a bit string 265 bits long, which has two sections “evolving” together, which means, roughly, that if the two sections matched entirely, the real “specified” length would by 265/2, and which produces “specificity” at only a handful of locations, with these locations moving around “randomly” (i.e., noise) once some minimum “specificity” is arrived at (which makes it then “maximum” specificity”). IOW, only about one third of the 265 bits at most are “specified”, and so you have a complexity of 2^88, or 10^30. This is inconsequential nonsense compared to one average sized paragraph in English. Even if Bill Dembski himself did the calculation for any of those scenarios, he would conclude that CSI is not present. So, how does that make ANY of those FOUR SCENARIOS worth five minutes worth of anyone’s attention?
PAV's rebuttal was well merited. As to whether our MG is the one whose CV you put up, I could be wrong, but that does not seem to be borne out by her response when mathematical specifics have been put up. And, the result of a simple transformation of Dembski's expression belie MG's confident assertions that the expression are mathematically ill defined and meaningless. A metric in bits beyond a threshold is perhaps unusual but it is not meaningless and it is arguably a powerful way to identify and measure what is being got at in the concept CSI. And in 23, you are right to identify an agenda. The questions, as I have said from the outset -- notice my for the record remarks, were heavily loaded with question begging agendas of the ilk, "have you stopped beating your wife." I clip:
I will say that the first challenge appears to be a thinly veiled attempt to get an admission that information in the genome can increase through a ‘simple’ gene duplication even. IOW, it wasn’t really about CSI at all.
Indeed. The storage capacity count obviously goes up, but the functionality and specificity of information does not, no more than when you print a copy of a book or get a copy of a program from aDownloads site. You may allow that functionality to be expressed in another way or location, even within the cell, but you have not created or originated new functionally specific complex information that did not previously exist, the material question at stake. There is no search space challenge to do that. So this is a strawman on a red herring as well. But, there is a self-defeating point, as I highlighted in the MG guest post thread and above. Copies don't appear by magic. Book implies printer or at least photocopier, or even scribe. Just so, novel copies of a gene imply a gene replicating mechanism. So, if the size of the copy is beyond the chance could easily or plausibly explain it threshold, its existence implies a copying mechanism, which will be enormously complex. So, the detection of a big enough copy implies the existence of considerable FSCO/I -- and likely it will be irreducibly complex -- to effect the copying process. That may need to be hunted down, but the inference to design on observing a big enough duplicate, is enough to support an inference to design. The computer science cases were addressed above starting at 19. And, indeed, the Hazen et al paper is significant, and the clipped ideas were part of the context of Durston's work. I particularly focus on:
Complex emergent systems of many interacting components, including complex biological systems, have the potential to perform quantifiable functions. Accordingly, we define “functional information,” I(Ex ), as a measure of system complexity . . . Functional information, which we illustrate with letter sequences, artificial life, and biopolymers, thus represents the probability that an arbitrary configuration of a system will achieve a specific function to a specified degree.
Thanks again GEM of TKIkairosfocus
April 17, 2011
April
04
Apr
17
17
2011
04:51 AM
4
04
51
AM
PDT
Talking point: I [N] cannot see any practical benefit in concluding that some phenomenon is due to supernatural intervention, even if it actually is. Science seeks always to explain nature in terms of nature. It does not tell the Truth This is of course based on the mistake of the false contrast: natural vs supernatural, joined to the imposition of evolutionary materialism as a censoring a priori on science. In actual fact, science can and does routinely study the empirical signs that distinguish the natural [= blind chance + mechanical necessity] from the ART-ificial (or intelligent). Just check your food labels in your kitchen for cases in point. Further to this, it is a blunder to impose naturalistic [= materialistic] explanations on science, censoring it from being able to fearlessly pursue the truth based on observed evidence. Instead, science should seek to be:
an unfettered (but ethically, epistemologially and intellectually responsible) progressive pursuit of the provisionally but well warranted knowledge of the truth about our world based on empirical evidence, observations, experiment, logical-mathematical analysis, explanatory modelling and theorising, and uncensored but mutually respectful discussion among the informed.
A more detailed discussion is in the same thread as N's post, at number 279.kairosfocus
April 17, 2011
April
04
Apr
17
17
2011
04:14 AM
4
04
14
AM
PDT
Talking point: ID is trying to reduce science to forensics, i.e. more or less to a cheap detective story whodunit. In fact, the design inference explanatory filter in large part is about the empirically reliable signs that point to causal patterns, and how we may properly infer on best explanation from sign to warranted cause: I: [si] --> O, on W That is, I (an observer) encounter and observe a pattern of signs, and infer an underlying object or objective state of affairs, on a credible warrant. In particular, in this thread, the warrant most in view is that for the empirically supported cause of CSI. On the point that specificity-complexity narrows possible configurations to independently describable, specific and hard- to- find- by- chance- dominated- search target zones in large config spaces, the Dembski inference identifies that we are warranted to conclude design if the embedded or expressed information in an object of investigation is such that:
Chi = Ip - (398 + K2) > 1, in bits beyond a threshold of sufficient complexity. Where, 398 + K2 will range up to 500 bits or so.
This is empirically well warranted [just the global collection of libraries and Internet and the ICTs industry alone constitute billions of successful tests without an exception], and it is supported by the implications of the infinite monkeys analysis. Notice, the inference it NOT to "whodunit" -- a subtle, veiled allusion to the "ID is creationism in a cheap tuxedo" smear dealt with just above -- but instead to that tweredun. To process, not to agent and identity. To use a corrective forensic example: after we have good reason to conclude arson, then we can go hunt us some possible miscreants. But, if there is no good -- empirically well-warranted -- reason to infer to such deliberate and intelligent cause, whodunit is irrelevant, and perhaps even dangerous. Repeat: the demonstrable focus of design theory -- as can easily be seen from the extended definition of ID hosted here at UD -- is on causal patterns and empirical evidence that warrants the conclusion that particular aspects of objects or states of affairs trace to a causal PROCESS involving factors of chance and/or mechanical necessity and/or design. Again: that tweredun, not whodunit. Going beyond that basic corrective, the above talking point is actually a case of the classic playground complaint: "But, he hit BACK first . . ." For, we are here dealing with a subtle turnabout false accusation. It needs to be spelled out, for this tactic works by confusing the onlookers and getting them to focus blame on the one trying to defend himself or correct a problem, instead of the one who started the fight or who hopes to benefit from the continuation of the error:
a: Those who have carried out a silent coup in science and so also b: have succeeded in imposing and institutionalising a censoring, question-begging a priori -- evolutionary materialism -- on science are now c: twisting the demonstrable but generally poorly understood facts d: to suggest that those who are exposing the coup and the need for reform to restore science to a sound basis, e: are coup-plotters and subversive trouble-makers, to be stopped.
Sadly, the tactic often works on those who do not know the real timeline of what happened and who don't know who really threw the first fist. But, in this case, the situation to be corrected is not in doubt: a priori imposition of materialist censorship on science obviously blocks science from seeking the truth about our world, in light of all the facts and all the relevant possible explanations. It is worth pausing to again hear what Lewontin had to say on this, yet again, as it needs to soak in until we fully understand what has been done to us:
To Sagan, as to all but a few other scientists, it is self-evident that the practices of science provide the surest method of putting us in contact with physical reality, and that, in contrast, the demon-haunted world rests on a set of beliefs and behaviors that fail every reasonable test . . . . It is not that the methods and institutions of science somehow compel us to accept a material explanation of the phenomenal world, but, on the contrary, that we are forced by our a priori adherence to material causes to create an apparatus of investigation and a set of concepts that produce material explanations, no matter how counter-intuitive, no matter how mystifying to the uninitiated. Moreover, that materialism is absolute, for we cannot allow a Divine Foot in the door. [[From: “Billions and Billions of Demons,” NYRB, January 9, 1997. ]
That plainly undermines the integrity of science and in the long run could lead to a collapse of the credibility of science. So the rot needs to be stopped, now. Before it is too late. GEM of TKI PS: The rebuttal made to the above talking point, here, brings out more specifics on how science can be restored, in light of the classic definition of the main methods of science by Newton, in his Opticks, Query 31.kairosfocus
April 17, 2011
April
04
Apr
17
17
2011
03:15 AM
3
03
15
AM
PDT
1 4 5 6 7

Leave a Reply