Uncommon Descent Serving The Intelligent Design Community

NEWS FLASH: Dembski’s CSI caught in the act

Share
Facebook
Twitter
LinkedIn
Flipboard
Print
Email

Dembski’s CSI concept has come under serious question, dispute and suspicion in recent weeks here at UD.

After diligent patrolling the cops announce a bust: acting on some tips from un-named sources,  they have caught the miscreants in the act!

From a comment in the MG smart thread, courtesy Dembski’s  NFL (2007 edn):

___________________

>>NFL as just linked, pp. 144 & 148:

144: “. . . since a universal probability bound of 1 in 10^150 corresponds to a universal complexity bound of 500 bits of information, (T, E) constitutes CSI because T [i.e. “conceptual information,” effectively the target hot zone in the field of possibilities] subsumes E [i.e. “physical information,” effectively the observed event from that field], T is detachable from E, and and T measures at least 500 bits of information . . . ”

148: “The great myth of contemporary evolutionary biology is that the information needed to explain complex biological structures can be purchased without intelligence. My aim throughout this book is to dispel that myth . . . . Eigen and his colleagues must have something else in mind besides information simpliciter when they describe the origin of information as the central problem of biology.

I submit that what they have in mind is specified complexity, or what equivalently we have been calling in this Chapter Complex Specified information or CSI . . . .

Biological specification always refers to function . . . In virtue of their function [a living organism’s subsystems] embody patterns that are objectively given and can be identified independently of the systems that embody them. Hence these systems are specified in the sense required by the complexity-specificity criterion . . . the specification can be cashed out in any number of ways . . . “

Here we see all the suspects together caught in the very act.

Let us line up our suspects:

1: CSI,

2: events from target zones in wider config spaces,

3: joint complexity-specification criteria,

4: 500-bit thresholds of complexity,

5: functionality as a possible objective specification

6: biofunction as specification,

7: origin of CSI as the key problem of both origin of life [Eigen’s focus] and Evolution, origin of body plans and species etc.

8: equivalence of CSI and complex specification.

Rap, rap, rap!

“How do you all plead?”

“Guilty as charged, with explanation your honour. We were all busy trying to address the scientific origin of biological information, on the characteristic of complex functional specificity. We were not trying to impose a right wing theocratic tyranny nor to smuggle creationism in the back door of the schoolroom your honour.”

“Guilty!”

“Throw the book at them!”

CRASH! >>

___________________

So, now we have heard from the horse’s mouth.

What are we to make of it, in light of Orgel’s conceptual definition from 1973 and the recent challenges to CSI raised by MG and others.

That is:

. . . In brief, living organisms are distinguished by their specified complexity. Crystals are usually taken as the prototypes of simple well-specified structures, because they consist of a very large number of identical molecules packed together in a uniform way. Lumps of granite or random mixtures of polymers are examples of structures that are complex but not specified. The crystals fail to qualify as living because they lack complexity; the mixtures of polymers fail to qualify because they lack specificity. [[The Origins of Life (John Wiley, 1973), p. 189.]

And, what about the more complex definition in the 2005 Specification paper by Dembski?

Namely:

define ϕS as . . . the number of patterns for which [agent] S’s semiotic description of them is at least as simple as S’s semiotic description of [a pattern or target zone] T. [26] . . . . where M is the number of semiotic agents [S’s] that within a context of inquiry might also be witnessing events and N is the number of opportunities for such events to happen . . . . [where also] computer scientist Seth Lloyd has shown that 10^120 constitutes the maximal number of bit operations that the known, observable universe could have performed throughout its entire multi-billion year history.[31] . . . [Then] for any context of inquiry in which S might be endeavoring to determine whether an event that conforms to a pattern T happened by chance, M·N will be bounded above by 10^120. We thus define the specified complexity [χ] of T given [chance hypothesis] H [in bits] . . . as  [the negative base-2 log of the conditional probability P(T|H) multiplied by the number of similar cases ϕS(t) and also by the maximum number of binary search-events in our observed universe 10^120]

χ = – log2[10^120 ·ϕS(T)·P(T|H)]  . . . eqn n1

How about this (we are now embarking on an exercise in “open notebook” science):

1 –> 10^120 ~ 2^398

2 –> Following Hartley, we can define Information on a probability metric:

I = – log(p) . . .  eqn n2

3 –> So, we can re-present the Chi-metric:

Chi = – log2(2^398 * D2 * p)  . . .  eqn n3

Chi = Ip – (398 + K2) . . .  eqn n4

4 –> That is, the Dembski CSI Chi-metric is a measure of Information for samples from a target zone T on the presumption of a chance-dominated process, beyond a threshold of at least 398 bits, covering 10^120 possibilities.

5 –> Where also, K2 is a further increment to the threshold that naturally peaks at about 100 further bits. In short VJT’s CSI-lite is an extension and simplification of the Chi-metric. He explains in the just linked (and building on the further linked):

The CSI-lite calculation I’m proposing here doesn’t require any semiotic descriptions, and it’s based on purely physical and quantifiable parameters which are found in natural systems. That should please ID critics. These physical parameters should have known probability distributions. A probability distribution is associated with each and every quantifiable physical parameter that can be used to describe each and every kind of natural system – be it a mica crystal, a piece of granite containing that crystal, a bucket of water, a bacterial flagellum, a flower, or a solar system . . . .

Two conditions need to be met before some feature of a system can be unambiguously ascribed to an intelligent agent: first, the physical parameter being measured has to have a value corresponding to a probability of 10^(-150) or less, and second, the system itself should also be capable of being described very briefly (low Kolmogorov complexity), in a way that either explicitly mentions or implicitly entails the surprisingly improbable value (or range of values) of the physical parameter being measured . . . .

my definition of CSI-lite removes Phi_s(T) from the actual formula and replaces it with a constant figure of 10^30. The requirement for low descriptive complexity still remains, but as an extra condition that must be satisfied before a system can be described as a specification. So Professor Dembski’s formula now becomes:

CSI-lite=-log2[10^120.10^30.P(T|H)]=-log2[10^150.P(T|H)] . . . eqn n1a

. . . .the overall effect of including Phi_s(T) in Professor Dembski’s formulas for a pattern T’s specificity, sigma, and its complex specified information, Chi, is to reduce both of them by a certain number of bits. For the bacterial flagellum, Phi_s(T) is 10^20, which is approximately 2^66, so sigma and Chi are both reduced by 66 bits. My formula makes that 100 bits (as 10^30 is approximately 2^100), so my CSI-lite computation represents a very conservative figure indeed.

Readers should note that although I have removed Dembski’s specification factor Phi_s(T) from my formula for CSI-lite, I have retained it as an additional requirement: in order for a system to be described as a specification, it is not enough for CSI-lite to exceed 1; the system itself must also be capable of being described briefly (low Kolmogorov complexity) in some common language, in a way that either explicitly mentions pattern T, or entails the occurrence of pattern T. (The “common language” requirement is intended to exclude the use of artificial predicates like grue.) . . . .

[As MF has pointed out] the probability p of pattern T occurring at a particular time and place as a result of some unintelligent (so-called “chance”) process should not be multiplied by the total number of trials n during the entire history of the universe. Instead one should use the formula (1–(1-p)^n), where in this case p is P(T|H) and n=10^120. Of course, my CSI-lite formula uses Dembski’s original conservative figure of 10^150, so my corrected formula for CSI-lite now reads as follows:

CSI-lite=-log2(1-(1-P(T|H))^(10^150)) . . . eqn n1b

If P(T|H) is very low, then this formula will be very closely approximated [HT: Giem] by the formula:

CSI-lite=-log2[10^150.P(T|H)]  . . . eqn n1c

6 –> So, the idea of the Dembski metric in the end — debates about peculiarities in derivation notwithstanding — is that if the Hartley-Shannon- derived information measure for items from a hot or target zone in a field of possibilities is beyond 398 – 500 or so bits, it is so deeply isolated that a chance dominated process is maximally unlikely to find it, but of course intelligent agents routinely produce information beyond such a threshold.

7 –> In addition, the only observed cause of information beyond such a threshold is the now proverbial intelligent semiotic agents.

8 –> Even at 398 bits that makes sense as the total number of Planck-time quantum states for the atoms of the solar system [most of which are in the Sun] since its formation does not exceed ~ 10^102, as Abel showed in his 2009 Universal Plausibility Metric paper. The search resources in our solar system just are not there.

9 –> So, we now clearly have a simple but fairly sound context to understand the Dembski result, conceptually and mathematically [cf. more details here]; tracing back to Orgel and onward to Shannon and Hartley. Let’s augment here [Apr 17], on a comment in the MG progress thread:

Shannon measured info-carrying capacity, towards one of his goals: metrics of the carrying capacity of comms channels — as in who was he working for, again?

CSI extended this to meaningfulness/function of info.

And in so doing, observed that this — due to the required specificity — naturally constricts the zone of the space of possibilities actually used, to island[s] of function.

That specificity-complexity criterion links:

I: an explosion of the scope of the config space to accommodate the complexity (as every added bit DOUBLES the set of possible configurations),  to

II: a restriction of the zone, T, of the space used to accommodate the specificity (often to function/be meaningfully structured).

In turn that suggests that we have zones of function that are ever harder for chance based random walks [CBRW’s] to pick up. But intelligence does so much more easily.

Thence, we see that if you have a metric for the information involved that surpasses a threshold beyond which a CBRW is a plausible explanation, then we can confidently infer to design as best explanation.

Voila, we need an info beyond the threshold metric. And, once we have a reasonable estimate of the direct or implied specific and/or functionally specific (especially code based) information in an entity of interest, we have an estimate of or credible substitute for the value of – log2(p(T|H)); especially if the value of information comes from direct inspection of storage capacity and code symbol patterns of use leading to an estimate of relative frequency, we may evaluate average [functionally or otherwise] specific information per symbol used. This is a version of Shannon’s weighted average information per symbol H-metric, H = –  Σ pi * log(pi), which is also known as informational  entropy [there is an arguable link to thermodynamic entropy, cf here)  or uncertainty.

As in (using Chi_500 for VJT’s CSI_lite [UPDATE, July 3: and S for a dummy variable that is 1/0 accordingly as the information in I is empirically or otherwise shown to be specific, i.e. from a narrow target zone T, strongly UNREPRESENTATIVE of the bulk of the distribution of possible configurations, W]):

Chi_500 = Ip*S – 500,  bits beyond the [solar system resources] threshold  . . . eqn n5

Chi_1000 = Ip*S – 1000, bits beyond the observable cosmos, 125 byte/ 143 ASCII character threshold . . . eqn n6

Chi_1024 = Ip*S – 1024, bits beyond a 2^10, 128 byte/147 ASCII character version of the threshold in n6, with a config space of 1.80*10^308 possibilities, not 1.07*10^301 . . . eqn n6a

[UPDATE, July 3: So, if we have a string of 1,000 fair coins, and toss at random, we will by overwhelming probability expect to get a near 50-50 distribution typical of the bulk of the 2^1,000 possibilities W. On the Chi-500 metric, I would be high, 1,000 bits, but S would be 0, so the value for Chi_500 would be – 500, i.e. well within the possibilities of chance.  However, if we came to the same string later and saw that the coins somehow now had the bit pattern of the ASCII codes for the first 143 or so characters of this post, we would have excellent reason to infer that an intelligent designer, using choice contingency, had intelligently reconfigured the coins. that is because, using the same I = 1,000 capacity value, S is now 1, and so Chi_500 = 500 bits beyond the solar system threshold. If the 10^57 or so atoms of our solar system, for its lifespan, were to be converted into coins and tables etc, and tossed at an impossibly fast rate, it would be impossible to sample enough of the possibilities space W to have confidence that something from so unrepresentative a zone T,  could reasonably be explained on chance. So, as long as an intelligent agent capable of choice is possible, choice — i.e. design — would be the rational, best explanation on the sign observed, functionally specific, complex information.]

10 –> Similarly, the work of Durston and colleagues, published in 2007, fits this same general framework. Excerpting:

Consider that there are usually only 20 different amino acids possible per site for proteins, Eqn. (6) can be used to calculate a maximum Fit value/protein amino acid site of 4.32 Fits/site [NB: Log2 (20) = 4.32]. We use the formula log (20) – H(Xf) to calculate the functional information at a site specified by the variable Xf such that Xf corresponds to the aligned amino acids of each sequence with the same molecular function f. The measured FSC for the whole protein is then calculated as the summation of that for all aligned sites. The number of Fits quantifies the degree of algorithmic challenge, in terms of probability [info and probability are closely related], in achieving needed metabolic function. For example, if we find that the Ribosomal S12 protein family has a Fit value of 379, we can use the equations presented thus far to predict that there are about 10^49 different 121-residue sequences that could fall into the Ribsomal S12 family of proteins, resulting in an evolutionary search target of approximately 10^-106 percent of 121-residue sequence space. In general, the higher the Fit value, the more functional information is required to encode the particular function in order to find it in sequence space. A high Fit value for individual sites within a protein indicates sites that require a high degree of functional information. High Fit values may also point to the key structural or binding sites within the overall 3-D structure.

11 –> So, Durston et al are targetting the same goal, but have chosen a different path from the start-point of the Shannon-Hartley log probability metric for information. That is, they use Shannon’s H, the average information per symbol, and address shifts in it from a ground to a functional state on investigation of protein family amino acid sequences. They also do not identify an explicit threshold for degree of complexity. [Added, Apr 18, from comment 11 below:] However, their information values can be integrated with the reduced Chi metric:

Using Durston’s Fits from his Table 1, in the Dembski style metric of bits beyond the threshold, and simply setting the threshold at 500 bits:

RecA: 242 AA, 832 fits, Chi: 332 bits beyond

SecY: 342 AA, 688 fits, Chi: 188 bits beyond

Corona S2: 445 AA, 1285 fits, Chi: 785 bits beyond  . . . results n7

The two metrics are clearly consistent, and Corona S2 would also pass the X metric’s far more stringent threshold right off as a single protein. (Think about the cumulative fits metric for the proteins for a cell . . . )

In short one may use the Durston metric as a good measure of the target zone’s actual encoded information content, which Table 1 also conveniently reduces to bits per symbol so we can see how the redundancy affects the information used across the domains of life to achieve a given protein’s function; not just the raw capacity in storage unit bits [= no.  of  AA’s * 4.32 bits/AA on 20 possibilities, as the chain is not particularly constrained.]

12 –> I guess I should not leave off the simple, brute force X-metric that has been knocking around UD for years.

13 –> The idea is that we can judge information in or reducible to bits, as to whether it is or is not contingent and complex beyond 1,000 bits. If so, C = 1 (and if not C = 0). Similarly, functional specificity can be judged by seeing the effect of disturbing the information by random noise [where codes will be an “obvious” case, as will be key-lock fitting components in a Wicken wiring diagram functionally organised entity based on nodes, arcs and interfaces in a network], to see if we are on an “island of function.” If so, S = 1 (and if not, S = 0).

14 –> We then look at the number of bits used, B — more or less the number of basic yes/no questions needed to specify the configuration [or, to store the data], perhaps adjusted for coding symbol relative frequencies — and form a simple product, X:

X = C * S * B, in functionally specific bits . . . eqn n8.

15 –> This is of course a direct application of the per aspect explanatory filter, (cf. discussion of the rationale for the filter here in the context of Dembski’s “dispensed with” remark) and the value in bits for a large file is the familiar number we commonly see such as a Word Doc of 384 k bits. So, more or less the X-metric is actually quite commonly used with the files we toss around all the time. That also means that on billions of test examples, FSCI in functional bits beyond 1,000 as a threshold of complexity is an empirically reliable sign of intelligent design.

______________

All of this adds up to a conclusion.

Namely, that there is excellent reason to see that:

i: CSI and FSCI are conceptually well defined (and are certainly not “meaningless”),

ii: trace to the work of leading OOL researchers in the 1970’s,

iii: have credible metrics developed on these concepts by inter alia Dembski and Durston, Chiu, Abel and Trevors, metrics that are based on very familiar mathematics for information and related fields, and

iv: are in fact — though this is hotly denied and fought tooth and nail — quite reliable indicators of intelligent cause where we can do a direct cross-check.

In short, the set of challenges recently raised by MG over the past several weeks has collapsed. END

Comments
F/N: The further response to Schneider's horse race page is at what is now 87. Sorry. Will correct above shortly.kairosfocus
April 20, 2011
April
04
Apr
20
20
2011
04:39 PM
4
04
39
PM
PDT
Observe MG, 106:
The issue isn’t simple probability calculations, but how those probabilities are determined in the first place. Please explain, in detail and with examples, how you arrive at the numbers you used for my four scenarios. [Of course, there are no details beyond that the reduction was used and the numbers come from her cites, links and the like (with the exception of the cite from PAV, who has come in thread and explained himself]; as cited above . . . so, did MG actually read what she posted and linked so she would know where the typical numbers clipped and plugged in were sourced? This, unfortunately is a direct foreshadowing of what is to follow in this post . . . ]
This, in response to my request of MG that she: "Kindly MATHEMATICALLY address the reduction of the Dembski type metric to the form in the OP" This convinces me that either MG has not bothered to read carefully before firing off dismissive talking points and even insinuations of dishonesty, or else that she is mathematically incompetent. For, the relevant calculations in the reduction are NOT "probability" calculations but direct application of the Hartley-Shannon information theory DEFINITIONS, and standard log of product logarithm calculations (of High School Algebra standard). Let us cite again, from OP:
Chi = – log2[10^120 ·phi_S(T)·P(T|H)] . . . eqn n1 How about this: 1 –> 10^120 ~ 2^398 2 –> Following Hartley, we can define Information on a probability metric: I = – log(p) . . . eqn n2 3 –> So, we can re-present the Chi-metric: Chi = – log2(2^398 * D2 * p) . . . eqn n3 Chi = Ip – (398 + K2) . . . eqn n4 4 –> That is, the Dembski CSI Chi-metric is a measure of Information for samples from a target zone T on the presumption of a chance-dominated process, beyond a threshold of at least 398 bits, covering 10^120 possibilities. 5 –> Where also, K2 is a further increment to the threshold that naturally peaks at about 100 further bits.
a: At step 1, I simply divided lg(10^120) = 120 by lg 2 = 0.3010, to derive the exponent of two more or less equal to 10^120. b: This is based on the rule that logA/logB = log_b (A), in turn derived from the laws of indices and definition of logarithms. c: In step 2, I applied the Harley -- as Shannon did and as has become standard in info theory -- suggestion that we measure information in log units, as is discussed as a 101 intro here in my always linked. d: This is NOT a "simple" probability calculation, it is a conversion of probabilities of symbols into the standard log metric of information, suggested by Demsbki's - log2 lead to the Chi eqn n1. INDEED, THIS IS REALLY JUST THE CITATION OF THE DEFINITION IN THE CONTEXT WHERE THE A POSTERIORI PROBABILITY IS 1, a common enough reduction. e: In step 3, I first re-present eqn n1, in a form more amenable to logarithmic reduction: Chi = – log2(2^398 * D2 * p) . . . eqn n3 f: Then, we use the well known rule derived from the laws of indices [I think Americans call these exponents], where Log(p*q) = log p + log q. So: Chi = Ip – (398 + K2) . . . eqn n4 g: We next use VJT's reduction,that K peaks out in praxis at about 100 bits: Chi = Ip - 500, in bits beyond a reasonable threshold of complexity h: So, once we have a reasonable measure of functionally specific information, Ip, whether probabilistically or by the classic assessment of storage used and/or storage used and code patterns that lead to the actual info content being less than the used capacity [As Durston et al did], we can identify whether or not it is beyond a threshold with a simple subtraction. i: For instance, it is known that there are 20 possible values for each AA in a polypeptide chain, and as Durston observe that shows a raw capacity of 4.32 bits/AA position. Similarly, notoriously DNA bases take 4 values and so are reduced to 2 bits per position. j: In praxis the functioning codes used or the sequences used to effect working proteins show somewhat uneven distributions, and that is what the Durston metric on H ground vs H functional addresses. k: Using his values of information stored in protein families, we deduce the results shown in point 11 the OP:
RecA: 242 AA, 832 fits, Chi: 332 bits beyond SecY: 342 AA, 688 fits, Chi: 188 bits beyond Corona S2: 445 AA, 1285 fits, Chi: 785 bits beyond . . . results n7
l: This both shows that the reduced Chi metric can be applied to biosystems and gives cases where the Chi metric indicates that these protein families are best explained as the result of design. m: I need not elaborate again the search space and isolation of islands of function [NB: proteins come in fold domains that are isolated in AA sequence space, due to very complex functional constraints] reason behind that conclusion. n: Let it suffice that the Chi metric as reduced will prove to be very reliable on empirical tests, where we directly know the cause. this, because this has already been abundantly shown. o: QED ______ But, it is time to render a verdict. A sad one. If MG has failed to read before dismissively remarking so blunderingly as above, she is in no credible position to responsibly make the contemptuous dismissals she has made over and over and over. And that is the better case. If instead she has been posing as deeply understanding the Shannon theory of information, when she cannot correctly interpret the basic reduction above, then she is incompetent and she is speaking as though the has a degree of intellectual authority that she does not have. That is worse, much, much worse than the first case. But, unfortunately, it does not stop there, in 105 she has capped off her remarks by quoting he traditionally reported sotto voce retort of Galileo when he had been forced to recant his theory by the Inquisition:
Eppur si muove. [= It still moves]
This is a snide insinuation of the worst order, of in effect a religiously motivated persecution against established scientific fact. That, in an immediate context where she would IMMEDIATELY go on to demonstrate that she has either failed to attend to basic facts of a simple derivation in information theory, or else that she is incompetent to follow such a derivation. Such irresponsible atmosphere poisoning behaviour is utterly reprehensible,a nd shouldbe stopped and apologised for. No religious inquisition is imposing a kangaroo court or threatening scholars with thumbscrews to get them to recant facts. Instead, we have carried on a SCIENTFIC discussion on well established facts, theory and evidence. We have deduced reasonable results and we have shown them in action. In addition, we have explained, over and over, step by step, just why the results of so-called evolutionary algorithms are perhaps somewhat analogous to micro-evolution but not to body plan origination macroevolution. Only to meet with brusque dismissals and now rude and disrespectful, slanderous insinuations. On the evidence of the simple reduction of eqn n1, we have reason to believe that MG is in no position to render a reasonable judgement on the challenge made to the Schneider type algorithms. For we have reason to believe that she has ether not paid enough attention to see what is being pointed out, or is incompetent to do so. Whichever way it goes, the matter is over on the merits. MG is mouthing dismissive hyperskeptical talking points, not responding responsibly and with due care and attention to a serious matter. So, we again see the force of the point at the foot of the OP:
In short, the set of challenges recently raised by MG over the past several weeks has collapsed.
QED Let us hope that MG will pause, look at what she has done, reflect, and do better in future. Good day, GEM of TKIkairosfocus
April 20, 2011
April
04
Apr
20
20
2011
04:02 PM
4
04
02
PM
PDT
MathGrrl:
I’m interested in learning what ID proponents consider to be the definition of CSI, as described by Dembski.
Shannon information with meaning/ function of a certain complexity. Shannon information- you figure out the number of possibilities for each position and that gives you the number of bits. for example- 4 possible nucleotides = 2^2 = 2 bits per nucleotide 64 coding codons = 2^6 = 6 bits per amino acid. Then you count and figure out the specificity. Or you keep acting like a little baby- your choice.Joseph
April 20, 2011
April
04
Apr
20
20
2011
03:55 PM
3
03
55
PM
PDT
You forgot one main point R0bb- ev is a targeted search, which means it is an irrelevant example. MathGrrl:
This is not correct.
Yes it is and I have provided the peer-reviewed paper that exposes it as such.Joseph
April 20, 2011
April
04
Apr
20
20
2011
03:50 PM
3
03
50
PM
PDT
Hi kairosfocus. Your link to IOSE above reminded me of: http://www.amazon.com/gp/product/0964130408 http://www.genome.com/life-Orgin.htmMung
April 20, 2011
April
04
Apr
20
20
2011
03:47 PM
3
03
47
PM
PDT
MG, You are boringly repetitious. If you'd like to have a discussion, by all means let's do so. Maybe one of us or even both of us can learn something. If you'd care to do something other than repeat the same two phrases over and over here's what I'd like to discuss: 1. Why you think your "specifications" qualify as a specification as that term is used by Dembski. 2. Just what it is you think the ev program actually does. (It's not what you think.)Mung
April 20, 2011
April
04
Apr
20
20
2011
03:42 PM
3
03
42
PM
PDT
MG, 102:
No one there [in my guest thread] was able to present a rigorous mathematical definition of CSI based on Dembski’s description. If you can, please do so and demonstrate how to calculate it for the four scenarios I describe there.
Pardon directness: this is the proverbial stuck record, repeating long since reasonably and credibly answered demands, without any responsiveness to the fact that they have been answered, again and again. Kindly cf the post at the head of this thread and the comments attached to comment no 1. That will give adequate answers to all real questions that reasonable onlookers will have. The onward links from the editorial comments at no 1 will I believe answer to all reasonable questions for all reasonable people. By contrast, no quantity of explanation, analysis or definition will ever suffice for the sufficiently hyperskeptical. There only remains the reductio ad absurdum of exposing that hyperskepticism as unreasonableness itself, in this case backed up with insinuations -- thus subtle accusations -- of dishonesty. And now, attempted denials. This is simply repetition of talking points that were long since answered. In the guest thread, there were reasonable analyses and calculations relative to the set scenarios. there was discussion of limitations on available information to answer others. And since there has been onward analysis culminating in the reduction of the Dembski chi at the head of this thread in the OP. And the four main challenges were answered point by point from my perspective at 19 - 20 above and onward objections were answered at 44 ff above. The first basic problem, though, is that MG refuses to acknowledge that a gene doubling even creates no novel FSCI. Thus the direct novel FSCI in this copied gene is zero. However the mechanism highlighted or implied by the process of copying will almost certainly be well beyond the FSCI threshold, and exhibits a rather specific function. The second, is that MG is unwilling to accept that the evolutionary algorithms she cites start on islands of function and proceeds to an implicit target on the structure of a so-called fitness function that has nicely trend-y behaviour and an artificially set up hill climbing algorithm. This I have dealt with twice above in a couple of days, starting with the comparison of the Mandelbrot set thought exercise, and again here in more elaborate detail on how computers work and how the relevant algorithms work. The real problem is to get to shores of such islands of function in the chemical pre biological world, and again to get to such islands for novel body plans in the biological world, crossing information origination gaps beyond the credible reach of blind chance and mechanical necessity. To cap off, when the cited challenges were scanned for stated values of claimed originated info, and these were plugged into he reduced Dembski metric to show illustrative cases from her challenges, they were dismissed as made up numbers. That is an unworthy false accusation that should be apologised for. Going yet further, the real case of the Durston measures of FSC were fed into the reduced Dembski metric and produced reasonable values of CSI in bits beyond 500 as a threshold of design as most credible explanation. These -- promoted to point 11 in the OP -- have been studiously ignored and/or dismissed. MG also needs to attend to the responses in what is now 82 to the onward set of questions/demands -- there is no end of inquisitorial questions once one is permitted to sit in the seat of a Torquemada who only demands and is not himself or herself accountable. It is to be noted that after several weeks and some direct invitations or even challenges, MG has yet to provide a single serious piece of mathematical analysis here at UD. What has consistently happened instead is the drumbeat repetition of hyperskeptical dismissals without reason, and repetition of demands. Finally, when VJT reasonably requested some details on the evolutionary algorithm cases that would show what was needed to work things out to a further level, he has been brusquely, even rudely dismissed. That, after he has been the consummate gentleman and has provided considerable efforts that reflect many voluntary hours of work. Some of which led to the reduced Chi metric above. Pardon my sharpness, but as of now MG -- sadly -- comes across as simply pushing hyperskeptical talking points and worse accusations, not as a serious partner for discussion. I hope she will turn over a new leaf. GEM of TKIkairosfocus
April 20, 2011
April
04
Apr
20
20
2011
02:12 PM
2
02
12
PM
PDT
MG, 100:
Schneider documents how a simple subset of evolutionary mechanisms can generate arbitrary amounts of Shannon information . . .
Strictly Shannon information is a metric on a signal of average info per symbol: H = - [SUM on i] pi log pi. More loosely, this relates to the Hartley suggested log metric I = - log p. The problem, as already highlighted is that a string of fair coin tosses produces Shannon info on such metrics [actually the highest info per symbol]. This is because such a random string would have in it no redundancy. But if your coin suddenly started to spit out the ASCII text codes for Hamlet, that would be a very different kettle of fish. The issue on complex, meaningfully or functionally specified info is that the ability to make a functional difference to a system or object on being in a particular set of configs is often important. Hence as Joseph keeps on pointing out we must reckon with such meaningfulness or functionality, as both Dembski, Abel et al and Durston et al among others, discuss. More fundamentally, as was pointed out in detail earlier, Schneider's whole approach misses the injection of active, intelligently created information in the whole exercise, which means that the exercise STARTS in an island of function already. From a quote given above, Schneider evidently did not even realise that he was carrying out an artificial not a natural selection process. The issue at the heart of the CSI/FSCI challenge is to arrive at the shores of such islands of function from arbitrary initial points in config spaces. As has been said over and over and over, only to be brushed aside or ignored. I am now of the opinion that many objectors to the FSCI/CSI concept -- astonishingly -- do not really understand the difference between:
1. [Class 1:] An ordered (periodic) and therefore specified arrangement: THE END THE END THE END THE END Example: Nylon, or a crystal . . . . 2. [Class 2:] A complex (aperiodic) unspecified arrangement: AGDCBFE GBCAFED ACEDFBG Example: Random polymers (polypeptides). 3. [Class 3:] A complex (aperiodic) specified arrangement: THIS SEQUENCE OF LETTERS CONTAINS A MESSAGE! Example: DNA, protein.
(For those who came in late, I just clipped an extract from the very first design theory technical work, Thaxton et al in TMLO, ch 8, 1984.) GEM of TKIkairosfocus
April 20, 2011
April
04
Apr
20
20
2011
01:35 PM
1
01
35
PM
PDT
MF: Re 98:
Please provide references to where I have done so . . .
Kindly cf 44 ff above, for my earlier responses; which relate to your 36 ff. A civil person does not raise suggestions of dishonesty without cast iron proof, for -- as common good sense and what in this region is called basic broughtupcy will tell us -- the mere suggestion is itself an accusation that demands a response. As was pointed out in my response from the first. The evasiveness of your response just above is sadly revealing. Please, correct your behaviour. GEM of TKIkairosfocus
April 20, 2011
April
04
Apr
20
20
2011
01:17 PM
1
01
17
PM
PDT
Joseph,
You forgot one main point R0bb- ev is a targeted search, which means it is an irrelevant example.
This is not correct. I strongly recommend that you read Schneider's PhD thesis as well as the ev paper to learn what ev really shows.MathGrrl
April 20, 2011
April
04
Apr
20
20
2011
12:17 PM
12
12
17
PM
PDT
kairosfocus,
Kindly MATHEMATICALLY address the reduction of the Dembski type metric to the form in the OP
The issue isn't simple probability calculations, but how those probabilities are determined in the first place. Please explain, in detail and with examples, how you arrive at the numbers you used for my four scenarios.MathGrrl
April 20, 2011
April
04
Apr
20
20
2011
12:17 PM
12
12
17
PM
PDT
kairosfocus,
If something is on an island of function deeply enough isolated that 10^102 quantum-time states cannot reasonably reach it on a random walk that works with trial and error selection, then if that thing is actually working, THAT IS BECAUSE IT IS NOT TRULY USING SUCH RW + TAE.
Eppur si muove. If your model of reality says something is impossible and someone like Schneider demonstrates that it is possible, the rational option is not to declare that reality is wrong and your model is correct.
More sophisticated GA’s do not load the target(s) EXPLICITLY, but do so implicitly. They have an intelligently designed well-behaved “fitness function” or objective function — one that has to have helpful trends pointing towards desired target zones — that is relevant to whatever you are trying to target that is mapped to the config space for the “genetic code” string or the equivalent; INTELLIGENTLY mapped, too.
You are confusing the simulator with that which is being simulated. GAs like Schneider's ev model (typically a subset of) known evolutionary mechanisms and a simplified version of the natural world. ev is particularly interesting in that it demonstrates the same results that Schneider observed in real biological organisms. That supports the theory that the modeled evolutionary mechanisms are responsible for the observed results. Unless you're claiming that it is impossible in principle to model evolutionary mechanisms, these GAs support the idea that known evolutionary mechanisms are capable of changing the allele frequency in a population such that subsequent generations are better able to reproduce in a particular environment.
Eventually, we find somewhere where changes don’t make for improvement. We are at a target and voila, information out of the thin air of random variation and selection.
That is not an accurate description of ev. There is no target and nothing limits changes in the simulation. That's what makes the results particularly interesting. I strongly recommend reading Schneider's paper.
GA’s overcome the physical limitations of atoms blindly scanning states through chance and blind necessity, because we are able to intelligently organise machines and program them to solve problems, including by hill-climbing within an island of function that after much painstaking effort, we have set up.
GAs model known evolutionary mechanisms. These mechanisms work without intelligent intervention, just as evolutionary mechanisms are observed to work in the real world. Are you suggesting that it is impossible, even in principle, to model evolutionary mechanisms?MathGrrl
April 20, 2011
April
04
Apr
20
20
2011
12:16 PM
12
12
16
PM
PDT
kairosfocus,
Dembski and others have produced quantitative models of what CSI is about, and have in so doing made more or less good enough for govt work metrics.
If that is the case, please demonstrate how to calculate CSI, as described by Dembski, for my four scenarios.
12.2 Provide real evidence for CSI claims 7 –> Willfully vague and hyperskeptical. If you mean that the only empirically known, observed source of CSI and especially FSCI is intelligence, look around you and ask where the web posts, library books, etc came from. Then ask if you have ever seen the like produced by chance plus blind mechanical necessity.
Thus far, I have not seen a rigorous definition of CSI, so your statement is literally meaningless. Please provide a rigorous mathematical definition of CSI and show how to objectively calculate it. The rest of your comment is equally devoid of definitions or calculations.MathGrrl
April 20, 2011
April
04
Apr
20
20
2011
12:15 PM
12
12
15
PM
PDT
kairosfocus,
Now, you know by now that MG has been simply recirculating objections and dismissals, even implying that the numbers I clipped — by way of illustrative example — from the excerpts she gave or that were otherwise accessible wee taken out of the air.
I'm trying to understand how to calculate CSI, as described by Dembski. Nowhere in your voluminous comments have you demonstrated how to do so. You provide no basis for the numbers you use above. Please provide your mathematically rigorous definition of CSI and explain exactly how you arrived at the numbers you used.MathGrrl
April 20, 2011
April
04
Apr
20
20
2011
12:14 PM
12
12
14
PM
PDT
Mung,
Specification: The Pattern That Signifies Intelligence This is the paper referenced in MathGrrl’s original OP. Look at the title. My gosh. I wonder if there’s anything in it about specification as that term is understood and used by Wm. Dembski himself.
That paper was discussed on my guest thread. No one there was able to present a rigorous mathematical definition of CSI based on Dembski's description. If you can, please do so and demonstrate how to calculate it for the four scenarios I describe there.MathGrrl
April 20, 2011
April
04
Apr
20
20
2011
12:14 PM
12
12
14
PM
PDT
Mung,
My participation here is solely so that I can understand CSI well enough to be able to test whether or not known evolutionary mechanisms can create it.
Schneider claims to have created it. Do you doubt him?
It seems that you haven't understood Schneider's summary. Schneider documents how a simple subset of evolutionary mechanisms can generate arbitrary amounts of Shannon information. ID proponents don't seem to accept that as being equivalent to CSI in this case. I'm interested in learning what ID proponents consider to be the definition of CSI, as described by Dembski.MathGrrl
April 20, 2011
April
04
Apr
20
20
2011
12:13 PM
12
12
13
PM
PDT
Mung,
Again, I don’t see where you’re getting your 266 bit value, but Schneider shows how to generate arbitrary amounts of Shannon information via ev.
MG, you have a bad habit of citing things without having read them.
That's an offensive and baseless assertion. If you think I've misrepresented Schneider, please explain exactly how, with reference to the page to which I linked.MathGrrl
April 20, 2011
April
04
Apr
20
20
2011
12:13 PM
12
12
13
PM
PDT
vjtorley,
You have yet to respond to my challenge regarding the four scenarios you describe:
Please provide us with a two- or three-page, detailed but completely jargon-free description of the four scenarios you are describing and post it up on UD. No references to other papers by biologists, please. Describe the problems in your own words, as you would to a non-biologist (which is what I am). Then I might be able to help you.
I’m still waiting.
What do you find confusing about my scenarios? What's wrong with referring to other papers?MathGrrl
April 20, 2011
April
04
Apr
20
20
2011
12:12 PM
12
12
12
PM
PDT
kairosfocus,
You have called people, for no good reason, dishonest
Please provide references to where I have done so.MathGrrl
April 20, 2011
April
04
Apr
20
20
2011
12:12 PM
12
12
12
PM
PDT
Joseph: Dead right. GEM of TKIkairosfocus
April 20, 2011
April
04
Apr
20
20
2011
08:17 AM
8
08
17
AM
PDT
Robb: 1: PaV’s calculation does not measure the same information that Schneider, Dembski, and Marks measure Please follow context. As can be seen from the freshly numbered eqn n5, Dembski's eqn n1 can be reduced to an expression that first measures Hartley-Shannon information [-log p] -- which can be equivalent to the suitable value of Shannon information proper H average bits per symbol or string as required -- is one term in a more complex expression and context. 500 bits taken off (in effect cf discussion) then gives the Chi_500 metric in bits beyond the complexity bound. In this context, the information has to be cross checked for specificity i.e. isolation to a zone of interest in a config space. Marks and Dembski then point out that once we are beyond such a threshold, the best explanation for event E being in zone T is that a quantum of intelligently directed active information has been a part of the cause of event E. Schneider in speaking of Shannon information is speaking to either Ip or at most H. PAV and I have addresses the numbers reported by Schneider as being on the I term, and he in particular has -- in my view quite correctly -- pointed out that the reported performance is below the threshold, especially when we apply a strict interpretation of what being in zone T means. In short, your dismissive remark is imprecise and off target in light of the analysis above. 2: PaV’s repeated objection that ev is irrelevant because it generates at most 96 bits of information doesn’t hold water, because we can run ev multiple times First, PaV is pointing out that the reported result is less impressive than a first glance will show. Second, as I showed above this morning in 87, step by step, every time Schneider runs his horse race, he shows that his whole exercise is a case of intelligently designed, implicitly targetted search. Schneider is evidently unaware that he 500 - 1,000 bit thresholds are not arbitrary, and so he is misled about the performance of his programs that start in an already narrow target zone of a functional program pre-loaded with a hill climbing filter and algorithm working on a fitness map that -- by the very act of constructing it -- has been loaded in with targetting information. 3: Your long response addressed neither of these points Actually, I spoke to the really material point, the one that you have unfortunately still missed as of the post cluster above. 4: You repeatedly argue that random walks, monkeys on typewriters, etc. are impotent, as if anyone is challenging that fact. Every time you address claims that nobody is making, you create more words for your readers to sift through in search of something responsive. Actually, the key point is that everyone nods assent to the search challenge issue then miss the implication: if you are outperforming such, it is because you have found a way to inject active information. The point that we are dealing with isolated islands of function and have to explain first and foremost getting to shores of function has plainly not got through the many shields of deflection and dismissal. And yet, time after time I see the equivalent of : assume you have already arrived at the island of function with a nice trendy fitness metric and this handy hill-climbing algorithm that can ride the trend like an escalator to the top; all duly intelligently designed. Voila, we run it for a few dozen or thousand cycles and hit a peak. There, in an afternoon we have shown that the search resource barrier is leaped in one swoop by magic powers of evolution on natural selection and chance variation. There is actually a clip in 89 above, HT Mung, from Schneider. I am not making this up. 5: Darwinian evolution, whether or not you believe that it actually occurs in nature, is a far cry from a random walk. The question is whether it occurs and to what extent, not the efficacy of random walks. Excuse me, please do not put words in my mouth that do not belong there, I am not the strawman you are projecting. Ecveryone, starting with Blythe, a Creationist, accepts that here is a natural selection effect of differential reproductive success. Everyone, including the much mocked and smeared young earth creationists, accepts that this accounts for certain small scale changes. However, in other cases the issue is more isolation on founder effects and selection that produces breeds or varieties through Mendelian inheritance. A fair amount of micro-evo can be explained on the two. Even the YECs commonly suggest this can go up to genera and families. In recent days I have cited the relationship between the red deer and the US elk, which are cross-breeding [they are interfertile] across species lines -- elk were recently classified as a species -- in New Zealand. But all of this is well within the major body plan level, and is within islands of function. If Schneider et al were content to say they have a model for evo up to that approximate level and maybe a bit beyond, that would not be of any great controversy, with empirical data to back up some and plausible arguments beyond. But the issue is that his is then turned into a claim that we have a theory that on selection of chance non-foresighted variations explains the history of life on the famed iconic branching tree of life model. the ACTUAL evidence, on study of fossil forms is that body plans are quite stable once they appear -- suddenly, and then they disappear or continue to the modern age. This fits with islands of function on irreducibly complex core based Wicken wiring diagrams, backed up by need to code for proteins and regulatory controls to express the plan on the fertilised ovum through embryonic development. It is those body plans which are credibly isolated that raise the body plan level macroevolutionary challenge: the need to use darwinian chance variation random walks and trial and error to hit shores of islands of function that are unexplained and lack empirical support. (If there were the classic tree of life, there would be overwhelming fossil support for it.There is not, as Gould bluntly testified.) If we take the geochronological timeline as more or less giving a broad -- but provisional -- picture [it more or less fits into the scale of the more observationally supportable cosmological timeline, e.g on apparent age of clusters on HR main sequence turnoffs . . . ], natural history may arguably fit an evolutionary pattern. But the darwinian mechanism is not enough to explain origin of body plans. The required information to explain such is only empirically known to come from design. And as for a sufficient designer, any nano-tech lab a few generations beyond Venter could do it. Front loading and adaptation to niches would work. Deliberate viral insertions would work, etc etc etc. Venter and genetic engineering more broadly have provided proof of concept. 6: I asked how H is defined in the chi formula. You didn’t answer directly, but you seem to be saying that we can just define it as a uniform distribution when we’re calculating the specified complexity of biological structures. But I wasn’t asking about H in general, not just for biology. Nope, I am saying something far more radical. H may be of analytical interest but it is operationally irrelevant. We have a world of information technology that allows us to arrive at bit values from many directions, and direct observation of code symbol patterns as can be documented for both proteins [Durston et al] and DNA [others], can be used to refine the flat-random case where necessary. Not that it makes much difference to the practical outcome. So it makes reasonable sense to use the carrying capacity and go with the bigger threshold of 1,000 bits to be conservative. When you load up a USB stick or a CD you do not worry to find out how the codes balance the symbols so you can report the communication bits [IIRC someone calls them shannons, I have seen binits too] vs the simple carrying capacity bits. In the relevant cases we are looking at genomes with 100,000- 300,000+ base pairs or 200 + kbits of carrying capacity; 1,000 bits is 1/2% where for every extra bit you double your config space. Experiments show that when they cut down bacterial genomes below 300 k bases, they tend to get disintegration of life function. For 200 k bits, config space is 9.98*10^60,205. 7: Given a hypothesis of a random walk, it’s extremely unlikely that the earth’s orbit would be elliptical around the sun. An ellipse is easy to describe, thus specified. And to boot, the earth’s trajectory is functional — a random walk trajectory would result in a dead planet. So the earth’s orbit is specified and complex. Therefore designed? Strawman. Astonishing one too, for someone who has been around at UD for years. Earth's orbit as a conical section modified by perturbations, has long been explained on mechanical necessity as in the first node of the filter. This is the equivalent of the simple order of the crystal as Orgel described. Where things do get complex is the structure and location of he solar system and having a planet like earth in it. From extrasolar systems, the commonest thing is to have disruptions from hot jupiters, some of which are on very eccentric orbits. A nice neat undisturbed system is rare. A moon like ours relative to our planet is a rarity too it seems. And a lot more. Please don't caricature the actual observations on the privileged nature of our solar system. Please view here. 8: you can come up with ad hoc reasons for denying that the earth’s orbit is complex, or specified, or both. Thus the call for an established rigorous definition, so we can come to agreement. As you well know the trichotomy chance necessity and art is NOT ad hoc. It is in fact foundational to how we design and implement experimental exercises, especially when we have to look at plots, treatments, control groups etc and address as well the inevitable statistical scatter. I suggest you take some time and read and reflect on this ID foundations series post. _________ Shakin' me head . . . GEM of TKIkairosfocus
April 20, 2011
April
04
Apr
20
20
2011
08:08 AM
8
08
08
AM
PDT
You forgot one main point R0bb- ev is a targeted search, which means it is an irrelevant example.Joseph
April 20, 2011
April
04
Apr
20
20
2011
06:38 AM
6
06
38
AM
PDT
kairosfocus, WRT random walks: You repeatedly argue that random walks, monkeys on typewriters, etc. are impotent, as if anyone is challenging that fact. Every time you address claims that nobody is making, you create more words for your readers to sift through in search of something responsive. Darwinian evolution, whether or not you believe that it actually occurs in nature, is a far cry from a random walk. The question is whether it occurs and to what extent, not the efficacy of random walks. I asked how H is defined in the chi formula. You didn't answer directly, but you seem to be saying that we can just define it as a uniform distribution when we're calculating the specified complexity of biological structures. But I wasn't asking about H in general, not just for biology. Given a hypothesis of a random walk, it's extremely unlikely that the earth's orbit would be elliptical around the sun. An ellipse is easy to describe, thus specified. And to boot, the earth's trajectory is functional -- a random walk trajectory would result in a dead planet. So the earth's orbit is specified and complex. Therefore designed? Of course, you can come up with ad hoc reasons for denying that the earth's orbit is complex, or specified, or both. Thus the call for an established rigorous definition, so we can come to agreement.R0bb
April 20, 2011
April
04
Apr
20
20
2011
06:37 AM
6
06
37
AM
PDT
kairosfocus, WRT random walks: You repeatedly argue that random walks, monkeys on typewriters, etc. are impotent, as if anyone is challenging that fact. Every time you address claims that nobody is making, you create more words for your readers to sift through in search of something responsive. Darwinian evolution, whether or not you believe that it actually occurs in nature, is a far cry from a random walk. The question is whether it occurs and to what extent, not the efficacy of random walks. I asked how H is defined in the chi formula. You didn't answer directly, but you seem to be saying that we can just define it as a uniform distribution when we're calculating the specified complexity of biological structures. But I wasn't asking about H in general, not just for biology. Given a hypothesis of a random walk, it's extremely unlikely that the earth's orbit would be elliptical around the sun. An ellipse is easy to describe, thus specified. And to boot, the earth's trajectory is functional -- a random walk trajectory would result in a dead planet. So the earth's orbit is specified and complex. Therefore designed? Of course, you can come up with ad hoc reasons for denying that the earth's orbit is complex, or specified, or both. Thus the call for an objective rigorousR0bb
April 20, 2011
April
04
Apr
20
20
2011
06:36 AM
6
06
36
AM
PDT
kairosfocus @ 61:
Pardon, but you seem to be falling into the same question-begging trap that MG has. (I was going to reply to her challenge here, but saw that you provided a live, fresh case in point.) The basic problem with programs like Ev etc...
I made only two points regarding ev. 1) PaV's calculation does not measure the same information that Schneider, Dembski, and Marks measure. 2) PaV's repeated objection that ev is irrelevant because it generates at most 96 bits of information doesn't hold water, because we can run ev multiple times. Your long response addressed neither of these points. Apparently you think I was making some other point.R0bb
April 20, 2011
April
04
Apr
20
20
2011
06:17 AM
6
06
17
AM
PDT
Further counter-challenges to MG and ilk: 4: Kindly MATHEMATICALLY address the reduction of the Dembski type metric to the form in the OP: Chi = - log 2( 10^120* D2* p) Chi = [- log2(p)] - [398 + K2] Ip = - log2(p) Chi_500 = Ip - 500, bits beyond the threshold of sufficient complexity. Chi_1000 = Ip - 1,000 bits beyond. Where it is held that once a reasonable estimate of information in a string or a nodes arcs and interfaces network and/or any codes found therein can be evaluated on ordinary ways to deduce such, Ip can be assigned such a value for relevant purposes. Where it is also held that 500 bits or 10^150 possibilities exceeds the quantum state capacity of the solar system since the big bang, and that similarly 1,000 bits exceeds the capacity of the observed cosmos of some 10^80 atoms. If this derived metric is "meaningless" on being undefined mathematically, kindly show why. It is taken that information is a neg log probability metric, on a flat distribution or a non-flat distribution. the VJT upper limit for phi_S(T) is taken.(It is taken that the Seth Lloyd and quantum state calculations for the solar system and cosmos can be done, as by Demsbki et al or by Abel et al.) 5: Kindly show that the use of the Durston metric values as shown in point 11 OP, in the reduced Dembski metric is meaningless if objected to, where on using these values and Chi_500, we may see for three samples from Durston's table of fits values for protein families carrying out various functions across the domain of cell based life:
Using Durston’s Fits from his Table 1, in the Dembski style metric of bits beyond the threshold, and simply setting the threshold at 500 bits:
RecA: 242 AA, 832 fits, Chi: 332 bits beyond SecY: 342 AA, 688 fits, Chi: 188 bits beyond Corona S2: 445 AA, 1285 fits, Chi: 785 bits beyond.
The two metrics are clearly consistent, and Corona S2 would also pass the X metric’s far more stringent threshold right off as a single protein. (Think about the cumulative fits metric for the proteins for a cell . . . )
6: If you disagree with the claims in the OP and similar places, kindly show why ad how the claims advanced are in material error to the point of being meaningless. 7: Further to this, on the assumption you can SHOW error to the point of meaninglessness [which normally implies a reductio ad absurdum], kindly then demonstrate why putting them forward constitutes DISHONESTY, rather than simple error as is a commonplace in scientific, mathematical and other work. 8: Also, show that it is UNREASONABLE and perhaps dishonest (not just an error, if you can show that) to infer that:
i: CSI as mathematically modelled by Dembski -- starting with NFL p 144 and 148 as cited in OP -- and ii: as reduced to metrics by him and/or others, is iii: materially similar in meaning and context of use to iv: the specified complexity discussed by Orgel [1973] and others using similar terms, and/or v: the complex functional organisation described by Wicken [1979] using terms like wiring diagrams and functional complexity that is information rich.
GEM of TKIkairosfocus
April 20, 2011
April
04
Apr
20
20
2011
05:06 AM
5
05
06
AM
PDT
Counter-Challenge to MG & ilk: 1:Kindly, explain why Schneider (as just shown [HT: Mung], on the issues raised above) was not able to perceive that his process of selection was patently artificial, and that the context of being locked down to an island of intelligently set up function, was also artificial, i.e intelligent. 2: Bonus: kindly explain why Schneider seems to think the plausibility limits discussed by Dembski, Abel and others [all the way back to the thermodynamicists -- cf the Infinite monkeys discussion here] is arbitrary and not rooted in a physical limit to random walks and chance variation in a wide config space that has isolated islands of function in it. 3: Bonus 2: If you disagree with the characterisation of islands of function, show how the general pattern of Wicken wiring diagram functionality based on codes and related nodes, arcs and interfaces organisation, is not like that, i.e that functionally specific complex organisation is not specific. (Remember, text is a string wiring network: s-t-r-i-n-g [where we have codes for meaningful symbols], and that this includes flowcharts, reaction networks, organisation of petroleum refineries, computer motherboards and similar circuits, the exploded view breakdown of multipart mechanical assemblies like a car engine, etc etc.) GEM of TKI PS: MG, I hope you can find it in yourself to face and withdraw the unwarranted accusations/insinuations of deception or dishonesty you have made for some days now in this thread.kairosfocus
April 20, 2011
April
04
Apr
20
20
2011
04:21 AM
4
04
21
AM
PDT
PPS: Schneider inadvertently reveals the depth of his misunderstanding here in his announcement of the "victory" in his horse race:
3 sites to go, 26,300 generations, Rsequence is now at 4.2 bits!! So we have 4.2 bits × 128 sites = 537 bits. We've beaten the so-called "Universal Probability Bound" in an afternoon using natural selection!
NOT The selection and the selection context were both plainly art-ificial, i.e. intelligent.kairosfocus
April 20, 2011
April
04
Apr
20
20
2011
04:09 AM
4
04
09
AM
PDT
PS: Mung's link to the PN discussion of ontogenetic depth is useful. The point Nelson brings out is that the cell types are important but so also is the body plan, on a Wicken wiring diagram that has somehow got to be set up. That is, the path from fertilised ovum to embryo then to developed body plan in the adult [for dozens of major plans and myriads of relatively minor variations on such], is dependent on not only coding for proteins and regulating their expression to release them, but on spatial organisation into a functional whole. The FSCO/I threshold already crossed when we look at coding for proteins, is only the beginning. The development of an embryologically viable individual and then a functioning adult and population that reproduces successfully, multiply the origin of functionally specific complexity challenge posed by life forms.kairosfocus
April 20, 2011
April
04
Apr
20
20
2011
04:05 AM
4
04
05
AM
PDT
Onlookers (& Mung): Mung links the Schneider page where he presents a horse race between his program -- notice, the intelligent designer is here presented [before he ducks behind the curtain to manipulate the puppet strings by the proxy of his algorithms loaded into the computer] -- and the Dembski style threshold for random walk based search. Of course, his program wins the race! Voila, Dembski -- that IDiot -- is exposed as an imbecile or fraud yet again! NOT . . . What Schneider is stubbornly (sorry, but only so strong a word will do, after this much opportunity and pleading to get things right have been willfully dismissed) overlooking -- despite being warned again and again and despite this warning going back as far as Dembski's NFL and even Thaxton et al in TMLO -- in the first instance is the physical basis for the Dembski or Abel type plausibility limit. (He seems to think the limit is arbitrary and can easily be exceeded by blind random chance feeding into an automatic selection process. This is the equivalent -- sorry but the comparison is well merited -- to clinging to a scheme to create a perpetual motion machine after its flaws have repeatedly been exposed.) For a 500-bit limit, we can show that the atoms in our solar system [by far and away most being in the sun], across its timeline since the usual date for the big bang, will have gone though at most 10^102 Planck-time quantum states. Where to get to just a strong force nuclear type interaction we need about 10^20 -- a hundred billion billion Planck times; i.e. we are well beyond the nanosecond or so times for computer cycles. Similarly, the atoms of the solar system could not have undergone as much as 10^120 bit operations in the available time. If something is on an island of function deeply enough isolated that 10^102 quantum-time states cannot reasonably reach it on a random walk that works with trial and error selection, then if that thing is actually working, THAT IS BECAUSE IT IS NOT TRULY USING SUCH RW + TAE. [OOPS, the comment inadvertently posted.] Why am I confident of that? Because I am highly confident in the physics and the underlying statistical thermodynamics reasoning that drives this sort of conclusion. If your search for a needle in a haystack is cursory, you are utterly unlikely to find it except by extremely good luck. And we cannot build a theory of what happens or is held to have happened fairly often within the ambit of our solar system -- namely body plan origination -- on such extraordinary good luck. What then is really happening with ev etc? The clue lies in the fact that ev etc are intelligently designed programs, and in effect that they seek a target iteratively by starting within a island of complex function, i.e an algorithm that was set up to work, and is coded with oodles of information. A second clue is the behaviour of the defining equations for the Mandelbrot set: an apparently simple expression, on being scanned across the complex plane [sort of a special x-y graph where the y values are multiples of the imaginary square root of -1, called i], and fed into an algorithm that detects how long before a given point escapes a bounded zone and colours accordingly, produces INFINITELY DEEP COMPLEXITY. Remember, there is an infinity of points in any line segment or area. We can relatively easily specify an infinitely deep degree of complexity to be brought out by that patient robot proxy for our intelligence, a computer. the real problem here is to tell it where to stop digging in further, and we have to be fairly clever to do so, or it will try to carry out a supertask and get nowhere, thrashing around in endless loops. (That, in effect is what happens when a PC freezes. You have to force a halt by ESC or reset.) This is what genetic algorithms -- that is a sales name -- like ev etc are really doing: ______________ >> When something like Weasel comes along and claims that information is being generated out of the thin air of random number sequences filtered by a selection process, we need to take a closer look. Weasel in fact has a “map” of a space of possible configs, where the total is about 10^40. This is of course within trial and error search capacity of modern PCs, and certainly that of the raw processing power of the cosmos. But what is crucial is that there is a way of measuring distance to target, basically off difference from the target value. This allowed Dawkins to convert the whole space of 10^40 possibilities into a warmer/colder map, without reference to whether or not the “nonsense phrases” had any meaning in themselves. At each stage, a ring of random changes in a seed phrase was made, and this was then compared with the seed itself and the warmest was picked to move to the next step. And so on until in about 40 – 60 generations in the published runs, the phrase sequence converged on the target. VOILA! Information created out of chance variation and selection! NOT. The target was already preloaded all along. In this case, quite explicitly. More sophisticated GA’s do not load the target(s) EXPLICITLY, but do so implicitly. They have an intelligently designed well-behaved “fitness function” or objective function — one that has to have helpful trends pointing towards desired target zones — that is relevant to whatever you are trying to target that is mapped to the config space for the “genetic code” string or the equivalent; INTELLIGENTLY mapped, too. Notice, all the neat little sales words that suggest a parallel to the biological world that is less real than appears. Then, when the seed genome value or config is fed in, it is tested for some figure of merit that looks at closeness to target or to desired performance. A ring of controlled random/pseudo-random samples is tossed out. The one or ones that trend warmest are picked and the process repeats. Eventually, we find somewhere where changes don’t make for improvement. We are at a target and voila, information out of the thin air of random variation and selection. NOT. Again, look at how much intelligently designed work was put in to set up the island of function to wander around in on a nice slope and detect warmer/colder, to move towards a local peak of performance or “fitness.” No nice fitness landscape and no effective progress. No CONTROL on the degree of randomness in the overall system, and chaos would soon dominate. No warmer/colder metric introduced at just the right times on that nice slope function, and you would wander around blindly. In short, the performance is impressive but the prestidigitation’s impact requires us to be distracted from the wires, curtains, smoke puffs, and trap doors under the stage. So, when we see claims that programs like avida, ev, tierra etc produce arbitrary quantities of shannon or even functionally specific information, we need to highlight the intelligently designed apparatus that makes this impressive performance possible. >> ______________ Maybe you doubt me, after all, like Demski I am an "IDiot" and maybe even a suspect closet -- shudder! -- Creationist. (Don't ask my students or clients what they think about whether I know what I am talking about as a rule.) So let us call wiki as an inadvertent witness against interest:
In a genetic algorithm, a population of strings (called chromosomes or the genotype of the genome), which encode candidate solutions (called individuals, creatures, or phenotypes) to an optimization problem, evolves toward better solutions. Traditionally, solutions are represented in binary as strings of 0s and 1s, but other encodings are also possible. The evolution usually starts from a population of randomly generated individuals and happens in generations. In each generation, the fitness of every individual in the population is evaluated, multiple individuals are stochastically selected from the current population (based on their fitness), and modified (recombined and possibly randomly mutated) to form a new population. The new population is then used in the next iteration of the algorithm. Commonly, the algorithm terminates when either a maximum number of generations has been produced, or a satisfactory fitness level has been reached for the population . . . . A typical genetic algorithm requires:
1] a genetic representation of the solution domain, 2] a fitness function to evaluate the solution domain.
A standard representation of the solution is as an array of bits. Arrays of other types and structures can be used in essentially the same way. The main property that makes these genetic representations convenient is that their parts are easily aligned due to their fixed size, which facilitates simple crossover operations . . . . The fitness function is defined over the genetic representation and measures the quality of the represented solution. The fitness function is always problem dependent. For instance, in the knapsack problem one wants to maximize the total value of objects that can be put in a knapsack of some fixed capacity. A representation of a solution might be an array of bits, where each bit represents a different object, and the value of the bit (0 or 1) represents whether or not the object is in the knapsack. Not every such representation is valid, as the size of objects may exceed the capacity of the knapsack. The fitness of the solution is the sum of values of all objects in the knapsack if the representation is valid, or 0 otherwise . . . . Once we have the genetic representation and the fitness function defined, GA proceeds to initialize a population of solutions randomly, then improve it through repetitive application of [operations].
The highlighted words and sentences are revealing. The GA is an intelligently constructed artifact that is critically dependent on being able to map a solution domain and evaluate fitness, based on a structured representation of the domain that can be CODED in a string or similar [extended] data structure. Solutions and representations are specific to the given problem, and so the search for a search exponentiation of search challenge issue Dembski highlights becomes an issue. In short, this is what such GA's end up doing: _______________ >> Too often, in the excitement over how wonderfully evolution has been modelled and how successful it is, this background of intelligent design and the way it drives the success of such exercises is forgotten. We tend to see what we expect to see or want to see . . . . GA’s model how already functioning genomes and life forms with effective body plans may adapt to niches in their environments, and to shifts in their environment. Given the way that there is so much closely matched co-adaptation of components . . . — i.e the dreaded irreducible complexity appears, if there is a drastic shift in environment, this explains why there is then a tendency to vanish from the annals of life. In short, we see something that fits the actual dominant feature of the fossil record: sudden appearance as a new body plan is introduced, stasis with a measure of adaptation and relatively minor variations within the overall body plan — that are explained on relatively minor mutations [especially on regulatory features such as size and proportions], disappearance or continuity into the modern world. And, Ilion’s point is still standing: in our observation, even through a machine that “cans” it, information transformation comes from mind. >> ________________ GA's overcome the physical limitations of atoms blindly scanning states through chance and blind necessity, because we are able to intelligently organise machines and program them to solve problems, including by hill-climbing within an island of function that after much painstaking effort, we have set up. So, since the search has now been narrowed down intelligently to the near vicinity of a relevant target, it becomes feasible to search on a restricted use of chance variation to detect the slope that points upward. Then, just as the advice to a lost hiker is to find a stream and go downhill as that way lies civilisation, hill climbing on a known nice slope will head for the peaks of performance. But that is critically dependent on and expresses the information we have loaded in. As the Mandelbrot set maps and video zooms show, that information can in principle be infinite, only limited by the resources to express it through that canned proxy and tireless mechanical calculator that does just what we tell it to no more and no less [hence GIGO], the computer. GEM of TKIkairosfocus
April 20, 2011
April
04
Apr
20
20
2011
02:50 AM
2
02
50
AM
PDT
1 2 3 4 5 6 7

Leave a Reply