Uncommon Descent Serving The Intelligent Design Community
meniscus

FOOTNOTE: On Einstein, Dembski, the Chi Metric and observation by the judging semiotic agent

Categories
Intelligent Design
Share
Facebook
Twitter/X
LinkedIn
Flipboard
Print
Email

(Follows up from here.)

Over at MF’s blog, there has been a continued stream of  objections to the recent log reduction of the chi metric in the recent CSI Newsflash thread.

Here is commentator Toronto:

__________

>> ID is qualifying a part of the equation’s terms with subjective observation.

If I do the same to Einstein’s, I might say;

E = MC^2, IF M contains more than 500 electrons,

BUT

E **MIGHT NOT** be equal to MC^2 IF M contains less than 500 electrons

The equation is no longer purely mathematical but subject to other observations and qualifications that are not mathematical at all.

Dembski claims a mathematical evaluation of information is sufficient for his CSI, but in practice, every attempt at CSI I have seen, requires a unique subjective evaluation of the information in the artifact under study.

The determination of CSI becomes a very small amount of math, coupled with an exhausting study and knowledge of the object itself.>>

_____________

A few thoughts in response:

a –> First, let us remind ourselves of the log reduction itself, starting with Dembski’s 2005 chi expression:

χ = – log2[10^120 ·ϕS(T)·P(T|H)]  . . . eqn n1

How about this (we are now embarking on an exercise in “open notebook” science):

1 –> 10^120 ~ 2^398

2 –> Following Hartley, we can define Information on a probability metric:

I = – log(p) . . .  eqn n2

3 –> So, we can re-present the Chi-metric:

Chi = – log2(2^398 * D2 * p)  . . .  eqn n3

Chi = Ip – (398 + K2) . . .  eqn n4

4 –> That is, the Dembski CSI Chi-metric is a measure of Information for samples from a target zone T on the presumption of a chance-dominated process, beyond a threshold of at least 398 bits, covering 10^120 possibilities.

5 –> Where also, K2 is a further increment to the threshold that naturally peaks at about 100 further bits . . . . As in (using Chi_500 for VJT’s CSI_lite):

Chi_500 = Ip – 500,  bits beyond the [solar system resources] threshold  . . . eqn n5

Chi_1000 = Ip – 1000, bits beyond the observable cosmos, 125 byte/ 143 ASCII character threshold . . . eqn n6

Chi_1024 = Ip – 1024, bits beyond a 2^10, 128 byte/147 ASCII character version of the threshold in n6, with a config space of 1.80*10^308 possibilities, not 1.07*10^301 . . . eqn n6a . . . .

Using Durston’s Fits from his Table 1, in the Dembski style metric of bits beyond the threshold, and simply setting the threshold at 500 bits:

RecA: 242 AA, 832 fits, Chi: 332 bits beyond

SecY: 342 AA, 688 fits, Chi: 188 bits beyond

Corona S2: 445 AA, 1285 fits, Chi: 785 bits beyond  . . . results n7

The two metrics are clearly consistent . . . .one may use the Durston metric as a good measure of the target zone’s actual encoded information content, which Table 1 also conveniently reduces to bits per symbol so we can see how the redundancy affects the information used across the domains of life to achieve a given protein’s function; not just the raw capacity in storage unit bits [= no.  of  AA’s * 4.32 bits/AA on 20 possibilities, as the chain is not particularly constrained.]

b –> In short, we are here reducing the explanatory filter to a formula. Once we have specific, observed functional information of Ip bits,  and we compare it to a threshold of a sufficiently large configuration space, we may infer that the instance of FSCI (or more broadly CSI)  is sufficiently isolated that the accessible search resources make it maximally unlikely that its best explanation is unintelligent cause by blind chance plus mechanical necessity. Instead, the best, and empirically massively supported causal explanation is design:

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

Fig 1: The ID Explanatory Filter

c –> This is especially clear when we use the 1,000 bit threshold, but in fact the “practical” universe we have is our solar system. And so, since the number of Planck time quantum states of our solar system since the usual date of the big bang is not more than 10^102, something that is in a config space of 10^150 [500 bits worth of possibilities] is 48 orders of magnitude beyond that threshold.

d –> So, something from a config space of 10^150 or more (500+ functionally specific bits) is on infinite monkey analysis grounds, comfortably beyond available search resources. 1,000 bits puts it beyond the resources of the observable cosmos:

Fig 2: The Observed Cosmos search window

e –> What the reduced Chi metric is telling us is that if say we had 140 functional bits [20 ASCII characters] , we would be 360 bits short of the threshold, and in principle a random walk based search could find something like that. For, while the reduced chi metric is giving us a value, it tells us we are falling short and by how much:

Chi_500(140 bits) = 140 – 500 = – 360 specific bits, within the threshold

f –> So, the Chi_500 metric tells us instances of this could happen by chance and trial and error testing.   Indeed, that is exactly what has happened with random text generation experiments:

One computer program run by Dan Oliver of Scottsdale, Arizona, according to an article in The New Yorker, came up with a result on August 4, 2004: After the group had worked for 42,162,500,000 billion billion monkey-years, one of the “monkeys” typed, “VALENTINE. Cease toIdor:eFLP0FRjWK78aXzVOwm)-‘;8.t” The first 19 letters of this sequence can be found in “The Two Gentlemen of Verona”. Other teams have reproduced 18 characters from “Timon of Athens”, 17 from “Troilus and Cressida”, and 16 from “Richard II”.[20]

A website entitled The Monkey Shakespeare Simulator, launched on July 1, 2003, contained a Java applet that simulates a large population of monkeys typing randomly, with the stated intention of seeing how long it takes the virtual monkeys to produce a complete Shakespearean play from beginning to end. For example, it produced this partial line from Henry IV, Part 2, reporting that it took “2,737,850 million billion billion billion monkey-years” to reach 24 matching characters:

RUMOUR. Open your ears; 9r"5j5&?OWTY Z0d

g –> But, 500 bits or 72 ASCII characters, and beyond this 1,000 bits or 143 ASCII characters, are a very different proposition, relative to the search resources of the solar system or the observed cosmos.

h –> That is why, consistently, we observe CSI beyond that threshold [e.g. Toronto’s comment] being produced by intelligence, and ONLY as produced by intelligence.

i –> So, on inference to best empirically warranted explanation, and on infinite monkeys analytical grounds, we have excellent reason to have high confidence that the threshold metric is credible.

j –> As a bonus, we have exposed the strawman suggestion that the Chi metric only applies beyond the threshold. Nope, it applies within the threshold and correctly indicates that something of such an order could come about by chance and necessity within the solar system’s search resources.

k –> is a threshold metric inherently suspicious? Not at all. In control system studies, for instance, we learn that once you reduce your expression to a transfer function of form

G = [(s – z1)(s- z2) . . . ]/[(s – p1)(s-p2)(s – p3) . . . ]

. . . then, if poles appear in the RH side of the complex s-plane, you have an unstable system.

l –> A threshold, and one that, when poles approach close to the threshold from the LH half-plane, will show up in a tendency that can be detected in the frequency response as peakiness.

m –> Is the simplicity of the math in question, in the end [after you have done the hard work of specifying information, and identifying thresholds], suspicious? No, again. For instance, let us compare:

v = i* R

q = v* C

n = sin i/ sin r

F = m*a

F2 = – F1

s = k log W

E = m0*c^2

v = H0D

Ik = – log2 (pk)

E = h*νφ

n –> Each of these is elegantly simple, but awesomely powerful; indeed, the last — precisely, a threshold relationship — was a key component of Einstein’s Nobel Prize (Relativity was just plain too controversial). And, once we put them to work in practical, empirical situations, each of them ” . . .  is no longer purely mathematical but subject to other observations and qualifications that are not mathematical at all.”

(The objection is clearly selectively hyperskeptical. Since when was an expression about an empirical quantity or situation “purely mathematical”? Let’s try another expression:

Y = C + I + G + [X – M].

How are its components measured and/or estimated, and with how much application of judgement calls, including those tracing to GAAP? [Cf discussion here.] Is this expression therefore meaningless and of no utility? What about M*VT = PT*T?)

o –> So, what about that horror, the involvement of the semiotic, judging agent as observer, who may even intervene and– shudder — judge? Of course, the observer is a major part of quantum mechanics, to the point where some are tempted to make it into a philosophical position. But the problem starts long before that, e.g. look at the problem of reading a meniscus! (Try, for Hg in glass, and for water in glass — the answers are different and can affect your results.)

Fig 3: Reading a meniscus to obtain volume of a liquid is both subjective and objective (Fair use clipping.)

p –> So, there is nothing in principle or in practice wrong with looking at information, and doing exercises — e.g. see the effect of deliberately injected noise of different levels, or of random variations — to test for specificity. Axe does just this, here, showing the islands of function effect dramatically. Clipping:

. . . if we take perfection to be the standard (i.e., no typos are tolerated) then P has a value of one in 10^60. If we lower the standard by allowing, say, four mutations per string, then mutants like these are considered acceptable:

no biologycaa ioformation by natutal means
no biologicaljinfommation by natcrll means
no biolojjcal information by natiral myans

and if we further lower the standard to accept five mutations, we allow strings like these to pass:

no ziolrgicgl informationpby natural muans
no biilogicab infjrmation by naturalnmaans
no biologilah informazion by n turalimeans

The readability deteriorates quickly, and while we might disagree by one or two mutations as to where we think the line should be drawn, we can all see that it needs to be drawn well below twelve mutations. If we draw the line at four mutations, we find P to have a value of about one in 10^50, whereas if we draw it at five mutations, the P value increases about a thousand-fold, becoming one in 10^47.

q –> Let us note how — when confronted with the same sort of skepticism regarding the link between information [a “subjective” quantity] and entropy [an “objective” one tabulated in steam tables etc] — Jaynes replied:

“. . . The entropy of a thermodynamic system is a measure of the degree of ignorance of a person whose sole knowledge about its microstate consists of the values of the macroscopic quantities . . . which define its thermodynamic state. This is a perfectly ‘objective’ quantity . . . it is a function of [those variables] and does not depend on anybody’s personality. There is no reason why it cannot be measured in the laboratory.”

r –> In short, subjectivity of the investigating observer is not a barrier to the objectivity of the conclusions reached, providing they are warranted on empirical and analytical grounds. As has been provided for the Chi metric, in reduced form.  END

Comments
kairosfocus,
Re: But you don’t have to search the whole search space if you’re lucky. If I lose my keys I rarely have to search ALL the possible places they could be . . .
In short, you admit that you believe in a lucky noise machine/ miracle, similar to a perpetual motion machine of the second kind.
Not at all. ellazimm is simply pointing out that evolutionary mechanisms are not analogous to random search. Evolutionary mechanisms can be modeled, with some caveats, as searching in the immediate vicinity of known functional regions of the fitness landscape. In the real world we observe that the "fitness landscape" (again, treated as a model with caveats) is amenable to this type of search. Once you have provided your rigorous mathematical definition of CSI and demonstrated how to calculate it for the first of my scenarios, we can have a very interesting conversation about how this affects probability distributions. That must wait for your other answers, though.MathGrrl
May 19, 2011
May
05
May
19
19
2011
06:08 AM
6
06
08
AM
PDT
Joseph,
Your brief description is not a rigorous mathematical definition of CSI and it is not aligned with Dembski’s description.
Yes, it is aligned with Dembski’s description and I have explained the mathematical rigor.
Simply asserting this does not make it so. You need to demonstrate it with references to Dembski's description. When you attempt to do so, you will find that your definition is not aligned with his.
You have also not shown how to calculate CSI objectively for any of the scenarios I’ve described.
I say that I have.
Again, it's easy to make assertions but supporting them requires effort. If you really had shown this, you would be able to copy and paste your calculation or provide a reference to it. You have done neither.
Thus far, no one has shown how to calculate CSI for any of the scenarios I described, nor has anyone provided a rigorous mathematical definition that is consistent with Dembski’s description.
I strongly disagree.
This isn't a matter of opinion. If you have examples of a rigorous mathematical definition of CSI, as described by Dembski, and calculations for my four scenarios, please produce them. The fact that you don't do so in response to my statements is telling.
I would expect someone making such claims to have already performed calculations similar to those I am requesting.
And I say it has been done, just not for your silly examples.
But you just said that it had been done for my "silly" examples (more of the civil discourse expected here, I see). Which is it?MathGrrl
May 19, 2011
May
05
May
19
19
2011
06:07 AM
6
06
07
AM
PDT
kairosfocus,
The way to do that is here, following on from Fig I.2 (cf. also I.1 and I.3) — and has been for many months, i.e reduce to a net list on the nodes-arcs wireframe, with some exploration of the noise to lose portraiture fidelity.
The calculation related to Hamlet on your referenced site suggests that you calculate CSI as two to the power of the number of bits required to describe the artifact under consideration. Is that correct?MathGrrl
May 19, 2011
May
05
May
19
19
2011
06:06 AM
6
06
06
AM
PDT
kairosfocus, I have read through all of your responses since my comment numbered 60 in this thread and have yet to see you address the two very direct questions I've asked. Let's try to make some progress by breaking this down into simple questions that can be answered succinctly. First, you repeatedly claim that CSI has been rigorously defined mathematically, but nowhere do you provide that rigorous mathematical definition. You could eliminate the need for your assertions by simply reproducing the definition here in this thread, in a single comment without any extraneous material. Could you please do so? Second, you have yet to reply to my question in comment 59:
CSI, I have explicitly said, many times, is a descriptive concept that describes an observed fact
By this, are you asserting that it is not possible to provide a mathematically rigorous definition of CSI, even in principle? If your answer is yes, I think you have a disagreement with some of your fellow ID proponents. If your answer is no, could you please simply state the mathematically rigorous definition of CSI, as described by Dembski, in a single, stand alone comment, without myriad tangential points, postscripts, and footnotes? It would go a long way to clarifying your position.
With these two questions answered, again as succinctly as possible, I believe we can make some progress in the discussion. Are you willing to work with me on this?MathGrrl
May 19, 2011
May
05
May
19
19
2011
06:05 AM
6
06
05
AM
PDT
F/N: Jospeh, just above, is using the known storage capacity of DNA as a metric of information, i.e. directly reading the bits of storage used, just like we do for program code, files, computer memory and DVDs or USB sticks. This is obviously a valid metric -- the most widely used and understood one, and it is the easiest one to use. DNA is based on strings of 4-state elements, and that makes it 2 bits per element. Codons for AAs in proteins use three bases, and so 6 bits. A 300 AA protein requires 1,800 bits. Durston et al were using a metric that takes into account a certain degree of redundancy in proteins due to flexibility in AAs being used in the chains. That, too, is valid, and it indicates that a certain fraction of the storage capacity is being used to effect the actual information; notice their example as excerpted, where for a given protein something like 10^-106 percent of the available space is valid for that protein. In my earlier remarks, I made an allowance for that in discussing on the FSCI in something like a Mycoplasma: 100,000+ bases, for a parasitical organism dependent on others for key components of life. You will recall, I wrote off half as "junk -- most unlikely. I then used just 1 bit per base, to allow for redundancy. The results are all going to be of the same order of magnitude, and all of them point well beyond the threshold where a cosmos-scope search on chance based random walks plus trial and error is a viable explanation for what we see. In short, we are in a situation where we can be roughly right, and that gives us orders of magnitude in hand, relative to the threshold for the search resource limit of the solar system or the observed cosmos. The resistance to the conclusion that FSCI beyond that threshold is -- on empirical and analytical grounds -- best explained by intelligence, is plainly ideological -- i.e. a priori materialism imposed regardless of the evidence, not properly scientific. At least, if we understand science as follows:
an unfettered (but ethically and intellectually responsible) progressive pursuit of the truth about our world in light of observation, experiment, theoretical modelling and analysis, testing and discussion among the informed.
kairosfocus
May 19, 2011
May
05
May
19
19
2011
05:48 AM
5
05
48
AM
PDT
arkF: Joseph (in addition to the points that Heinrich makes above)– as I am sure you know, the Shannon information in a message is relative to a probability model of the possible outcomes. That is why 2 bits per nucleotide and 6 bits per amino acid.
For some reason you seem to be assuming a model that all nucleotides are equally likely in all cases.
There aren't any physio-chemical laws that tell us the order of nucleotides along one side of a DNA sequence.
But that clearly is not true of real biological events such as gene duplication.
You are clearly confused. First there isn't any evidence that a gene duplication is via blind, undirected chemical processes. And second the way the nucleotides are strung is not decided by any physio-chemical law.
Hence the request to do the calculation for real events i.e. you cannot just look at a gene or amino acid and work out the amount of Shannon information it contains.
Yes, you can.
You need to understand the context in which that gene or amino acid was created to calculate the Shannon information.
Good luck showing that to be true. You do realize that your "word" is not a valid reference.
But no doubt you understand all this – what I would like to see is the calculation in a real context.
I would like to see you produce positive eviodence for your position but we both know that ain't going to happen. And I would love to see the methodology used that determined gene duplications are blind watchmaker processes.Joseph
May 19, 2011
May
05
May
19
19
2011
04:31 AM
4
04
31
AM
PDT
CSI is Shannon information, of a specified complexity, with meaning/ function. The math behind Shannon information tells us, with resect to biology, thee are 2 bits of information per nucleotide as there are 4 possible nucleotides, 4 = 2^2 =2 bits/ nuc. For amino acids it will be 6 bits as there are 64 possible coding codons (including STOP), 64 = 2^6 = 6 bits. Do you understand that bit of math, MathGrrl? Are we OK with the fact that Shannon informationhas been duly defined with mathematical rigor? For the complexity part Dembski provided the math in “No Free Lunch”. And for specification you need to determine the variation tolerance- and the math for that also exists and has been presented to you. Heinrich:
How do you formally define “specified complexity”, and “meaning/ function”?
I told you already- Dembski took care of the complexity part in NFL and he also covered "meaning/ function". In biology specification refers to biological function. IOW "information" as it is used by IDists is the same as every day use.Joseph
May 19, 2011
May
05
May
19
19
2011
04:21 AM
4
04
21
AM
PDT
F/N: Kindly note my point by point response to MG at 23 above, which she is ignoring in her latest round of drumbeat repetition of already answered talking points.kairosfocus
May 19, 2011
May
05
May
19
19
2011
03:21 AM
3
03
21
AM
PDT
F/N: Let me cite from NFL, pp. 144 and 148, as well as Dembski's 2005 paper, attention also being drawn to the remarks in 35 above on why the CSI metric in the log reduced form is sufficiently well warranted empirically and analytically to be used with reasonable confidence: ____________________ CSI: >> p. 148: “The great myth of contemporary evolutionary biology is that the information needed to explain complex biological structures can be purchased without intelligence. My aim throughout this book is to dispel that myth . . . . Eigen and his colleagues must have something else in mind besides information simpliciter when they describe the origin of information as the central problem of biology. I submit that what they have in mind is specified complexity [cf. here], or what equivalently we have been calling in this Chapter Complex Specified information or CSI . . . . Biological specification always refers to function . . . In virtue of their function [a living organism's subsystems] embody patterns that are objectively given and can be identified independently of the systems that embody them. Hence these systems are specified in the sense required by the complexity-specificity criterion . . . the specification can be cashed out in any number of ways [[through observing the requisites of functional organisation within the cell, or in organs and tissues or at the level of the organism as a whole] . . .” >> Notice, the CSI concept is primarily a description of an empirical reality, commonplace in the technological world, and as highlighted for the biological world by Orgel and Wicken et al, from the 1970's. Dembski's contribution is to have constructed a model and metric for when something is sufficiently complex AND specified jointly that it is reasonable to infer to intelligence as the only empirically and analytically credible source. In effect, he did so by applying the principle of the search challenge of finding a needle in a haystack, a metaphor that he in fact used. If there is sufficient haystack, the search resources accessible to us in the solar system or the cosmos as a while will be grossly inadequate and we would be entitled to infer to the only routinely observed cause of such CSI, on empirically anchored best explanation. Namely, design. EZ's appeal to statistical miracles, above, show just how fundamentally sound the approach is, despite the drumbeat talking point about "not rigorous." (Which is turning into a code word for, to stick with my ideological commitment to a priori materialism, or as a fellow traveller of those who do so, I will exert selective hyperskepticism.) DEFINITION: >> p. 144: [[Specified complexity can be defined:] “. . . since a universal probability bound of 1 [[chance] in 10^150 corresponds to a universal complexity bound of 500 bits of information, [[the cluster] (T, E) constitutes CSI because T [[ effectively the target hot zone in the field of possibilities] subsumes E [[ effectively the observed event from that field], T is detachable from E, and and T measures at least 500 bits of information . . . ” >> 2005 Eqn (excerpted from cite in 26 - 27, Weak Argument Correctives, top right this and every UD page for years): >> pp. 17 – 24, he argues: define ?S as . . . the number of patterns for which [agent] S’s semiotic description of them is at least as simple as S’s semiotic description of [a pattern or target zone] T. [26] . . . . where M is the number of semiotic agents [S's] that within a context of inquiry might also be witnessing events and N is the number of opportunities for such events to happen . . . . [where also] computer scientist Seth Lloyd has shown that 10^120 constitutes the maximal number of bit operations that the known, observable universe could have performed throughout its entire multi-billion year history.[31] . . . [Then] for any context of inquiry in which S might be endeavoring to determine whether an event that conforms to a pattern T happened by chance, M·N will be bounded above by 10^120. We thus define the specified complexity [?] of T given [chance hypothesis] H [in bits] . . . as [the negative base-2 logarithm of the conditional probability P(T|H) multiplied by the number of similar cases ?S(t) and also by the maximum number of binary search-events in our observed universe 10^120] ? = – log2[10^120 ·?S(T)·P(T|H)]. >> WAC 27 goes on to say: >> Debates over Dembski’s models and metrics notwithstanding, the basic point of a specification is that it stipulates a relatively small target zone in so large a configuration space that the reasonably available search resources — on the assumption of a chance-based information-generating process — will have extremely low odds of hitting the target. So low, that random information generation becomes an inferior and empirically unreasonable explanation relative to the well-known, empirically observed source of CSI: design. >> I add as well the log reduction from the OP above, that turns this into the directly applicable expression: >> ? = – log2[10^120 ·?S(T)·P(T|H)] . . . eqn n1 How about this (we are now embarking on an exercise in “open notebook” science): 1 –> 10^120 ~ 2^398 2 –> Following Hartley, we can define Information on a probability metric: I = – log(p) . . . eqn n2 3 –> So, we can re-present the Chi-metric: Chi = – log2(2^398 * D2 * p) . . . eqn n3 Chi = Ip – (398 + K2) . . . eqn n4 4 –> That is, the Dembski CSI Chi-metric is a measure of Information for samples from a target zone T on the presumption of a chance-dominated process, beyond a threshold of at least 398 bits, covering 10^120 possibilities. 5 –> Where also, K2 is a further increment to the threshold that naturally peaks at about 100 further bits. >> Applying this, we see that in he case of some of Durston et al's 35 protein families: >> Using Durston’s Fits from his Table 1, in the Dembski style metric of bits beyond the threshold, and simply setting the threshold at 500 bits: RecA: 242 AA, 832 fits, Chi: 332 bits beyond SecY: 342 AA, 688 fits, Chi: 188 bits beyond Corona S2: 445 AA, 1285 fits, Chi: 785 bits beyond . . . results n7 The two metrics are clearly consistent, and Corona S2 would also pass the X metric’s far more stringent threshold right off as a single protein. (Think about the cumulative fits metric for the proteins for a cell . . . ) In short one may use the Durston metric as a good measure of the target zone’s actual encoded information content, which Table 1 also conveniently reduces to bits per symbol so we can see how the redundancy affects the information used across the domains of life to achieve a given protein’s function; not just the raw capacity in storage unit bits [= no. of AA's * 4.32 bits/AA on 20 possibilities, as the chain is not particularly constrained.] >> From Taub and Schilling, Princs of Comm Systems, 2nd edn, Taub and Schilling (McGraw Hill, 1986), p. 512, Sect. 13.2 [cf as well my notes here, building on F R Connor] we can see the underlying definition and quantification of information -- which Schneider tried to "correct" Dembski for using, by substituting the far rarer synonym, "surprisal": >> Let us consider a communication system in which the allowable messages are m1, m2, . . ., with [observed, e.g. by studying the proportion of letters in typical text, as printers were long aware of] probabilities of occurrence p1, p2, . . . . Of course p1 + p2 + . . . = 1. Let the transmitter select message mk of probability pk; let us further assume that the receiver has correctly identified the message [My nb: i.e. the a posteriori probability in my online discussion is 1]. Then we shall say, by way of definition of the term information, that the system has communicated an amount of information Ik given by Ik = (def) log2 1/pk (13.2-1) [i.e. Ik = - log2 pk] >> ______________________ That, folks is what is being dodged and obfuscated behind a smoke cloud of selectively hyperskeptical objections, as I showed in 34 - 35 above. I again suggest the interested onlooker look at my discussion here. GEM of TKIkairosfocus
May 19, 2011
May
05
May
19
19
2011
01:45 AM
1
01
45
AM
PDT
Mr Mark F: I see that you have now managed to make a comment in a thread that I have posted, but have also managed to ignore the substantial matter. I hope that you have by now managed to clean up the behaviour at your own blog, where there was violation of my privacy, in what looks very much like a case of attempted "outing" to harm. Kindly, inform me of such corrective measures as you have taken. When it comes to the substantial matter you tried to address above by cherry picking -- a form of strawman, onlookers; MF being a trained philosopher -- something Joseph said while ignoring my elaboration on it, I note that you have again used the tactic of uncivilly ignoring what I have had to say, in the interests of making debate points off what Jospeh had said in brief. This is instantly utterly revealing on the true want of merits in your claims. I therefore suggest that if you intend to continue commenting in this thread, kindly address the specific matters in the original post (including the post to which this is a footnote), and the relevant matters in no 71 above, which you have artfully dodged. Good day, sir. GEM of TKIkairosfocus
May 19, 2011
May
05
May
19
19
2011
01:24 AM
1
01
24
AM
PDT
Heinrich: Kindly, start with the definitions cited in NFL, pp. 144 and 148, as were already quoted in the post to which this is a footnote. Kindly, explain -- specifically -- just what part of the definition there is inadequate: an observed event E within an independently identifiable limited island of function or specificity [zone T] coming from a space of 10^150 or more possibilities, i.e. beyond the lifetime, quantum state search resources of our solar system of ~10^55- 57 atoms or so: 10^102 states. Remember, the very fastest chemical reactions take up about 10^30 Planck times. (My citation and discussion here may be helpful.) You may then go to Dembski's 2005 paper on specification, which elaborates. Observe that the objection on the quantification there evaporates once we move the equation for Chi one step forward by doing a log reduction. In each case, your objections must be "rigorously defined," and must pass at least the criterion of inference to best explanation in light of empirical test. In case of mathematical objections, show your working relative to first principles and/or established results. (In short, I am saying that your objection falls apart once you have to meet your own standard. In contrast, onlookers, I invite you to look to 34 - 35 above, as already linked but studiously ignored or brushed aside -- over at MF's blog, with ad hominem attacks, on the subject of "rigour" and scientific work of morelling and metrics.) Heinrich, constantly repeating a false talking point -- "not rigorous" -- does not make it into truth, regardless of how intensely you want to believe the falsehoods. GEM of TKIkairosfocus
May 19, 2011
May
05
May
19
19
2011
01:13 AM
1
01
13
AM
PDT
EZ:
Re: But you don’t have to search the whole search space if you’re lucky. If I lose my keys I rarely have to search ALL the possible places they could be . . .
In short, you admit that you believe in a lucky noise machine/ miracle, similar to a perpetual motion machine of the second kind. This is a case of inference to the inferior explanation. You are able to find your keys because the scope of search is small enough that it is reasonable to find on trial and error. Once the scope of search becomes large enough, the island of function -- location where you can see the keys in this case -- becomes so isolated that search becomes infeasible. For instance, if you go boating and your key drops off somewhere unknown into a lake of reasonable size, you go get a new lock, you don't try to search for it. No prizes for guessing why. We are dealing here -- for OOL and/or OO novel body plans -- with scopes of search where the resources of the whole cosmos are grossly inadequate, i.e. essentially the same grounds that are the basis for the second law of thermodynamics. I suggest that you read Abel here on the universal plausibility bound, here on the implications of thermodynamics, and here on the problem of relying on lucky noise. Also, Wiki on the infinite monkey theorem, here. (Pay particular attention to the results of actual tests cited at f in the original post above that show searches on a scope of 10^50 being feasible, which is about where Borel put the limit many years ago in a thermodynamics context.) In short, the problem is that any search of a config space of at least 500 - 1,000 bits worth of possibilities [10^150 - 10^301 possibilities] on the scope of the cosmos will round down to effectively zero search. That, BTW is why in hardware and software design contexts, beyond a certain point you not only cannot search by exhaustion, but you cannot sample enough to make a difference if you are looking for a rare condition in the space of possibilities, so you need analytical or heuristic methods that are intelligently directed and much more likely to succeeed than traial and error. The odds of these searches are so remote that they make a search that consists of marking a single atom in our cosmos for one Planck time at random across the lifespan of the cosmos, then moving about by chance to any place and time in the entire history of the cosmos and picking up just one atom -- lo and behold, the marked one at the precise 5* 10^-44 s when it is marked -- look like a sure thing by contrast. In short, to maintain belief in chance and necessity, you are appealing to a statistical miracle. Not once, but dozens and dozens of times over. ___________ So, the bottom-line is plain: when you have a routinely observed, empirically reliable explanation of the source of FSCI -- intelligence -- and you are forced to resort to appealing to many times over repeated statistical miracles to stick with the blind chance and necessity explanation, this is a strong sign that the problem is ideology not scientific reasonableness. GEM of TKI PS: Just in case you want to raise the winnability of lotteries (as I have seen in former times), lotteries are winnable because they are DESIGNED to be winnable. That is the scope of search is very carefully balanced indeed so the sponsors will by overwhelming probability make money, and some few lucky individuals (a proportion very carefully designed) will win enough to encourage enough people to tip in.kairosfocus
May 19, 2011
May
05
May
19
19
2011
12:55 AM
12
12
55
AM
PDT
KF: "2^50,000 possibilities is a search space of 3.16*10^15,051, far beyond the search resources of the cosmos, allowing for redundancy in the code." But you don't have to search the whole search space if you're lucky. If I lose my keys I rarely have to search ALL the possible places they could be. And I stop searching when I've found them. Even if I don't use my intelligence to help direct the search I would hardly ever have to search all possibilities.ellazimm
May 18, 2011
May
05
May
18
18
2011
11:00 PM
11
11
00
PM
PDT
Joseph at #61
CSI is Shannon information, of a specified complexity, with meaning/ function. The math behind Shannon information tells us, with resect to biology, thee are 2 bits of information per nucleotide as there are 4 possible nucleotides, 4 = 2^2 =2 bits/ nuc. For amino acids it will be 6 bits as there are 64 possible coding codons (including STOP), 64 = 2^6 = 6 bits.
Joseph (in addition to the points that Heinrich makes above)– as I am sure you know, the Shannon information in a message is relative to a probability model of the possible outcomes.  For some reason you seem to be assuming a model that all nucleotides are equally likely in all cases.  But that clearly is not true of real biological events such as gene duplication.  Hence the request to do the calculation for real events i.e. you cannot just look at a gene or amino acid and work out the amount of Shannon information it contains.  You need to understand the context in which that gene or amino acid was created to calculate the Shannon information. But no doubt you understand all this – what I would like to see is the calculation in a real context.markf
May 18, 2011
May
05
May
18
18
2011
10:40 PM
10
10
40
PM
PDT
Joseph @61 -
CSI is Shannon information, of a specified complexity, with meaning/ function.
How do you formally define "specified complexity", and "meaning/ function"?Heinrich
May 18, 2011
May
05
May
18
18
2011
10:45 AM
10
10
45
AM
PDT
Onlookers: To get an idea of what is really going on with these drumbeat repetition falsehood-based objections, let us clip MG at 62:
Given the desire on the part of many ID proponents for ID to be accepted as science
[NB: ID will never be accepted as science so long as the institutions of science are captive to a priori, question begging evolutionary materialism, so the issue is not acceptability but the ending of ideological captivity of science to materialism; once the materialist censorship is ended, it is at once obvious that the design inference is a scientific process]
, I would expect those proponents to be eagerly applying CSI calculations to real world artifacts
[done, MG just refuses to acknowledge this]
. Scientists, in my experience, don’t sit back and ask others to research their hypotheses.
[I must protest this ad hominem laced strawman caricature immediately: Really now, and what did Einstein do with his gravitational lens prediction of General Relativity, of 1916; i.e. I allude to the famous observations of 1919? There is a reason why there is commonly a division of labour in science between experimentalists and theoreticians. So, even if it were true -- and it is not -- that design researchers have not provided real world values of CSI on design inference principles, starting with Dembski's value for the flagellum about 10 years ago in NFL . . . so this was false from the beginning, as MG has been explicitly corrected but as usual has brushed aside and proceeds to tell the false as though it were true . . . and MG claims to be familiar with NFL!]
Now, the above is manifestly false and accusatory, even in the immediate context of this discussion here at UD. (And Joseph is fundamentally correct to raise the simple point that the known storage capacity of DNA is 2 bits per base, so a genome of 100,000 bases, even if we (overgenerously) write off half as "junk," and round down from 2 to 1 bit per base is storing 50 k bits of functionally specific info. Known to be specific, as it is coded and/or serves regulatory functions for the code, starting with Mycoplasma. Let's go for just 1 bit per symbol (as just noted) to take into account redundancies -- cf Durston et al and you will see this is reasonable or even generous. 2^50,000 possibilities is a search space of 3.16*10^15,051, far beyond the search resources of the cosmos, allowing for redundancy in the code. The search space challenge to get to the islands of function for first life are clearly beyond the capacity of the observed cosmos. The only empirically credible , observed cause of such functionally specific complex information is design. So, we have excellent reason to infer that first cell based life is designed, and the reasoning can be extended into seeing that body plans requiring 10+ m bases of further specific info to build the plans, dozens of times over, are also going to be credibly designed. Once we remove materialistic blinkers. On reasoning very closely related to that which warrants the 2nd law of thermodynamics. In short, MG's complaints on want of rigour boil down to being equivalent to objecting to the credibility of the 2nd law of thermodynamics. She is suggesting that the natural world has in it the equivalent of a perpetual motion machine, i.e a lucky noise machine capable of building FSCI-rich first life and body plans our of blind chance and mechanical necessity. Okay, physician heal thyself: provide empirical demonstration of chance and necessity giving rise to the equivalent of first life. We already know that the level of FSCI involved can be built by intelligences, routinely. Look around you onlookers: for years and years, evo mat objectors at UD have been asked to provide a clear example. The most they have been able to do is to point to something like ev, which it turns out is targetted search that works within an island of function, i.e it is at most an illustration of intelligently designed micro evolution. (Cf Mung's dissection of ev here, as just one example in point. this is one of the points that MG is refusing to explain herself on, as linked above and we may as well give it again here.) Conclusion: MG's selectively hyperskeptical objections are plainly motivated ideologically, not scientifically. To see how this has led her to make false accusations that she should or does know are false, compare the clip in the original post above (which one reads or should read before commenting), on the Durston et al metric being fed into Chi_500:
Using Durston’s Fits from his Table 1, in the Dembski style metric of bits beyond the threshold, and simply setting the threshold at 500 bits:
RecA: 242 AA, 832 fits, Chi: 332 bits beyond SecY: 342 AA, 688 fits, Chi: 188 bits beyond Corona S2: 445 AA, 1285 fits, Chi: 785 bits beyond . . . results n7
The two metrics are clearly consistent . . . .one may use the Durston metric as a good measure of the target zone’s actual encoded information content, which Table 1 also conveniently reduces to bits per symbol so we can see how the redundancy affects the information used across the domains of life to achieve a given protein’s function; not just the raw capacity in storage unit bits [= no. of AA's * 4.32 bits/AA on 20 possibilities, as the chain is not particularly constrained.]
Finally, let us observe point 26 from comment 24, where -- in correcting false statements by MG et al [which I specifically drew her attention to] I clipped a previously cited excerpt from the 2007 Durston et al paper, i.e. this is a repeat correction:
Consider that there are usually only 20 different amino acids possible per site for proteins, Eqn. (6) can be used to calculate a maximum Fit value/protein amino acid site of 4.32 Fits/site [NB: Log2 (20) = 4.32]. We use the formula log (20) – H(Xf) to calculate the functional information at a site specified by the variable Xf such that Xf corresponds to the aligned amino acids of each sequence with the same molecular function f. The measured FSC for the whole protein is then calculated as the summation of that for all aligned sites. The number of Fits quantifies the degree of algorithmic challenge, in terms of probability [info and probability are closely related], in achieving needed metabolic function. For example, if we find that the Ribosomal S12 protein family has a Fit value of 379, we can use the equations presented thus far to predict that there are about 10^49 different 121-residue sequences that could fall into the Ribsomal S12 family of proteins, resulting in an evolutionary search target of approximately 10^-106 percent of 121-residue sequence space. In general, the higher the Fit value, the more functional information is required to encode the particular function in order to find it in sequence space. A high Fit value for individual sites within a protein indicates sites that require a high degree of functional information. High Fit values may also point to the key structural or binding sites within the overall 3-D structure.
In short, Durston et al effectively invite slotting in their value of H in fits -- average info per symbol, recall -- as a measure of Ip. That is, as I did for the three examples: Chi_500 = Ip - 500, bits beyond the solar system threshold So, again, MG is repeating a falsehood, in hopes of winning debate points on drumbeat repetition. ________________ This is willful falsehood. Inexcusable. See why I have lost patience with the sort of bland, drumbeat declarations of falsehoods that MG et al know or should know are false? GEM of TKIkairosfocus
May 18, 2011
May
05
May
18
18
2011
10:37 AM
10
10
37
AM
PDT
ellazimm:
MathGrrl wants to be sure she uses a definition that is one you agree to and consistent with Dr Dembski’s definition. AND she’s asking to see that definition in action. I don’t understand why it’s so hard to give her what she wants. Or why you are casting so many aspersions on her motivations. It’s your chance to make a point. Take it!!
I have made my point. It is all a waste of time now. As I said until MathGrrl can provide something from evolutionary biology with mathematical rigor she will always just move the goal posts and be unsatisfied.Joseph
May 18, 2011
May
05
May
18
18
2011
10:19 AM
10
10
19
AM
PDT
MathGrrl:
Your brief description is not a rigorous mathematical definition of CSI and it is not aligned with Dembski’s description.
Yes, it is aligned with Dembski's description and I have explained the mathematical rigor.
You have also not shown how to calculate CSI objectively for any of the scenarios I’ve described.
I say that I have.
Thus far, no one has shown how to calculate CSI for any of the scenarios I described, nor has anyone provided a rigorous mathematical definition that is consistent with Dembski’s description.
I strongly disagree.
Your comment on funding strikes me as rather odd for at least two reasons.
Your requests strike me as rather odd.
First, many ID proponents make strong claims about CSI being a clear indicator of the involvement of intelligent agency.
And we have explained why that is- cause and effect relationships.
I would expect someone making such claims to have already performed calculations similar to those I am requesting.
And I say it has been done, just not for your silly examples.
Second, CSI is quite possibly the most clearly testable concept associated with ID.
I don't agree.Joseph
May 18, 2011
May
05
May
18
18
2011
10:14 AM
10
10
14
AM
PDT
Dr Bot: The way to do that is here, following on from Fig I.2 (cf. also I.1 and I.3) -- and has been for many months, i.e reduce to a net list on the nodes-arcs wireframe, with some exploration of the noise to lose portraiture fidelity. The detailed exercise itself would be non-trivial, but we already know that -- from 3-d computer animation -- we are well beyond the 500 or 1,000 bit FSCI thresholds and the result would affirm what we already directly know. GEM of TKIkairosfocus
May 18, 2011
May
05
May
18
18
2011
09:53 AM
9
09
53
AM
PDT
Just curious, can anyone give me a figure (and the math behind it) for the CSI in the Mount Rushmore National Memorial?DrBot
May 18, 2011
May
05
May
18
18
2011
09:08 AM
9
09
08
AM
PDT
Joseph: Please watch your tone. GEM of TKIkairosfocus
May 18, 2011
May
05
May
18
18
2011
09:03 AM
9
09
03
AM
PDT
PS: In addition, you are now associated with a site that has entertained abusive and uncivil behaviour, such as attempted outing. I have notified that site that I will only notify for the record. The bottomline remains that for weeks you have made no serious attempt to address cogent corrections, nor have you explained yourself in light of points of corrective concern highlighted in summary here. Instead you have tried to spread that which is false or highly misleading, and that in the teeth of what you know or should know. Persisting in such behaviour removes you from the circle of civil discussion.kairosfocus
May 18, 2011
May
05
May
18
18
2011
08:59 AM
8
08
59
AM
PDT
MG: Why do you keep on repeating already cogently and repeatedly answered objections as though they have merit? Do you not see that you are simply showing a textbook example of hyperskeptical, closed minded objectionism that does not even care to respond to substantial corrections, and brazenly asserts falsehoods [e.g. your four cases were answered to adequately many times and in particular the first one, that you repeated above has been answered yet again and again: a duplicate is a copy, i.e there is no fresh FSCI involved, but the process of duplication once the duplicate is beyond the reasonable threshold, implies duplicating machinery and algorithms, implying a lot of FSCI beyond the threshold. And the direct answer is therefore ZERO, as you received immediately, weeks ago. Just, you seem incapable/unwilling to understand that the question is misconceived and demands an answer that is based on the involved logic.]?
Thus far, no one has shown how to calculate CSI for any of the scenarios I described, nor has anyone provided a rigorous mathematical definition that is consistent with Dembski’s description
This is now willfully false in the teeth of what you know or should know, as has already been addressed above, and in previous threads over the course of two to three months now. All you seem able to do is repeat what you know or should know is false, or distorted, and which is intended to be damaging. Again and again, on being corrected and provided with details, you have simply ignored mere facts and cogent arguments relative to facts, and resorted to drumbeat repetition of selectively hyperskeptical, false, strawman tactic, and accusatory talking points. SHAME ON YOU! You again have several points of explanation to do, on matters outlined and linked on just above. These include questions of definition, rigour [cf here above in this thread on that specific point], and the cases you have put up. At no point have you showed the slightest responsiveness on merits. The conclusion is sadly plain, as your misbehaviour is speaking for itself. Please, do better than this. Good day GEM of TKIkairosfocus
May 18, 2011
May
05
May
18
18
2011
08:37 AM
8
08
37
AM
PDT
Joseph, MathGrrl wants to be sure she uses a definition that is one you agree to and consistent with Dr Dembski's definition. AND she's asking to see that definition in action. I don't understand why it's so hard to give her what she wants. Or why you are casting so many aspersions on her motivations. It's your chance to make a point. Take it!!ellazimm
May 18, 2011
May
05
May
18
18
2011
08:22 AM
8
08
22
AM
PDT
Joseph, Your brief description is not a rigorous mathematical definition of CSI and it is not aligned with Dembski's description. You have also not shown how to calculate CSI objectively for any of the scenarios I've described.
That said if you want your little project completed, ie your guest post, then I suggest you fund it or do it yourself with all the information you have been given.
Thus far, no one has shown how to calculate CSI for any of the scenarios I described, nor has anyone provided a rigorous mathematical definition that is consistent with Dembski's description. I'm still very interested in performing some tests of my own once an ID proponent has explained how to actually calculate CSI objectively. Your comment on funding strikes me as rather odd for at least two reasons. First, many ID proponents make strong claims about CSI being a clear indicator of the involvement of intelligent agency. I would expect someone making such claims to have already performed calculations similar to those I am requesting. Without doing so, such claims are baseless. Second, CSI is quite possibly the most clearly testable concept associated with ID. Given the desire on the part of many ID proponents for ID to be accepted as science, I would expect those proponents to be eagerly applying CSI calculations to real world artifacts. Scientists, in my experience, don't sit back and ask others to research their hypotheses.MathGrrl
May 18, 2011
May
05
May
18
18
2011
07:15 AM
7
07
15
AM
PDT
MathGrrl:
Please note that the rigorous mathematical definition of CSI as described by Dembski is an essential prerequisite to any calculations. Without that definition, the calculations are without context and therefore meaningless.
Your entire position on CSI is meaningless as you have been provided a mathematically rigorous definition of CSI- more mathematically rigorous than anthing the theory of evolution has to offer- and you just handwave it away. So here it is AGAIN: CSI is Shannon information, of a specified complexity, with meaning/ function. The math behind Shannon information tells us, with resect to biology, thee are 2 bits of information per nucleotide as there are 4 possible nucleotides, 4 = 2^2 =2 bits/ nuc. For amino acids it will be 6 bits as there are 64 possible coding codons (including STOP), 64 = 2^6 = 6 bits. Do you understand that bit of math, MathGrrl? Are we OK with the fact that Shannon informationhas been duly defined with mathematical rigor? For the complexity part Dembski provided the math in "No Free Lunch". And for specification you need to determine the variation tolerance- and the math for that also exists and has been presented to you. That said if you want your little project completed, ie your guest post, then I suggest you fund it or do it yourself with all the information you have been given. And if you cannot then change your moniker because you give math a bad name.
Based on my experience with working scientists, I would expect proponents of a non-mainstream hypothesis such as ID to welcome the attention and participation of such people.
Based on my experience dealing with them is as much a waste of time as dealing with you and the "mainstream" scientists should focus on their position as it is in need of a good colonic.Joseph
May 17, 2011
May
05
May
17
17
2011
05:10 PM
5
05
10
PM
PDT
Joseph,
You have also generated a large amount of text without directly addressing the issue at hand, namely whether or not you can provide a mathematically rigorous definition of CSI (as described in Dembski’s papers and books) and detailed examples of how to calculate it.
I provied that for you- complete with examples.
If that is, in fact, the case, please repeat it here, as succinctly as possible. I do not recall you doing so over the course of our interactions on this topic. Please note that the rigorous mathematical definition of CSI as described by Dembski is an essential prerequisite to any calculations. Without that definition, the calculations are without context and therefore meaningless.MathGrrl
May 17, 2011
May
05
May
17
17
2011
04:12 PM
4
04
12
PM
PDT
kairosfocus, My original requests to you, in comment 16 of this thread, were to provide a mathematically rigorous definition of CSI as described by Dembski and to show how to calculate CSI so defined for the first of the scenarios I proposed in my guest post:
A simple gene duplication, without subsequent modification, that increases production of a particular protein from less than X to greater than X. The specification of this scenario is "Produces at least X amount of protein Y."
Despite posting literally thousands of words since I made those requests, you have thus far not responded to them. The closest I have seen you come is this statement:
CSI, I have explicitly said, many times, is a descriptive concept that describes an observed fact
in comment 23. By this, are you asserting that it is not possible to provide a mathematically rigorous definition of CSI, even in principle? If your answer is yes, I think you have a disagreement with some of your fellow ID proponents. If your answer is no, could you please simply state the mathematically rigorous definition of CSI, as described by Dembski, in a single, stand alone comment, without myriad tangential points, postscripts, and footnotes? It would go a long way to clarifying your position.MathGrrl
May 17, 2011
May
05
May
17
17
2011
04:12 PM
4
04
12
PM
PDT
Mung,
By the way, as noted by Toronto on Mark Frank’s blog, a number of the participants there are not allowed to post comments here at UD. In the spirit of open discussion, I hope you will respond there.
You can’t carry your own water here?
I'm more than happy to continue to participate in the discussion here. I'm simply pointing out that there are others who would like to participate but are prevented from doing so.
Like having more monkeys typing on keyboards is somehow going to help you make your case?
Is that another example of the civility expected here at UD? Many of the people discussing this topic at Mark Frank's blog are doing so calmly, rationally, and civilly. They have expertise in a broad range of different scientific disciplines. Based on my experience with working scientists, I would expect proponents of a non-mainstream hypothesis such as ID to welcome the attention and participation of such people.MathGrrl
May 17, 2011
May
05
May
17
17
2011
04:11 PM
4
04
11
PM
PDT
Mung,
My belief is that MathGrrl felt she was losing the argument here and therefore ran away to some place where she hoped to get some help.
Your belief is unsubstantiated. I have had a ridiculously busy week at work and am just now finding the time to return to the discussion. I apologize to the other participants on this thread for my temporary disappearance; I'm sure you've all had similar real world demands on your time.
She was allowed to guest post here, she owes us the courtesty of remaining here to carry on her argument (if she has one).
I quite agree, although I don't see this discussion as an argument. From my guest post onward I have simply been requesting clarification of the mathematics behind Dembski's CSI metric. I still have yet to receive such clarification. I do take exception to one of your previous comments, apropos of that:
MathGrrl, you’re not interested in moving the debate along. You have nothing to offer beyond repeating ad nauseam the same ttwo demands.
When I get an answer to my requests (not demands) for sufficient information to be able to calculate Dembski's CSI objectively, I fully intend to run some tests of my own. Until someone here can provide that information, it isn't unreasonable to continue to request it. Rather than criticize me for doing so, perhaps you could help move the conversation forward by providing a mathematically rigorous definition of CSI, as described by Dembski, and demonstrate in detail how to calculate it for the four scenarios described in my guest thread?MathGrrl
May 17, 2011
May
05
May
17
17
2011
04:10 PM
4
04
10
PM
PDT
1 6 7 8 9 10

Leave a Reply