Uncommon Descent Serving The Intelligent Design Community

Is the CSI concept well-founded mathematically, and can it be applied to the real world, giving real and useful numbers?

Share
Facebook
Twitter
LinkedIn
Flipboard
Print
Email

Those who have been following the recently heated up exchanges on the theory of intelligent design and the key design inference on tested, empirically reliable signs, through the ID explanatory filter, will know that a key move in recent months was the meteoric rise of the mysterious internet persona MathGrrl (who is evidently NOT the Calculus Prof who has long used the same handle).

MG as the handle is abbreviated, is well known for “her” confident-manner assertion — now commonly stated as if it were established fact in the Darwin Zealot fever swamps that are backing the current cyberbullying tactics that have tried to hold my family hostage —  that:

without a rigorous mathematical definition and examples of how to calculate [CSI], the metric is literally meaningless. Without such a definition and examples, it isn’t possible even in principle to associate the term with a real world referent.

As the strike-through emphasises, every one of these claims has long been exploded.

You doubt me?

Well, let us cut down the clip from the CSI Newsflash thread of April 18, 2011, which was again further discussed in a footnote thread of 10th May (H’mm, anniversary of the German Attack in France in 1940), which was again clipped yesterday at fair length.

( BREAK IN TRANSMISSION: BTW, antidotes to the intoxicating Darwin Zealot fever swamp “MG dunit” talking points were collected here — Graham, why did you ask the question but never stopped by to discuss the answer? And the “rigour” question was answered step by step at length here.  In a nutshell, as the real MathGrrl will doubtless be able to tell you, the Calculus itself, historically, was founded on sound mathematical intuitive insights on limits and infinitesimals, leading to the warrant of astonishing insights and empirically warranted success, for 200 years. And when Math was finally advanced enough to provide an axiomatic basis — at the cost of the sanity of a mathematician or two [doff caps for a minute in memory of Cantor] — it became plain that such a basis was so difficult that it could not have been developed in C17. Had there been an undue insistence on absolute rigour as opposed to reasonable warrant, the great breakthroughs of physics and other fields that crucially depended on the power of Calculus, would not have happened.  For real world work, what we need is reasonable warrant and empirical validation of models and metrics, so that we know them to be sufficiently reliable to be used.  The design inference is backed up by the infinite monkeys analysis tracing to statistical thermodynamics, and is strongly empirically validated on billions of test cases, the whole Internet and the collection of libraries across the world being just a sample of the point that the only credibly known source for functionally specific complex information and associated organisation [FSCO/I]  is design.  )

After all, a bit of  careful citation always helps:

_________________

>>1 –> 10^120 ~ 2^398

I = – log(p) . . .  eqn n2
3 –> So, we can re-present the Chi-metric:
[where, from Dembski, Specification 2005,  χ = – log2[10^120 ·ϕS(T)·P(T|H)]  . . . eqn n1]
Chi = – log2(2^398 * D2 * p)  . . .  eqn n3
Chi = Ip – (398 + K2) . . .  eqn n4
4 –> That is, the Dembski CSI Chi-metric is a measure of Information for samples from a target zone T on the presumption of a chance-dominated process, beyond a threshold of at least 398 bits, covering 10^120 possibilities.
5 –> Where also, K2 is a further increment to the threshold that naturally peaks at about 100 further bits . . . .
6 –> So, the idea of the Dembski metric in the end — debates about peculiarities in derivation notwithstanding — is that if the Hartley-Shannon- derived information measure for items from a hot or target zone in a field of possibilities is beyond 398 – 500 or so bits, it is so deeply isolated that a chance dominated process is maximally unlikely to find it, but of course intelligent agents routinely produce information beyond such a threshold.

7 –> In addition, the only observed cause of information beyond such a threshold is the now proverbial intelligent semiotic agents.
8 –> Even at 398 bits that makes sense as the total number of Planck-time quantum states for the atoms of the solar system [most of which are in the Sun] since its formation does not exceed ~ 10^102, as Abel showed in his 2009 Universal Plausibility Metric paper. The search resources in our solar system just are not there.
9 –> So, we now clearly have a simple but fairly sound context to understand the Dembski result, conceptually and mathematically [cf. more details here]; tracing back to Orgel and onward to Shannon and Hartley . . . .
As in (using Chi_500 for VJT’s CSI_lite [UPDATE, July 3: and S for a dummy variable that is 1/0 accordingly as the information in I is empirically or otherwise shown to be specific, i.e. from a narrow target zone T, strongly UNREPRESENTATIVE of the bulk of the distribution of possible configurations, W]):
Chi_500 = Ip*S – 500,  bits beyond the [solar system resources] threshold  . . . eqn n5
Chi_1000 = Ip*S – 1000, bits beyond the observable cosmos, 125 byte/ 143 ASCII character threshold . . . eqn n6
Chi_1024 = Ip*S – 1024, bits beyond a 2^10, 128 byte/147 ASCII character version of the threshold in n6, with a config space of 1.80*10^308 possibilities, not 1.07*10^301 . . . eqn n6a
[UPDATE, July 3: So, if we have a string of 1,000 fair coins, and toss at random, we will by overwhelming probability expect to get a near 50-50 distribution typical of the bulk of the 2^1,000 possibilities W. On the Chi-500 metric, I would be high, 1,000 bits, but S would be 0, so the value for Chi_500 would be – 500, i.e. well within the possibilities of chance.  However, if we came to the same string later and saw that the coins somehow now had the bit pattern of the ASCII codes for the first 143 or so characters of this post, we would have excellent reason to infer that an intelligent designer, using choice contingency, had intelligently reconfigured the coins. that is because, using the same I = 1,000 capacity value, S is now 1, and so Chi_500 = 500 bits beyond the solar system threshold. If the 10^57 or so atoms of our solar system, for its lifespan, were to be converted into coins and tables etc, and tossed at an impossibly fast rate, it would be impossible to sample enough of the possibilities space W to have confidence that something from so unrepresentative a zone T,  could reasonably be explained on chance. So, as long as an intelligent agent capable of choice is possible, choice — i.e. design — would be the rational, best explanation on the sign observed, functionally specific, complex information.]
10 –> Similarly, the work of Durston and colleagues, published in 2007, fits this same general framework . . . .
We use the formula log (20) – H(Xf) to calculate the functional information at a site specified by the variable Xf such that Xf corresponds to the aligned amino acids of each sequence with the same molecular function f. The measured FSC for the whole protein is then calculated as the summation of that for all aligned sites. The number of Fits quantifies the degree of algorithmic challenge, in terms of probability [info and probability are closely related], in achieving needed metabolic function. For example, if we find that the Ribosomal S12 protein family has a Fit value of 379, we can use the equations presented thus far to predict that there are about 10^49 different 121-residue sequences that could fall into the Ribsomal S12 family of proteins, resulting in an evolutionary search target of approximately 10^-106 percent of 121-residue sequence space. In general, the higher the Fit value, the more functional information is required to encode the particular function in order to find it in sequence space . . . .
11 –> So, Durston et al are targetting the same goal, but have chosen a different path from the start-point of the Shannon-Hartley log probability metric for information. That is, they use Shannon’s H, the average information per symbol, and address shifts in it from a ground to a functional state on investigation of protein family amino acid sequences. They also do not identify an explicit threshold for degree of complexity. [Added, Apr 18, from comment 11 below:] However, their information values can be integrated with the reduced Chi metric:
Using Durston’s Fits from his Table 1, in the Dembski style metric of bits beyond the threshold, and simply setting the threshold at 500 bits:
RecA: 242 AA, 832 fits, Chi: 332 bits beyond
SecY: 342 AA, 688 fits, Chi: 188 bits beyond
Corona S2: 445 AA, 1285 fits, Chi: 785 bits beyond  . . . results n7
The two metrics are clearly consistent . . .  (Think about the cumulative fits metric for the proteins for a cell . . . )
In short one may use the Durston metric as a good measure of the target zone’s actual encoded information content, which Table 1 also conveniently reduces to bits per symbol so we can see how the redundancy affects the information used across the domains of life to achieve a given protein’s function; not just the raw capacity in storage unit bits [= no.  of  AA’s * 4.32 bits/AA on 20 possibilities, as the chain is not particularly constrained.]>>

_________________

So, there we have it folks:

I: Dembski’s CSI metric is closely related to standard and widely used work in Information theory, starting with I = – log p

II: It is reducible on taking the appropriate logs, to an information beyond a threshold value

III: The threshold is reasonably set by referring to the accessible search resources of a relevant system, i.e. our solar system or the observed cosmos as a whole.

IV: Where, once an observed configuration — event E, per NFL — that bears or implies information is from a separately and “simply” describable narrow zone T that is strongly unrepresentative — that’s key — of the space of possible configurations, W, then

V: since the search applied is of a very small fraction of W, it is unreasonable to expect that chance can reasonably account for E in T, instead of the far more typical possibilities in W of in aggregate, overwhelming statistical weight.

(For instance the 10^57 or so atoms of our solar system will go through about 10^102 Planck-time Quantum states in the time since its founding on the usual timeline. 10^150 possibilities [500 bits worth of possibilities] is 48 orders of magnitude beyond that reach, where it takes 10^30 P-time states to execute the fastest chemical reactions.  1,000 bits worth of possibilities is 150 orders of magnitude beyond the 10^150 P-time Q-states of the about 10^80 atoms of our observed cosmos. When you are looking for needles in haystacks, you don’t expect to find them on relatively tiny and superficial searches.)

VI: Where also, in empirical investigations we observe that an aspect of an object, system, process or phenomenon that is controlled by mechanical necessity will show itself in low contingency. A dropped, heavy object falls reliably at g. We can make up a set of differential equations and model how events will play out on a given starting condition, i.e we identify an empirically reliable natural law.

VII: By contrast, highly contingent outcomes — those that vary significantly on similar initial conditions, reliably trace to chance factors and/or choice, e.g we may drop a fair die and it will tumble to a value essentially by chance. (This is in part an ostensive definition, by key example and family resemblance.)  Or, I may choose to compose a text string, writing it this way or the next. Or as the 1,000 coins in a string example above shows, coins may be strung by chance or by choice.

VIII: Choice and chance can be reliably empirically distinguished, as we routinely do in day to day life, decision-making, the court room, and fields of science like forensics.  FSCO/I is one of the key signs for that and the Dembski-style CSI metric helps us quantify that, as was shown.

IX:  Shown, based on a reasonable reduction from standard approaches, and shown by application to real world cases, including biologically relevant ones.

We can safely bet, though, that you would not have known that this was done months ago — over and over again — in response to MG’s challenge, if you were going by the intoxicant fulminations billowing up from the fever swamps of the Darwin zealots.

Let that be a guide to evaluating their credibility — and, since this was repeatedly drawn to their attention and just as repeatedly brushed aside in the haste to go on beating the even more intoxicating talking point drums,  sadly, this also raises serious questions on the motives and attitudes of the chief ones responsible for those drumbeat talking points and for the fever swamps that give off the poisonous, burning strawman rhetorical fumes that make the talking points seem stronger than they are.  (If that is offensive to you, try to understand: this is coming from a man whose argument as summarised above has repeatedly been replied to by drumbeat dismissals without serious consideration, led on to the most outrageous abuses by the more extreme Darwin zealots (who were too often tolerated by host sites advocating alleged “uncensored commenting,” until it was too late), culminating now in a patent threat to his family by obviously unhinged bigots.)

And, now also you know the most likely why of TWT’s attempt to hold my family hostage by making the mafioso style threat: we know you, we know where you are and we know those you care about. END

Comments
Notice: no explanation for the trumpeted, drumbeat assertions against CSI, and the Chi metric, or its reduction. Gkairosfocus
July 5, 2011
July
07
Jul
5
05
2011
02:01 PM
2
02
01
PM
PDT
As ever, I point out re: "Chance and necessity" that the 'chance' is assumed, from the point of view of science. Intelligent agents are capable of making large, sudden changes, I admit. They're also capable of making small, incremental changes, or of employing such changes. That said, it's also reasonable to suppose real and practical limits to the small and incremental.nullasalus
July 5, 2011
July
07
Jul
5
05
2011
12:58 PM
12
12
58
PM
PDT
Liz: And so my answer would be that your brain has been “programmed” (scare quotes deliberate) both by evolution and by your own actions and experiences to do what you rightly say it can do.
Prove it.mike1962
July 5, 2011
July
07
Jul
5
05
2011
12:44 PM
12
12
44
PM
PDT
Elizabeth, I think I've been honest and fair towards you but this is the weakest argument I've seen from you, besides "thriving life = Genetic Entropy is not true". You are simply begging the question of how/when chance and necessity have ever created CSI. If you can provide evidence of this you could single-handedly shut down the ID movement.
your error is the basic ID error – the assumption Chance and Necessity can’t create CSI!
It's not an assumption l, it's an observation from decades of experiments and computer simulations and thousands of years of human experience, as well as the extinguished fantasies of many a computer programmer.
your brain has been “programmed” (scare quotes deliberate) both by evolution and by your own actions and experiences to do what you rightly say it can do
Who/what is "you" and what does it have to do with mindless materialism? If my brain is simply an adaptable algorithm, what does it matter what "I" think my brain can do? Are you not simply dumping the CSI-production into the mysterious algorithm "you"? And would you mind linking some of the "systems materialism" information you speak of?uoflcard
July 5, 2011
July
07
Jul
5
05
2011
10:36 AM
10
10
36
AM
PDT
Elizabeth Liddle:
Well, I think your error is the basic ID error – the assumption Chance and Necessity can’t create CSI!
Do you deny that humans do so all the time? Do you agree that humans can and do create computer programs, such as genetic algorithms? Do these programs require CSI to function? Do you think genetic algorithms can generate CSI? If so, where does the CSI come from?
I don’t think this is anything like demonstrated, and, indeed, I think it is demonstrably untrue.
So now you want the ID theorist to demonstrate that Chance + Necessity could not possibly create CSI? Where have we seen this before?
Bring on “systems materialism”!
Does that mean the old materialism has been falsified?Mung
July 5, 2011
July
07
Jul
5
05
2011
09:15 AM
9
09
15
AM
PDT
Quick notes: ULC: Please cf the discussion of the Glasgow coma scale here. Dr Liddle: Kindly provide a case of chance and mechanical necessity without intelligent direction, producing CSI, especially FSCI. (The just linked would also help, I think.) UB: Indeed, I would like to see that response. GEM of TKIkairosfocus
July 5, 2011
July
07
Jul
5
05
2011
09:14 AM
9
09
14
AM
PDT
Hello Dr Liddle, I have returned, and have looked around for your response to our previous conversation (where you were going to demonstrate the rise of information processing from chance and necessity). The last post to which you said you would address yourself was dated June 17th. If you have made such a response, then please provide a link. If you have not, then, what have you got for me?Upright BiPed
July 5, 2011
July
07
Jul
5
05
2011
09:04 AM
9
09
04
AM
PDT
Well, I think your error is the basic ID error - the assumption Chance and Necessity can't create CSI! I don't think this is anything like demonstrated, and, indeed, I think it is demonstrably untrue. And so my answer would be that your brain has been "programmed" (scare quotes deliberate) both by evolution and by your own actions and experiences to do what you rightly say it can do. We even know a lot about just how those programs work. And yes, although it sounds circular, I don't think it is - a spiral is not a circle, and bootstrapping is of course possible. I'd say that we bootstrap ourselves into consciousness, will, and identity, and it is our brains, honed by billions of years of evolution, that allow us to do this. But another flaw in your logic (from my PoV!) is to characterise materialism as "reducing" our brains and bodies to atoms. As I've said a few times now, reductive materialism is neither effective nor necessary. Bring on "systems materialism"! It's what we do in any case. And the systems by which we make sense of our environment, and decide on appropriate action in light not just of our immediate needs, but our long term goals, desires, and abstract principles is not only fascinating, but well studied :) (And interesting, whether or not you buy the Ghost in the Machine or not).Elizabeth Liddle
July 5, 2011
July
07
Jul
5
05
2011
08:15 AM
8
08
15
AM
PDT
Here is something I've been wondering about lately. I think it's on topic, but sorry if it's not... Our minds output CSI constantly while conscious. This is completely undeniable, even more undeniable than CSI in a genome. How can a materialist possibly account for this? If ultimately our entire body ("mind" included) is reducible to atoms, and the arrangement of these atoms are all reducible to DNA (at least "templates" are reducible to DNA), then our minds must be in our DNA. But how do we get so much CSI (thoughts) out of, relatively, so little (genome)? Shouldn't our brains need to be at least as complex and specified as any thought that comes out of it? But then how do they get programmed to be this complex and specified? It's common knowledge among computer programmers that whatever intelligent information comes out of a program needed to be previously programmed into it, yet this is exactly what "mainstream" atheis...err, I mean biologists must violate in order for their assumptions to be accommodated. And input from the environment doesn't seem to explain it because an input, for a purely material system, must have a pre-programmed "decoder" in order to make sense of it. A security scanner at an airport is great at recognizing guns hidden in suitcases because it has been programmed to look for certain features of known guns. But if I hook a microphone up to it (assuming it even had a port for one), I could tell it I was going to hi-jack flight # xxxx and it wouldn't do a thing. Why? Because it hadn't been programmed to interpret such information. So how does a mindless brain accomplish this task? What is the mainstream response to this prompt? There might be some error in my logic because I'm a little out of my comfort zone with this topic, but I'm curious for a response from both sides of the debate.uoflcard
July 5, 2011
July
07
Jul
5
05
2011
07:32 AM
7
07
32
AM
PDT
F/N: Did a bit of checking on the Chi_500 expression and saw that some are trying the old "painting the target after the arrow hits" talking point. This is doubly erroneous. First, as the observed event E is recognised as coming from a separately definable zone of interest T that must on good grounds be seen as narrow and unrepresentative of the bulk of possibilities W. Second, the detection of specificity of such a configuration is amenable to empirical test or analytical test. That is:
a: We can simply perturb the explicit or implied data string on the ground or in a validated simulation, and see if it makes a difference -- direct or simulation testing of the island of function type case. b: We can see if valid values for T are constrained in such a way as will make them come from a narrow and unrepresentative cluster in W. c: For instance, if the 1,000 coin string is such that the ASCII code equivalents are going to be --
as I have often given as an example, and as has been used as a classic example since Thaxton et al in TMLO, 1985, the very first ID technical work [there is no excuse] --
. . . a contextually responsive string in English, then it is sharply constrained by the requirements for English spelling, grammar and of course the issue of being meaningful and related to a context. So, beyond dispute there will only be a narrow and unrepresentative cross section of the field of possibilities W that will be acceptable. d: Similarly, computer code or data structures starting with strings used in information systems must conform to the requirements of symbolisation conventions, and rules for expressing meaningful and functional possibilities. e: Going further, something like a blueprint, a wiring diagram or an exploded diagram or the specs for a part will be reducible to a set of structured strings, e.g. as CAD software routinely does [PC active storage is based on ordered, stacked strings AKA memory]. f: taking such an entity, one may perturb the string by injecting noise triggering a random walk in a config space. tests can then show whether the resulting varied item will fit within an island of function or whether there is a wide variability that will still allow for a working part. g: We already know -- or should know -- the general result. While there are tolerances, parts in complex systems generally need to be within such ranges and to be appropriately connected to the other parts. h: Just to pick a jet aircraft case, back in the 1950's for the old Saberjet F 86, IIRC, there was a particular bolt in the wing that needed to be put in in a way that was opposite to what had been the usual way around in previous aircraft. There were several fatal crashes, and on tracing, it was found that the problems correlated with shifts by a particular older worker. He had followed his habit, not the blueprint, and had unintentionally killed several men. (I gather the investigators spared him the horrific consequences of that error, but made sure it did not happen again.) i: This is a case where there was a tolerance for error of zero bits. j: Similarly, it is notorious in military aircraft maintenance, that an aircraft that is unfit for service may have a cluster of parts that are all individually in-tolerance but collectively yield a non-functional subsystem. k: In living systems, embryological feasibility of body-plan affecting mutations -- which must be expressed early in development, while the body plan is being laid down in the embryo as it forms -- is a notorious source of lethality of mutation. (That is, body plans per evidence, come in islands of function.)
In short, this is not a case of painting the target after the arrow hits. And, BTW, Dembski pointed this out in NFL a decade or more ago. There is a reason why T is independently and simply describable. GEM of TKI PS: It seems that some have taken up the mistaken epistemological view that if one can object, one may then easily dismiss. This is self-referentially incoherent selective hyperskepticism. Before one justifiably claims to know that A or to know that not-A, alike, one has the same duty of warrant. For, Not-A is just as much a commitment as A. And, for things that are really important, "I don't know so I dismiss" is even less defensible. There are things that one knows or SHOULD know. And if one is actually accepting substantially similar cases but rejects on a prejudice against a particular example, that is irresponsible. Let's make this concrete: we routinely read text and confidently know it comes from an intelligent source, precisely because of the utter unlikelihood that such would result form blind chance and/or necessity. We have in hand something that draws out the reason behind this. To then turn around and say I don't like that this might let a Divine Foot in the door so I dismiss is prejudice not responsible thought. (Especially, as Lewontin's excuse that a world in which miracles are possible is a chaos in which science is impossible is flat out contradicted by the fact that it was theistic thinkers who founded modern science, and did so on the understanding that God is a God of order who sustains the cosmos by his powerful word, which is precisely why we use the odd little term: "LAW of nature." Miracles, as C S Lewis often pointed out, will only work as signs pointing beyond the usual course of the world, if there is such a usual course, i.e., a world in which miracles are possible is ALSO a world in which science is possible.)kairosfocus
July 5, 2011
July
07
Jul
5
05
2011
12:48 AM
12
12
48
AM
PDT
Arkady: Thanks. Given your likely background, do you have some counsel for us? I would love to hear it. GEM of TKIkairosfocus
July 4, 2011
July
07
Jul
4
04
2011
10:53 AM
10
10
53
AM
PDT
Historical footnote.kairosfocus
July 4, 2011
July
07
Jul
4
04
2011
10:19 AM
10
10
19
AM
PDT
KF: "(H’mm, anniversary of the German Attack in France in 1940)" is that just an aside for the history buffs here or is it more significant?paragwinn
July 4, 2011
July
07
Jul
4
04
2011
10:15 AM
10
10
15
AM
PDT
This situation has occured, I think, because this is not about science, but about maintaining control of a system that depends upon accepting a consensus of opinion without question. Unfortunately, it's a cultural hegemony. Lay-persons are as dependant on the system as are the hard-line professional contenders for it. I suspect this is because the philosophical implications of the consensus are congenial to the majority stakeholder's assumptions about the nature of life at a social level. In my opinion, with the exception of a small minority, this is unlikely likely to change, even in the presence of overwhelming evidence. Anyone with a substantial scientific argument for design will likely, and sadly, be be subjected to the same types of attack, and virulence of those attacks will likely be in direct proportion to the strength of the evidence supporting the design inference.arkady967
July 4, 2011
July
07
Jul
4
04
2011
07:26 AM
7
07
26
AM
PDT
1 4 5 6

Leave a Reply