Skip to comments.A Mathematician's View of Evolution
Posted on 09/20/2006 9:51:34 AM PDT by SirLinksalot
A Mathematician's View of Evolution
University of Texas El Paso
The Mathematical Intelligencer 22, no. 4 (2000), pp5-7
Copyright held by Springer Verlag, NY, LLC
In 1996, Lehigh University biochemist Michael Behe published a book entitled "Darwin's Black Box" [Free Press], whose central theme is that every living cell is loaded with features and biochemical processes which are "irreducibly complex"--that is, they require the existence of numerous complex components, each essential for function. Thus, these features and processes cannot be explained by gradual Darwinian improvements, because until all the components are in place, these assemblages are completely useless, and thus provide no selective advantage. Behe spends over 100 pages describing some of these irreducibly complex biochemical systems in detail, then summarizes the results of an exhaustive search of the biochemical literature for Darwinian explanations. He concludes that while biochemistry texts often pay lip-service to the idea that natural selection of random mutations can explain everything in the cell, such claims are pure "bluster", because "there is no publication in the scientific literature that describes how molecular evolution of any real, complex, biochemical system either did occur or even might have occurred."
When Dr. Behe was at the University of Texas El Paso in May of 1997 to give an invited talk, I told him that I thought he would find more support for his ideas in mathematics, physics and computer science departments than in his own field. I know a good many mathematicians, physicists and computer scientists who, like me, are appalled that Darwin's explanation for the development of life is so widely accepted in the life sciences. Few of them ever speak out or write on this issue, however--perhaps because they feel the question is simply out of their domain. However, I believe there are two central arguments against Darwinism, and both seem to be most readily appreciated by those in the more mathematical sciences.
1. The cornerstone of Darwinism is the idea that major (complex) improvements can be built up through many minor improvements; that the new organs and new systems of organs which gave rise to new orders, classes and phyla developed gradually, through many very minor improvements. We should first note that the fossil record does not support this idea, for example, Harvard paleontologist George Gaylord Simpson ["The History of Life," in Volume I of "Evolution after Darwin," University of Chicago Press, 1960] writes:
"It is a feature of the known fossil record that most taxa appear abruptly. They are not, as a rule, led up to by a sequence of almost imperceptibly changing forerunners such as Darwin believed should be usual in evolution...This phenomenon becomes more universal and more intense as the hierarchy of categories is ascended. Gaps among known species are sporadic and often small. Gaps among known orders, classes and phyla are systematic and almost always large. These peculiarities of the record pose one of the most important theoretical problems in the whole history of life: Is the sudden appearance of higher categories a phenomenon of evolution or of the record only, due to sampling bias and other inadequacies?"
An April, 1982, Life Magazine article (excerpted from Francis Hitching's book, "The Neck of the Giraffe: Where Darwin Went Wrong") contains the following report:
"When you look for links between major groups of animals, they simply aren't there...'Instead of finding the gradual unfolding of life', writes David M. Raup, a curator of Chicago's Field Museum of Natural History, 'what geologists of Darwin's time and geologists of the present day actually find is a highly uneven or jerky record; that is, species appear in the fossil sequence very suddenly, show little or no change during their existence, then abruptly disappear.' These are not negligible gaps. They are periods, in all the major evolutionary transitions, when immense physiological changes had to take place."
Even among biologists, the idea that new organs, and thus higher categories, could develop gradually through tiny improvements has often been challenged. How could the "survival of the fittest" guide the development of new organs through their initial useless stages, during which they obviously present no selective advantage? (This is often referred to as the "problem of novelties".) Or guide the development of entire new systems, such as nervous, circulatory, digestive, respiratory and reproductive systems, which would require the simultaneous development of several new interdependent organs, none of which is useful, or provides any selective advantage, by itself? French biologist Jean Rostand, for example, wrote ["A Biologist's View," Wm. Heinemann Ltd. 1956]:
"It does not seem strictly impossible that mutations should have introduced into the animal kingdom the differences which exist between one species and the next...hence it is very tempting to lay also at their door the differences between classes, families and orders, and, in short, the whole of evolution. But it is obvious that such an extrapolation involves the gratuitous attribution to the mutations of the past of a magnitude and power of innovation much greater than is shown by those of today."
Behe's book is primarily a challenge to this cornerstone of Darwinism at the microscopic level. Although we may not be familiar with the complex biochemical systems discussed in this book, I believe mathematicians are well qualified to appreciate the general ideas involved. And although an analogy is only an analogy, perhaps the best way to understand Behe's argument is by comparing the development of the genetic code of life with the development of a computer program. Suppose an engineer attempts to design a structural analysis computer program, writing it in a machine language that is totally unknown to him. He simply types out random characters at his keyboard, and periodically runs tests on the program to recognize and select out chance improvements when they occur. The improvements are permanently incorporated into the program while the other changes are discarded. If our engineer continues this process of random changes and testing for a long enough time, could he eventually develop a sophisticated structural analysis program? (Of course, when intelligent humans decide what constitutes an "improvement", this is really artificial selection, so the analogy is far too generous.)
If a billion engineers were to type at the rate of one random character per second, there is virtually no chance that any one of them would, given the 4.5 billion year age of the Earth to work on it, accidentally duplicate a given 20-character improvement. Thus our engineer cannot count on making any major improvements through chance alone. But could he not perhaps make progress through the accumulation of very small improvements? The Darwinist would presumably say, yes, but to anyone who has had minimal programming experience this idea is equally implausible.
Major improvements to a computer program often require the addition or modification of hundreds of interdependent lines, no one of which makes any sense, or results in any improvement, when added by itself. Even the smallest improvements usually require adding several new lines. It is conceivable that a programmer unable to look ahead more than 5 or 6 characters at a time might be able to make some very slight improvements to a computer program, but it is inconceivable that he could design anything sophisticated without the ability to plan far ahead and to guide his changes toward that plan.
If archeologists of some future society were to unearth the many versions of my PDE solver, PDE2D , which I have produced over the last 20 years, they would certainly note a steady increase in complexity over time, and they would see many obvious similarities between each new version and the previous one. In the beginning it was only able to solve a single linear, steady-state, 2D equation in a polygonal region. Since then, PDE2D has developed many new abilities: it now solves nonlinear problems, time-dependent and eigenvalue problems, systems of simultaneous equations, and it now handles general curved 2D regions.
Over the years, many new types of graphical output capabilities have evolved, and in 1991 it developed an interactive preprocessor, and more recently PDE2D has adapted to 3D and 1D problems. An archeologist attempting to explain the evolution of this computer program in terms of many tiny improvements might be puzzled to find that each of these major advances (new classes or phyla??) appeared suddenly in new versions; for example, the ability to solve 3D problems first appeared in version 4.0. Less major improvements (new families or orders??) appeared suddenly in new subversions, for example, the ability to solve 3D problems with periodic boundary conditions first appeared in version 5.6. In fact, the record of PDE2D's development would be similar to the fossil record, with large gaps where major new features appeared, and smaller gaps where minor ones appeared. That is because the multitude of intermediate programs between versions or subversions which the archeologist might expect to find never existed, because-- for example--none of the changes I made for edition 4.0 made any sense, or provided PDE2D any advantage whatever in solving 3D problems (or anything else) until hundreds of lines had been added.
Whether at the microscopic or macroscopic level, major, complex, evolutionary advances, involving new features (as opposed to minor, quantitative changes such as an increase in the length of the giraffe's neck*, or the darkening of the wings of a moth, which clearly could occur gradually) also involve the addition of many interrelated and interdependent pieces. These complex advances, like those made to computer programs, are not always "irreducibly complex"--sometimes there are intermediate useful stages. But just as major improvements to a computer program cannot be made 5 or 6 characters at a time, certainly no major evolutionary advance is reducible to a chain of tiny improvements, each small enough to be bridged by a single random mutation.
2. The other point is very simple, but also seems to be appreciated only by more mathematically-oriented people. It is that to attribute the development of life on Earth to natural selection is to assign to it--and to it alone, of all known natural "forces"--the ability to violate the second law of thermodynamics and to cause order to arise from disorder. It is often argued that since the Earth is not a closed system--it receives energy from the Sun, for example-- the second law is not applicable in this case. It is true that order can increase locally, if the local increase is compensated by a decrease elsewhere, ie, an open system can be taken to a less probable state by importing order from outside. For example, we could transport a truckload of encyclopedias and computers to the moon, thereby increasing the order on the moon, without violating the second law. But the second law of thermodynamics--at least the underlying principle behind this law--simply says that natural forces do not cause extremely improbable things to happen**, and it is absurd to argue that because the Earth receives energy from the Sun, this principle was not violated here when the original rearrangement of atoms into encyclopedias and computers occurred.
The biologist studies the details of natural history, and when he looks at the similarities between two species of butterflies, he is understandably reluctant to attribute the small differences to the supernatural. But the mathematician or physicist is likely to take the broader view. I imagine visiting the Earth when it was young and returning now to find highways with automobiles on them, airports with jet airplanes, and tall buildings full of complicated equipment, such as televisions, telephones and computers. Then I imagine the construction of a gigantic computer model which starts with the initial conditions on Earth 4 billion years ago and tries to simulate the effects that the four known forces of physics (the gravitational, electromagnetic and strong and weak nuclear forces) would have on every atom and every subatomic particle on our planet (perhaps using random number generators to model quantum uncertainties!). If we ran such a simulation out to the present day, would it predict that the basic forces of Nature would reorganize the basic particles of Nature into libraries full of encyclopedias, science texts and novels, nuclear power plants, aircraft carriers with supersonic jets parked on deck, and computers connected to laser printers, CRTs and keyboards? If we graphically displayed the positions of the atoms at the end of the simulation, would we find that cars and trucks had formed, or that supercomputers had arisen? Certainly we would not, and I do not believe that adding sunlight to the model would help much. Clearly something extremely improbable has happened here on our planet, with the origin and development of life, and especially with the development of human consciousness and creativity.
*Ironically, W.E.Loennig's article "The Evolution of the Long-necked Giraffe," has since convinced me that even this feature could not, and did not, arise gradually.
**An unfortunate choice of words, for which I was severely chastised. I should have said, the underlying principle behind the second law is that natural forces do not do macroscopically describable things which are extremely improbable from the microscopic point of view. See "A Second Look at the Second Law," for a more thorough treatment of this point.
Granville Sewell completed his PhD at Purdue University. He has subsequently been employed by (in chronological order) Universidad Simon Bolivar (Caracas), Oak Ridge National Laboratory, Purdue University, IMSL (Houston), The University of Texas Center for High Performance Computing (Austin), and the University of Texas El Paso; he spent Fall 1999 at Universidad Nacional de Tucuman in Argentina on a Fulbright grant. He has written three books on numerical analysis.
This may seem like a pedant debate - but I really have an issue with this statement. Maybe I am wrong but with my current data this statement doesn't sound correct. I am not trying to be a pain - I was concede if you show me I am wrong.
Wolves stalk their prey - the first time the prey realizes the wolf is there, the wolf is somewhat near - no more than 300 feet (guessing). A wolf can sprint at 40 MPH for say two minutes and then it can run at 7 MPH for as much 7 hours (I got that from a link). Man can run say 18 MPH for a long time. The wolf is going to over take the man in the first few minutes. (math is not my strong subject so maybe I am miscalculating - or not calculating at all)
Our running ability is useful for running things down, not for getting away from things.
After reading that post I am really confused what yours could be!
All the calculations in the world are useless if you're not asking the right question. Their calculations assume that an enzyme or bacterium springs fully formed from the forehead of Zeus, when in actuality nothing of the sort happened.
Why run something down when you can lie in wait for it? It would seem that for survival, the ability to run faster than the thing that's chasing you to eat you would be more critical than running down prey.
Why expend that kind of energy for that long when there are far easier things to find and eat that don't require that kind of energy output. I cannot imagine that some pre-human is going to chase things for those kind of long distances anyway. They'd be getting too far from home.
So what would the evolutionary advantage be to be able to outrun ANY creature for those kind of long distances?
Why don't you go ask all of the creatures that chase down their prey and also are slower than some other predator?
What? Don't the evolutionists have the answer to that one that I have to be sent off packing to ask an animal?
"All the calculations in the world are useless if you're not asking the right question."
The question: What is the probability of life arriving solely by chance and natural process? is the right question. It is the crux of the issue.
a) Calculations of Sir Fred Hoyle and Chandra Wickramasinghe for random generation of a simple enzyme and calculations for a single celled bacterium.
Although he is an evolutionist, and an atheist, Hoyle sees the mathematical statistical difficulty in producing a single bacterium like E. coli. In his calculations of the probability of life emerging from chance interactions with chemicals, Hoyle assumed that the first living cell was much simpler than todays bacteria. However, his calculation for the likelihood of even one very simple enzyme arising at the right time in the right place was only chance in 10^(20). Because there are thousands of different enzymes with different functions, to produce the simplest living cell, Hoyle calculated that about 2,000 enzymes were needed with each one performing a specific task to form a single bacterium lie E coli.
No matter how large the environment one considers, life cannot have a random beginning .there are about two thousand enzymes, and the chance of obtaining them all in a random trial is only one part in (10^20)^2000 = 10^40,000, an outrageously small probability that could not be faced even if the whole universe consisted of organic soup. If one is not prejudiced either by social beliefs or by a scientific training into the conviction that life originated on the Earth, this simple calculation wipes the idea entirely out of court .the enormous information content of even the simplest living systems .cannot in out view be generated by what are often called natural processes, For life to have originated on the Earth it would be necessary that quite explicit instruction should have been provided for its assembly There is no way in which we can expect to avoid the need for information, no way in which we can simply get by with a bigger and better organic soup, as we ourselves hoped might be possible a year or two ago.
-Hoyle & Wickramasinghe, Evolution from Space (London: J.M. Dent & Sons, 1981).
Other calculation models posted here:
Um, I think if you were patient enough to read my entire post, you would see that that was sarcasm......
Ah....and I posted my correction without reading your next post.....a lesson for us all!
Then the numbers are even worse - the overwhelming sprint speed advantage of four legged mammals would mean man is not going to run down much.
You don't make much sense
I suppose if repeating arguments I've already conceded or excluded could be called pedantry. Among other things, I started by conceding that many animals could outsprint a human.
No would you like to consider the situation where the human is chasing the prey, and the prey is not a carnivore, or not chasing the human?
Our ability to cover a lot of ground at a steady pace is also useful for foraging as well as for hunting.
That's what I think about you. Glad we are of one mind.
You are assuming a prey animal can keep up a sprint speed over a long distance in the summer heat. They can't. A typical human jogger can kill his pet dog by overheating it.
Disclaimer: Opinions posted on Free Republic are those of the individual posters and do not necessarily represent the opinion of Free Republic or its management. All materials posted herein are protected by copyright law and the exemption for fair use of copyrighted works.