Tuesday, July 21, 2009

Why should self-organisation be limited?

My main research project of the past decade has been an investigation of physical constraints on self organisation. This sounds like a dry topic, I know, but it is rich with practical applications as well as being of great philosophical interest. My latest paper on the subject may be downloaded here.

Certain kinds of object, even rather large, intricate ones, will appear spontaneously under fairly general conditions. Crystals are the most obvious example of this. A sodium chloride crystal, for example, is a very specific arrangement of sodium and chlorine atoms. The chance of a large crystal forming by chance is too remote to consider. If, for example, you were to arrange black counters and white counters randomly on a large square grid, you would not produce the perfect pattern of alternation, like a chess board, that we see in sodium chloride. The appearance of sodium chloride is therefore a case of self-organisation.

Probably the most impressive examples of self organisation are fractals. The Mandelbrot set, for example, is enormously intricate, one might even say infinitely so, yet it is generated by a rather simple rule. Also, while the Mandelbrot set is a mathematical abstraction rather than a natural object, there are apparently natural fractals.

These examples, while impressive, point to a possible limitation of self organisation, however. For crystals and fractals share the feature of being highly self similar. If you move a crystal over by one atom, it maps exactly onto itself. It is "invariant", as they say, "under spatial translation". Fractals don't have this kind of self similarity. But they are self-similar, or invariant, under changes of scale. As you progressively "zoom in" on the Mandelbrot set, for example, you just see more and more of the same. So, despite being very intricate, crystals and fractals are kinda boring. This raises the following questions:
  1. Do self-organised objects have to be self similar?
  2. If they do, then why is this?
These are hard questions that I have long studied, and can now finally answer with some confidence. The answers are:

  1. Yes
  2. It results from symmetries in the underlying dynamics, specifically causal locality and invariance under spatial translation.
If these answers are indeed correct, then there are some important consequences for biology, concerning theories of the origin of life. For living organisms are quite unlike crystals and fractals in this respect. While there is a lot of self similarity in animals and plants, there is also a lot of differentiation. It appears then that living organisms cannot be products of self organisation, as is usually supposed.

I'm not going to repeat the whole argument here, as it can be found in my paper. But I will reflect on the argument, from three different perspectives. The argument can be seen supporting a limitative principle, analogous to Gödel's incompleteness theorem, the Löwenheim-Skolem theorem, or Bell's theorem. The argument is a fairly typical symmetry argument. The argument also stands in a tradition of arguing that a cause can produce an effect only if the effect is somehow "present" in the cause.


I. Limitative Results: Gödel, Bell and Jeremiah

There is a fairly honourable tradition of limitative, or "no go" theorems in science. One of Gödel's theorems, for example, entails that there is no complete system of rules of inference in second-order logic, or indeed in any formal language in which a categorical set of axioms of arithmetic exists. This result is disappointing to many, since it shows that rational inference cannot be "captured", as it were, in rules that manipulate symbols. (Whether human inference can be so captured is another matter. I like to say that Gödel's theorem shows that God isn't a machine.)

Another delicious result is Bell's theorem, concerning quantum theory, which says that no local, hidden-variable theory can be empirically equivalent to standard quantum mechanics.

Each of these theorems, like all such theorems, is negative. It says that, while you might keep trying to do a certain thing, it will never work. You might say they are logical Jeremiahs, or prophets of doom. This would be a little unfair to Jeremiah, however, since even that much-maligned prophet did urge an alternative to defeat and exile. These results, by contrast, make no positive suggestion at all. Bell's theorem gives no clue about how to interpret quantum mechanics, and Gödel's theorem doesn't suggest any better understanding of rational inference.

My theorem is also pure doom and gloom. Let's be honest: It offers no positive suggestion at all.

Can such negative claims be part of science? It is often said that a scientist must propose hypotheses that are empirically testable. That's not really true, however. While that's a big part of science, a lot of good scientific work is indeed negative. Much useful work is done by experimentalists who show that, while hypothesis H might predict empirical result E, E doesn't actually occur. Also, while most theorists are busily showing that H predicts E, other theorists very helpfully point out that H doesn't really predict E at all, even though we thought it did. A really negative scientist might even show that no hypothesis of a certain type will ever predict E.

I'm afraid I'm one of those really negative scientists. I've shown that no hypothesis in a very broad class predicts the existence of complex living organisms. More precisely, life cannot self organise in any dynamical system whose laws are local and invariant under spatial translation.

At this point a worrying possibility emerges. This no-go theorem is so broad that it rules out just about any naturalistic theory of the origin of life! It certainly seems to rule out all the naturalistic theories presently proposed. Yet, the whole business of science is to provide natural explanations for phenomena, so this result is unscientific after all. (Even if it is technically correct, take note.)

Well, this is an awkward business! What are we to do?

1. It is surely irrational to disregard results that attack the viability of one's project, simply on the grounds that they attack the viability of one's project!

2. While I regard my limitative result as scientific, I myself am a philosopher (of science) by trade, so I don't mind too much if it's considered philosophy. But it is science really.

3. This kind of thing has happened before. Maxwell found that his equations for the electric and magnetic fields were incompatible with the mechanical philosophy of his day. In those days, to give a scientific explanation of something was to give a mechanical model for it. Yet Maxwell, despite much effort, was unable to understand his equations as describing displacement fields in any kind of solid. He had, in effect, a limitative theorem that no such model was possible. Yet science didn't end in the late 19th century! Instead, they realised that non-mechanical explanations are better than none. And if the world isn't a giant machine, then those are going to be the true explanations.

4. So I think we have to broaden our horizons, and be open to new kinds of explanation. Perhaps it won't be that bad? And we have no other choice, if we want our explanations to be true.

[A mischievous aside. I like to say that Maxwell and Einstein refuted the mechanical philosophy over a century ago, but no one told the biologists! Now, one might reasonably object to this, saying that a model that's been shown to be inadequate in one context might work perfectly well elsewhere. I see. So mechanical models are too clumsy to handle the subtleties of the electromagnetic field, but are adequate for such simple phenomena as life, consciousness and rational thought?]


II. Symmetry Arguments

Symmetry arguments are common in logic, mathematics and physics. In logic, it is generally held that logical consequence preserves symmetry, so that if the premises of an argument (considered together) are symmetric in some respect, then the conclusion must also be symmetric in that respect. I will take this for granted. For example, suppose you're investigating a murder, and you have some evidence that Fred did it. On the other hand, you have exactly the same sort of evidence that Mike did it. (Perhaps you know that the murderer is tall, left-handed and male, and both Fred and Mike have these attributes. They have equal motives, and so on.) In this epistemic situation, you cannot draw a conclusion about Fred doing the murder without drawing a similar conclusion about Mike. In other words, you should assign the same probability to each.

In some physical experiments, the initial situation is symmetric in some respect. For example, the state might have reflective symmetry about some plane, so that the state maps to itself under reflection in that plane. What should we believe about the outcome of this experiment, given this initial symmetry?

Two answers have been considered here:

  1. The outcome will have reflective symmetry about that plane.
  2. The probability function over the outcomes will have reflective symmetry about that plane.
Which of these answers you should accept depends on whether or not you are a determinist. I shall first prove (2), and then derive (1) from (2) by adding the assumption of determinism.

Nothing interesting can be proved without premises, of course. What premise(s) will allow me to prove (2)? The first premise is the one above, that logical consequence preserves symmetry. The second premise is what philosophers call the Principal Principle, as follows:

If the chance of a future event E is known, then the epistemic probability of E is equal to it.

So, for example, if you know that a coin is biased in such a way that its chance of heads is 0.331 on each toss, then the subjective utility of a gamble that pays $1 on heads (and nothing on tails) is 33.1 cents. The Principal Principle, called PP for short, is almost universally accepted.

The third premise is no less certain. It says that the chance of an outcome in an experiment depends only on the possible causes of that outcome, i.e. on the causally relevant physical conditions. In other words, if you repeat an experiment exactly, perfectly duplicating all the relevant physical circumstances, then the same chances obtain for the outcomes. This doesn't mean that the outcomes will be the same, except in the deterministic case. For example, suppose you do some experiment with the Stern-Gerlach apparatus, and the chance of the particle going UP is q. Then, in every exact repetition of this experiment, the chance of UP will be q as well.

Let's refer to this third premise as the claim that chances are physical.

With these three premises in place, we now consider some ideal physicist who has maximal knowledge of the initial state of some experiment, together with maximal knowledge of all the relevant dynamical laws. Since chances are physical, it follows that this physicist will know the chances of all possible outcomes of the experiment. Now we assume that the initial conditions are symmetric in some way, and consider what this physicist can infer about the outcome of the experiment. Any such conclusion can only be inferred from what the physicist knows, i.e. from his knowledge of the initial conditions. Now, since logical consequence preserves symmetry, any symmetric outcomes A and B must have equal epistemic probability for this physicist. Then, using the Principal Principle, the outcomes A and B must have equal chances as well. (If they did not, then PP entails that their epistemic probabilities would be different.)

This concludes our proof of (2). Now suppose that the system is deterministic, so that the chance of each outcome is either 0 or 1. We will argue by reductio ad absurdum. Suppose that the initial state of the experiment has symmetry with respect to transformation R, so that the initial state maps to itself under R, but the outcome of the experiment does not. In other words, if the actual outcome is E, then R(E) ≠ E. Now, assuming determinism, whatever occurs has chance 1, and whatever doesn't occur has chance 0. R(E) doesn't occur, and so has chance 0. E occurs, and so has chance 1. But, using (2), symmetric outcomes have equal chance. This is a contradiction. Thus the result is proven, i.e. the outcome must be R-symmetric as well in the deterministic case.

My proofs of (1) and (2) show that symmetry arguments are valid in physics as well as in logic and mathematics.

My limitative theorem is established, in part, by just such a symmetry argument. I assume that the dynamical laws operate locally, and are invariant under arbitrary spatial translation. These are rather strong symmetries to impose, but they are well accepted by physicists. I assume a random initial state, so that every possible initial state has the same chance. Such an initial chance distribution has just about every kind of symmetry you might name.

To be continued ...









Tuesday, June 30, 2009

Column resistance -- some thoughts on Chandler, Szamboti and Bazant

There has been some recent discussion on 911Blogger about one of David Chandler's videos on his YouTube channel.

http://www.youtube.com/watch?v=xG2y50Wyys4

I have some points to make here.

1. Chandler's argument, while correct as far as it goes, establishes only an upper bound on the column resistance. A better empirical estimate is roughly 0 - 0.1mg. This can be seen as follows.

The video measures the downward acceleration of the roofline of WTC1, during the first few seconds of collapse, and measures a downward acceleration of around 0.64g. Chandler argues, correctly, that this implies that the block is resisted by an upward force of (1-0.64)mg = 0.36mg, where m is the mass of the block. Yet this is puzzling, since the columns were (just a moment before) supporting the full weight (mg) of this block. (Moreover, assuming a safety factor of 3, they the columns were capable of supporting at least 3mg, almost 10 times their measured resistance.)

Chandler's video is, in my opinion, a useful device to attract attention to this important puzzle. However, while the reasoning is correct as far as it goes, the true situation is actually far more disturbing than even his argument suggests. (I assume that he is aware of this, but wants to keep the argument as simple as possible.) Let me explain why.

Chandler's argument is usefully seen as being based on an idealised model of the tower, where the floors below the airplane impacts have zero mass, and hence no inertia. Also, in his model there is no energy consumed in the pulverisation of the concrete to a fine powder. The only force resisting the descent of the upper floors (above the plane impacts) therefore is due to the steel columns. In reality, of course, the lower floors have inertia, and concrete requires force to crush it. Hence Chandler's quick argument greatly over-estimates the column resistance, and so should be regarded as an upper bound rather than an estimate. The true situation is even more worrying than his argument suggests. But what really was the column resistance?

The best way to measure this from the video evidence, I think, is to use a different idealised model, based on what Kenneth Kuttler has called a "floating floor" model. In this model, the floors have inertia, which resists the acceleration of the falling block. One also includes an energy drain for the crushing of concrete, as well as a variable drain for the deformation of the columns. (For convenience, it's best to calculate using energies rather than forces, although of course the work done by a given force is simply the integral of that force with respect to displacement.) In such a model one easily calculates the downward acceleration of the top block, for a given value of the column resistance. One then adjusts that resistance value to bring the model's behaviour into line with reality, as seen in the videos. This is then a reasonably accurate empirical estimate of the mean column resistance.

What value is obtained by such a procedure? I haven't yet finished my analysis, but it is certainly very low -- much less than Chandler's upper bound of 0.36mg. I believe it is in the range 0 to 0.1mg -- somewhere around that.


2. The papers by Zdenek Bazant on the WTC collapses provide an alleged theoretical estimate that roughly agrees with this empirical estimate. Bazant compares the potential energy released as the "top block" descends one story with the energy absorbed by the plastic deformation of the columns over the same distance. He concludes (without, unfortunately, showing much of his calculation) that the energy released is at least 8.4 times the energy absorbed, so that the collapse will inevitably progress. In other words, the columns absorb about 12%, at most, of the energy released. Now, this energy fraction is easily converted to a force, using the fact that the work done to the columns must equal the integral of the force applied to them, with respect to the distance fallen. In other words, the average resistance of the columns is at most 12% of the weight of the block. This is at the upper end of what is allowed by the empirical data.


3. We see therefore that Chandler's argument will not disturb Bazant in any way, who has already accepted a much lower value for the column resistance. On the other hand, if Chandler's empirical upper bound for the column resistance is worryingly low, then Bazant's theoretical estimate is far more so! One may even wonder whether Bazant's estimate is genuine, given that he doesn't explain how he obtained it.

These darker suspicions are supported by the fact that Bazant's figure lies at the upper end of what is empirically possible. If one were to invent a figure, say to reassure the engineering community, then this would be the perfect value to choose. One would choose the highest, i.e. most plausible, value that was consistent with the data.


4. How is a non-specialist like me to evaluate the reasonableness of Bazant's upper bound? One obvious way is to compare it with the estimates of other experts. I am aware of two of these, one due to Gregory Szuladzinski in Journal of Engineering Mechanics and the other from Tony Szamboti in "The Missing Jolt", Journal of 9/11 Studies.

The estimates are similar in approach. They calculate the energy absorbed from the elastic compression, then plastic shortening, then plastic buckling of the columns. The resistive force of a column drops sharply after it begins to buckle, but it seems to remain above about 25% of the yield stress until the thing breaks ("fractures").

The outcomes of the two calculations are also similar, giving average resistances above mg, certainly far in excess of Bazant's alleged upper bound of 0.12mg. Szuladzinski and Szamboti both predict collapse arrest.


5. In JEM, Bazant replies to Szuladzinski, finding some ten or so errors in his calculation. (Although there seems to be some padding here, as correcting some of these "errors" would increase, rather than decrease, the resistance.) It is hard for me, as a non-specialist, to judge the validity of these objections, although I suspect that they are minor quibbles. It is telling, I think, that Bazant does not correct the calculation, showing what it ought to be. This indicates to me that the outcome would not be too different.

To avoid such trifling objections, if that's what they are, it would be useful to calculate a robust lower bound for the column resistance. Can we have that, please? By "robust" I mean that all reasonable allowances for lowering the column resistance have already been made, so that it is the lowest possible value. It is quibble-proof, so to speak.

I repeat (to all relevant experts): Can we have that, please?